aws-sdk-neptunedata 1.0.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +7 -0
- data/CHANGELOG.md +8 -0
- data/LICENSE.txt +202 -0
- data/VERSION +1 -0
- data/lib/aws-sdk-neptunedata/client.rb +2958 -0
- data/lib/aws-sdk-neptunedata/client_api.rb +1958 -0
- data/lib/aws-sdk-neptunedata/customizations.rb +0 -0
- data/lib/aws-sdk-neptunedata/endpoint_parameters.rb +66 -0
- data/lib/aws-sdk-neptunedata/endpoint_provider.rb +54 -0
- data/lib/aws-sdk-neptunedata/endpoints.rb +618 -0
- data/lib/aws-sdk-neptunedata/errors.rb +970 -0
- data/lib/aws-sdk-neptunedata/plugins/endpoints.rb +154 -0
- data/lib/aws-sdk-neptunedata/resource.rb +26 -0
- data/lib/aws-sdk-neptunedata/types.rb +3988 -0
- data/lib/aws-sdk-neptunedata.rb +57 -0
- metadata +94 -0
@@ -0,0 +1,2958 @@
|
|
1
|
+
# frozen_string_literal: true
|
2
|
+
|
3
|
+
# WARNING ABOUT GENERATED CODE
|
4
|
+
#
|
5
|
+
# This file is generated. See the contributing guide for more information:
|
6
|
+
# https://github.com/aws/aws-sdk-ruby/blob/version-3/CONTRIBUTING.md
|
7
|
+
#
|
8
|
+
# WARNING ABOUT GENERATED CODE
|
9
|
+
|
10
|
+
require 'seahorse/client/plugins/content_length.rb'
|
11
|
+
require 'aws-sdk-core/plugins/credentials_configuration.rb'
|
12
|
+
require 'aws-sdk-core/plugins/logging.rb'
|
13
|
+
require 'aws-sdk-core/plugins/param_converter.rb'
|
14
|
+
require 'aws-sdk-core/plugins/param_validator.rb'
|
15
|
+
require 'aws-sdk-core/plugins/user_agent.rb'
|
16
|
+
require 'aws-sdk-core/plugins/helpful_socket_errors.rb'
|
17
|
+
require 'aws-sdk-core/plugins/retry_errors.rb'
|
18
|
+
require 'aws-sdk-core/plugins/global_configuration.rb'
|
19
|
+
require 'aws-sdk-core/plugins/regional_endpoint.rb'
|
20
|
+
require 'aws-sdk-core/plugins/endpoint_discovery.rb'
|
21
|
+
require 'aws-sdk-core/plugins/endpoint_pattern.rb'
|
22
|
+
require 'aws-sdk-core/plugins/response_paging.rb'
|
23
|
+
require 'aws-sdk-core/plugins/stub_responses.rb'
|
24
|
+
require 'aws-sdk-core/plugins/idempotency_token.rb'
|
25
|
+
require 'aws-sdk-core/plugins/jsonvalue_converter.rb'
|
26
|
+
require 'aws-sdk-core/plugins/client_metrics_plugin.rb'
|
27
|
+
require 'aws-sdk-core/plugins/client_metrics_send_plugin.rb'
|
28
|
+
require 'aws-sdk-core/plugins/transfer_encoding.rb'
|
29
|
+
require 'aws-sdk-core/plugins/http_checksum.rb'
|
30
|
+
require 'aws-sdk-core/plugins/checksum_algorithm.rb'
|
31
|
+
require 'aws-sdk-core/plugins/request_compression.rb'
|
32
|
+
require 'aws-sdk-core/plugins/defaults_mode.rb'
|
33
|
+
require 'aws-sdk-core/plugins/recursion_detection.rb'
|
34
|
+
require 'aws-sdk-core/plugins/sign.rb'
|
35
|
+
require 'aws-sdk-core/plugins/protocols/rest_json.rb'
|
36
|
+
|
37
|
+
Aws::Plugins::GlobalConfiguration.add_identifier(:neptunedata)
|
38
|
+
|
39
|
+
module Aws::Neptunedata
|
40
|
+
# An API client for Neptunedata. To construct a client, you need to configure a `:region` and `:credentials`.
|
41
|
+
#
|
42
|
+
# client = Aws::Neptunedata::Client.new(
|
43
|
+
# region: region_name,
|
44
|
+
# credentials: credentials,
|
45
|
+
# # ...
|
46
|
+
# )
|
47
|
+
#
|
48
|
+
# For details on configuring region and credentials see
|
49
|
+
# the [developer guide](/sdk-for-ruby/v3/developer-guide/setup-config.html).
|
50
|
+
#
|
51
|
+
# See {#initialize} for a full list of supported configuration options.
|
52
|
+
class Client < Seahorse::Client::Base
|
53
|
+
|
54
|
+
include Aws::ClientStubs
|
55
|
+
|
56
|
+
@identifier = :neptunedata
|
57
|
+
|
58
|
+
set_api(ClientApi::API)
|
59
|
+
|
60
|
+
add_plugin(Seahorse::Client::Plugins::ContentLength)
|
61
|
+
add_plugin(Aws::Plugins::CredentialsConfiguration)
|
62
|
+
add_plugin(Aws::Plugins::Logging)
|
63
|
+
add_plugin(Aws::Plugins::ParamConverter)
|
64
|
+
add_plugin(Aws::Plugins::ParamValidator)
|
65
|
+
add_plugin(Aws::Plugins::UserAgent)
|
66
|
+
add_plugin(Aws::Plugins::HelpfulSocketErrors)
|
67
|
+
add_plugin(Aws::Plugins::RetryErrors)
|
68
|
+
add_plugin(Aws::Plugins::GlobalConfiguration)
|
69
|
+
add_plugin(Aws::Plugins::RegionalEndpoint)
|
70
|
+
add_plugin(Aws::Plugins::EndpointDiscovery)
|
71
|
+
add_plugin(Aws::Plugins::EndpointPattern)
|
72
|
+
add_plugin(Aws::Plugins::ResponsePaging)
|
73
|
+
add_plugin(Aws::Plugins::StubResponses)
|
74
|
+
add_plugin(Aws::Plugins::IdempotencyToken)
|
75
|
+
add_plugin(Aws::Plugins::JsonvalueConverter)
|
76
|
+
add_plugin(Aws::Plugins::ClientMetricsPlugin)
|
77
|
+
add_plugin(Aws::Plugins::ClientMetricsSendPlugin)
|
78
|
+
add_plugin(Aws::Plugins::TransferEncoding)
|
79
|
+
add_plugin(Aws::Plugins::HttpChecksum)
|
80
|
+
add_plugin(Aws::Plugins::ChecksumAlgorithm)
|
81
|
+
add_plugin(Aws::Plugins::RequestCompression)
|
82
|
+
add_plugin(Aws::Plugins::DefaultsMode)
|
83
|
+
add_plugin(Aws::Plugins::RecursionDetection)
|
84
|
+
add_plugin(Aws::Plugins::Sign)
|
85
|
+
add_plugin(Aws::Plugins::Protocols::RestJson)
|
86
|
+
add_plugin(Aws::Neptunedata::Plugins::Endpoints)
|
87
|
+
|
88
|
+
# @overload initialize(options)
|
89
|
+
# @param [Hash] options
|
90
|
+
# @option options [required, Aws::CredentialProvider] :credentials
|
91
|
+
# Your AWS credentials. This can be an instance of any one of the
|
92
|
+
# following classes:
|
93
|
+
#
|
94
|
+
# * `Aws::Credentials` - Used for configuring static, non-refreshing
|
95
|
+
# credentials.
|
96
|
+
#
|
97
|
+
# * `Aws::SharedCredentials` - Used for loading static credentials from a
|
98
|
+
# shared file, such as `~/.aws/config`.
|
99
|
+
#
|
100
|
+
# * `Aws::AssumeRoleCredentials` - Used when you need to assume a role.
|
101
|
+
#
|
102
|
+
# * `Aws::AssumeRoleWebIdentityCredentials` - Used when you need to
|
103
|
+
# assume a role after providing credentials via the web.
|
104
|
+
#
|
105
|
+
# * `Aws::SSOCredentials` - Used for loading credentials from AWS SSO using an
|
106
|
+
# access token generated from `aws login`.
|
107
|
+
#
|
108
|
+
# * `Aws::ProcessCredentials` - Used for loading credentials from a
|
109
|
+
# process that outputs to stdout.
|
110
|
+
#
|
111
|
+
# * `Aws::InstanceProfileCredentials` - Used for loading credentials
|
112
|
+
# from an EC2 IMDS on an EC2 instance.
|
113
|
+
#
|
114
|
+
# * `Aws::ECSCredentials` - Used for loading credentials from
|
115
|
+
# instances running in ECS.
|
116
|
+
#
|
117
|
+
# * `Aws::CognitoIdentityCredentials` - Used for loading credentials
|
118
|
+
# from the Cognito Identity service.
|
119
|
+
#
|
120
|
+
# When `:credentials` are not configured directly, the following
|
121
|
+
# locations will be searched for credentials:
|
122
|
+
#
|
123
|
+
# * `Aws.config[:credentials]`
|
124
|
+
# * The `:access_key_id`, `:secret_access_key`, and `:session_token` options.
|
125
|
+
# * ENV['AWS_ACCESS_KEY_ID'], ENV['AWS_SECRET_ACCESS_KEY']
|
126
|
+
# * `~/.aws/credentials`
|
127
|
+
# * `~/.aws/config`
|
128
|
+
# * EC2/ECS IMDS instance profile - When used by default, the timeouts
|
129
|
+
# are very aggressive. Construct and pass an instance of
|
130
|
+
# `Aws::InstanceProfileCredentails` or `Aws::ECSCredentials` to
|
131
|
+
# enable retries and extended timeouts. Instance profile credential
|
132
|
+
# fetching can be disabled by setting ENV['AWS_EC2_METADATA_DISABLED']
|
133
|
+
# to true.
|
134
|
+
#
|
135
|
+
# @option options [required, String] :region
|
136
|
+
# The AWS region to connect to. The configured `:region` is
|
137
|
+
# used to determine the service `:endpoint`. When not passed,
|
138
|
+
# a default `:region` is searched for in the following locations:
|
139
|
+
#
|
140
|
+
# * `Aws.config[:region]`
|
141
|
+
# * `ENV['AWS_REGION']`
|
142
|
+
# * `ENV['AMAZON_REGION']`
|
143
|
+
# * `ENV['AWS_DEFAULT_REGION']`
|
144
|
+
# * `~/.aws/credentials`
|
145
|
+
# * `~/.aws/config`
|
146
|
+
#
|
147
|
+
# @option options [String] :access_key_id
|
148
|
+
#
|
149
|
+
# @option options [Boolean] :active_endpoint_cache (false)
|
150
|
+
# When set to `true`, a thread polling for endpoints will be running in
|
151
|
+
# the background every 60 secs (default). Defaults to `false`.
|
152
|
+
#
|
153
|
+
# @option options [Boolean] :adaptive_retry_wait_to_fill (true)
|
154
|
+
# Used only in `adaptive` retry mode. When true, the request will sleep
|
155
|
+
# until there is sufficent client side capacity to retry the request.
|
156
|
+
# When false, the request will raise a `RetryCapacityNotAvailableError` and will
|
157
|
+
# not retry instead of sleeping.
|
158
|
+
#
|
159
|
+
# @option options [Boolean] :client_side_monitoring (false)
|
160
|
+
# When `true`, client-side metrics will be collected for all API requests from
|
161
|
+
# this client.
|
162
|
+
#
|
163
|
+
# @option options [String] :client_side_monitoring_client_id ("")
|
164
|
+
# Allows you to provide an identifier for this client which will be attached to
|
165
|
+
# all generated client side metrics. Defaults to an empty string.
|
166
|
+
#
|
167
|
+
# @option options [String] :client_side_monitoring_host ("127.0.0.1")
|
168
|
+
# Allows you to specify the DNS hostname or IPv4 or IPv6 address that the client
|
169
|
+
# side monitoring agent is running on, where client metrics will be published via UDP.
|
170
|
+
#
|
171
|
+
# @option options [Integer] :client_side_monitoring_port (31000)
|
172
|
+
# Required for publishing client metrics. The port that the client side monitoring
|
173
|
+
# agent is running on, where client metrics will be published via UDP.
|
174
|
+
#
|
175
|
+
# @option options [Aws::ClientSideMonitoring::Publisher] :client_side_monitoring_publisher (Aws::ClientSideMonitoring::Publisher)
|
176
|
+
# Allows you to provide a custom client-side monitoring publisher class. By default,
|
177
|
+
# will use the Client Side Monitoring Agent Publisher.
|
178
|
+
#
|
179
|
+
# @option options [Boolean] :convert_params (true)
|
180
|
+
# When `true`, an attempt is made to coerce request parameters into
|
181
|
+
# the required types.
|
182
|
+
#
|
183
|
+
# @option options [Boolean] :correct_clock_skew (true)
|
184
|
+
# Used only in `standard` and adaptive retry modes. Specifies whether to apply
|
185
|
+
# a clock skew correction and retry requests with skewed client clocks.
|
186
|
+
#
|
187
|
+
# @option options [String] :defaults_mode ("legacy")
|
188
|
+
# See {Aws::DefaultsModeConfiguration} for a list of the
|
189
|
+
# accepted modes and the configuration defaults that are included.
|
190
|
+
#
|
191
|
+
# @option options [Boolean] :disable_host_prefix_injection (false)
|
192
|
+
# Set to true to disable SDK automatically adding host prefix
|
193
|
+
# to default service endpoint when available.
|
194
|
+
#
|
195
|
+
# @option options [Boolean] :disable_request_compression (false)
|
196
|
+
# When set to 'true' the request body will not be compressed
|
197
|
+
# for supported operations.
|
198
|
+
#
|
199
|
+
# @option options [String] :endpoint
|
200
|
+
# The client endpoint is normally constructed from the `:region`
|
201
|
+
# option. You should only configure an `:endpoint` when connecting
|
202
|
+
# to test or custom endpoints. This should be a valid HTTP(S) URI.
|
203
|
+
#
|
204
|
+
# @option options [Integer] :endpoint_cache_max_entries (1000)
|
205
|
+
# Used for the maximum size limit of the LRU cache storing endpoints data
|
206
|
+
# for endpoint discovery enabled operations. Defaults to 1000.
|
207
|
+
#
|
208
|
+
# @option options [Integer] :endpoint_cache_max_threads (10)
|
209
|
+
# Used for the maximum threads in use for polling endpoints to be cached, defaults to 10.
|
210
|
+
#
|
211
|
+
# @option options [Integer] :endpoint_cache_poll_interval (60)
|
212
|
+
# When :endpoint_discovery and :active_endpoint_cache is enabled,
|
213
|
+
# Use this option to config the time interval in seconds for making
|
214
|
+
# requests fetching endpoints information. Defaults to 60 sec.
|
215
|
+
#
|
216
|
+
# @option options [Boolean] :endpoint_discovery (false)
|
217
|
+
# When set to `true`, endpoint discovery will be enabled for operations when available.
|
218
|
+
#
|
219
|
+
# @option options [Boolean] :ignore_configured_endpoint_urls
|
220
|
+
# Setting to true disables use of endpoint URLs provided via environment
|
221
|
+
# variables and the shared configuration file.
|
222
|
+
#
|
223
|
+
# @option options [Aws::Log::Formatter] :log_formatter (Aws::Log::Formatter.default)
|
224
|
+
# The log formatter.
|
225
|
+
#
|
226
|
+
# @option options [Symbol] :log_level (:info)
|
227
|
+
# The log level to send messages to the `:logger` at.
|
228
|
+
#
|
229
|
+
# @option options [Logger] :logger
|
230
|
+
# The Logger instance to send log messages to. If this option
|
231
|
+
# is not set, logging will be disabled.
|
232
|
+
#
|
233
|
+
# @option options [Integer] :max_attempts (3)
|
234
|
+
# An integer representing the maximum number attempts that will be made for
|
235
|
+
# a single request, including the initial attempt. For example,
|
236
|
+
# setting this value to 5 will result in a request being retried up to
|
237
|
+
# 4 times. Used in `standard` and `adaptive` retry modes.
|
238
|
+
#
|
239
|
+
# @option options [String] :profile ("default")
|
240
|
+
# Used when loading credentials from the shared credentials file
|
241
|
+
# at HOME/.aws/credentials. When not specified, 'default' is used.
|
242
|
+
#
|
243
|
+
# @option options [Integer] :request_min_compression_size_bytes (10240)
|
244
|
+
# The minimum size in bytes that triggers compression for request
|
245
|
+
# bodies. The value must be non-negative integer value between 0
|
246
|
+
# and 10485780 bytes inclusive.
|
247
|
+
#
|
248
|
+
# @option options [Proc] :retry_backoff
|
249
|
+
# A proc or lambda used for backoff. Defaults to 2**retries * retry_base_delay.
|
250
|
+
# This option is only used in the `legacy` retry mode.
|
251
|
+
#
|
252
|
+
# @option options [Float] :retry_base_delay (0.3)
|
253
|
+
# The base delay in seconds used by the default backoff function. This option
|
254
|
+
# is only used in the `legacy` retry mode.
|
255
|
+
#
|
256
|
+
# @option options [Symbol] :retry_jitter (:none)
|
257
|
+
# A delay randomiser function used by the default backoff function.
|
258
|
+
# Some predefined functions can be referenced by name - :none, :equal, :full,
|
259
|
+
# otherwise a Proc that takes and returns a number. This option is only used
|
260
|
+
# in the `legacy` retry mode.
|
261
|
+
#
|
262
|
+
# @see https://www.awsarchitectureblog.com/2015/03/backoff.html
|
263
|
+
#
|
264
|
+
# @option options [Integer] :retry_limit (3)
|
265
|
+
# The maximum number of times to retry failed requests. Only
|
266
|
+
# ~ 500 level server errors and certain ~ 400 level client errors
|
267
|
+
# are retried. Generally, these are throttling errors, data
|
268
|
+
# checksum errors, networking errors, timeout errors, auth errors,
|
269
|
+
# endpoint discovery, and errors from expired credentials.
|
270
|
+
# This option is only used in the `legacy` retry mode.
|
271
|
+
#
|
272
|
+
# @option options [Integer] :retry_max_delay (0)
|
273
|
+
# The maximum number of seconds to delay between retries (0 for no limit)
|
274
|
+
# used by the default backoff function. This option is only used in the
|
275
|
+
# `legacy` retry mode.
|
276
|
+
#
|
277
|
+
# @option options [String] :retry_mode ("legacy")
|
278
|
+
# Specifies which retry algorithm to use. Values are:
|
279
|
+
#
|
280
|
+
# * `legacy` - The pre-existing retry behavior. This is default value if
|
281
|
+
# no retry mode is provided.
|
282
|
+
#
|
283
|
+
# * `standard` - A standardized set of retry rules across the AWS SDKs.
|
284
|
+
# This includes support for retry quotas, which limit the number of
|
285
|
+
# unsuccessful retries a client can make.
|
286
|
+
#
|
287
|
+
# * `adaptive` - An experimental retry mode that includes all the
|
288
|
+
# functionality of `standard` mode along with automatic client side
|
289
|
+
# throttling. This is a provisional mode that may change behavior
|
290
|
+
# in the future.
|
291
|
+
#
|
292
|
+
#
|
293
|
+
# @option options [String] :sdk_ua_app_id
|
294
|
+
# A unique and opaque application ID that is appended to the
|
295
|
+
# User-Agent header as app/<sdk_ua_app_id>. It should have a
|
296
|
+
# maximum length of 50.
|
297
|
+
#
|
298
|
+
# @option options [String] :secret_access_key
|
299
|
+
#
|
300
|
+
# @option options [String] :session_token
|
301
|
+
#
|
302
|
+
# @option options [Boolean] :stub_responses (false)
|
303
|
+
# Causes the client to return stubbed responses. By default
|
304
|
+
# fake responses are generated and returned. You can specify
|
305
|
+
# the response data to return or errors to raise by calling
|
306
|
+
# {ClientStubs#stub_responses}. See {ClientStubs} for more information.
|
307
|
+
#
|
308
|
+
# ** Please note ** When response stubbing is enabled, no HTTP
|
309
|
+
# requests are made, and retries are disabled.
|
310
|
+
#
|
311
|
+
# @option options [Aws::TokenProvider] :token_provider
|
312
|
+
# A Bearer Token Provider. This can be an instance of any one of the
|
313
|
+
# following classes:
|
314
|
+
#
|
315
|
+
# * `Aws::StaticTokenProvider` - Used for configuring static, non-refreshing
|
316
|
+
# tokens.
|
317
|
+
#
|
318
|
+
# * `Aws::SSOTokenProvider` - Used for loading tokens from AWS SSO using an
|
319
|
+
# access token generated from `aws login`.
|
320
|
+
#
|
321
|
+
# When `:token_provider` is not configured directly, the `Aws::TokenProviderChain`
|
322
|
+
# will be used to search for tokens configured for your profile in shared configuration files.
|
323
|
+
#
|
324
|
+
# @option options [Boolean] :use_dualstack_endpoint
|
325
|
+
# When set to `true`, dualstack enabled endpoints (with `.aws` TLD)
|
326
|
+
# will be used if available.
|
327
|
+
#
|
328
|
+
# @option options [Boolean] :use_fips_endpoint
|
329
|
+
# When set to `true`, fips compatible endpoints will be used if available.
|
330
|
+
# When a `fips` region is used, the region is normalized and this config
|
331
|
+
# is set to `true`.
|
332
|
+
#
|
333
|
+
# @option options [Boolean] :validate_params (true)
|
334
|
+
# When `true`, request parameters are validated before
|
335
|
+
# sending the request.
|
336
|
+
#
|
337
|
+
# @option options [Aws::Neptunedata::EndpointProvider] :endpoint_provider
|
338
|
+
# The endpoint provider used to resolve endpoints. Any object that responds to `#resolve_endpoint(parameters)` where `parameters` is a Struct similar to `Aws::Neptunedata::EndpointParameters`
|
339
|
+
#
|
340
|
+
# @option options [URI::HTTP,String] :http_proxy A proxy to send
|
341
|
+
# requests through. Formatted like 'http://proxy.com:123'.
|
342
|
+
#
|
343
|
+
# @option options [Float] :http_open_timeout (15) The number of
|
344
|
+
# seconds to wait when opening a HTTP session before raising a
|
345
|
+
# `Timeout::Error`.
|
346
|
+
#
|
347
|
+
# @option options [Float] :http_read_timeout (60) The default
|
348
|
+
# number of seconds to wait for response data. This value can
|
349
|
+
# safely be set per-request on the session.
|
350
|
+
#
|
351
|
+
# @option options [Float] :http_idle_timeout (5) The number of
|
352
|
+
# seconds a connection is allowed to sit idle before it is
|
353
|
+
# considered stale. Stale connections are closed and removed
|
354
|
+
# from the pool before making a request.
|
355
|
+
#
|
356
|
+
# @option options [Float] :http_continue_timeout (1) The number of
|
357
|
+
# seconds to wait for a 100-continue response before sending the
|
358
|
+
# request body. This option has no effect unless the request has
|
359
|
+
# "Expect" header set to "100-continue". Defaults to `nil` which
|
360
|
+
# disables this behaviour. This value can safely be set per
|
361
|
+
# request on the session.
|
362
|
+
#
|
363
|
+
# @option options [Float] :ssl_timeout (nil) Sets the SSL timeout
|
364
|
+
# in seconds.
|
365
|
+
#
|
366
|
+
# @option options [Boolean] :http_wire_trace (false) When `true`,
|
367
|
+
# HTTP debug output will be sent to the `:logger`.
|
368
|
+
#
|
369
|
+
# @option options [Boolean] :ssl_verify_peer (true) When `true`,
|
370
|
+
# SSL peer certificates are verified when establishing a
|
371
|
+
# connection.
|
372
|
+
#
|
373
|
+
# @option options [String] :ssl_ca_bundle Full path to the SSL
|
374
|
+
# certificate authority bundle file that should be used when
|
375
|
+
# verifying peer certificates. If you do not pass
|
376
|
+
# `:ssl_ca_bundle` or `:ssl_ca_directory` the the system default
|
377
|
+
# will be used if available.
|
378
|
+
#
|
379
|
+
# @option options [String] :ssl_ca_directory Full path of the
|
380
|
+
# directory that contains the unbundled SSL certificate
|
381
|
+
# authority files for verifying peer certificates. If you do
|
382
|
+
# not pass `:ssl_ca_bundle` or `:ssl_ca_directory` the the
|
383
|
+
# system default will be used if available.
|
384
|
+
#
|
385
|
+
def initialize(*args)
|
386
|
+
super
|
387
|
+
end
|
388
|
+
|
389
|
+
# @!group API Operations
|
390
|
+
|
391
|
+
# Cancels a Gremlin query. See [Gremlin query cancellation][1] for more
|
392
|
+
# information.
|
393
|
+
#
|
394
|
+
#
|
395
|
+
#
|
396
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/gremlin-api-status-cancel.html
|
397
|
+
#
|
398
|
+
# @option params [required, String] :query_id
|
399
|
+
# The unique identifier that identifies the query to be canceled.
|
400
|
+
#
|
401
|
+
# @return [Types::CancelGremlinQueryOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
402
|
+
#
|
403
|
+
# * {Types::CancelGremlinQueryOutput#status #status} => String
|
404
|
+
#
|
405
|
+
# @example Request syntax with placeholder values
|
406
|
+
#
|
407
|
+
# resp = client.cancel_gremlin_query({
|
408
|
+
# query_id: "String", # required
|
409
|
+
# })
|
410
|
+
#
|
411
|
+
# @example Response structure
|
412
|
+
#
|
413
|
+
# resp.status #=> String
|
414
|
+
#
|
415
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/CancelGremlinQuery AWS API Documentation
|
416
|
+
#
|
417
|
+
# @overload cancel_gremlin_query(params = {})
|
418
|
+
# @param [Hash] params ({})
|
419
|
+
def cancel_gremlin_query(params = {}, options = {})
|
420
|
+
req = build_request(:cancel_gremlin_query, params)
|
421
|
+
req.send_request(options)
|
422
|
+
end
|
423
|
+
|
424
|
+
# Cancels a specified load job. This is an HTTP `DELETE` request.
|
425
|
+
#
|
426
|
+
# See [Neptune Loader Get-Status API][1] for more information.
|
427
|
+
#
|
428
|
+
#
|
429
|
+
#
|
430
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/load-api-reference-status.htm
|
431
|
+
#
|
432
|
+
# @option params [required, String] :load_id
|
433
|
+
# The ID of the load job to be deleted.
|
434
|
+
#
|
435
|
+
# @return [Types::CancelLoaderJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
436
|
+
#
|
437
|
+
# * {Types::CancelLoaderJobOutput#status #status} => String
|
438
|
+
#
|
439
|
+
# @example Request syntax with placeholder values
|
440
|
+
#
|
441
|
+
# resp = client.cancel_loader_job({
|
442
|
+
# load_id: "String", # required
|
443
|
+
# })
|
444
|
+
#
|
445
|
+
# @example Response structure
|
446
|
+
#
|
447
|
+
# resp.status #=> String
|
448
|
+
#
|
449
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/CancelLoaderJob AWS API Documentation
|
450
|
+
#
|
451
|
+
# @overload cancel_loader_job(params = {})
|
452
|
+
# @param [Hash] params ({})
|
453
|
+
def cancel_loader_job(params = {}, options = {})
|
454
|
+
req = build_request(:cancel_loader_job, params)
|
455
|
+
req.send_request(options)
|
456
|
+
end
|
457
|
+
|
458
|
+
# Cancels a Neptune ML data processing job. See [The `dataprocessing`
|
459
|
+
# command][1].
|
460
|
+
#
|
461
|
+
#
|
462
|
+
#
|
463
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-dataprocessing.html
|
464
|
+
#
|
465
|
+
# @option params [required, String] :id
|
466
|
+
# The unique identifier of the data-processing job.
|
467
|
+
#
|
468
|
+
# @option params [String] :neptune_iam_role_arn
|
469
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
470
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
471
|
+
# group or an error will occur.
|
472
|
+
#
|
473
|
+
# @option params [Boolean] :clean
|
474
|
+
# If set to `TRUE`, this flag specifies that all Neptune ML S3 artifacts
|
475
|
+
# should be deleted when the job is stopped. The default is `FALSE`.
|
476
|
+
#
|
477
|
+
# @return [Types::CancelMLDataProcessingJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
478
|
+
#
|
479
|
+
# * {Types::CancelMLDataProcessingJobOutput#status #status} => String
|
480
|
+
#
|
481
|
+
# @example Request syntax with placeholder values
|
482
|
+
#
|
483
|
+
# resp = client.cancel_ml_data_processing_job({
|
484
|
+
# id: "String", # required
|
485
|
+
# neptune_iam_role_arn: "String",
|
486
|
+
# clean: false,
|
487
|
+
# })
|
488
|
+
#
|
489
|
+
# @example Response structure
|
490
|
+
#
|
491
|
+
# resp.status #=> String
|
492
|
+
#
|
493
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/CancelMLDataProcessingJob AWS API Documentation
|
494
|
+
#
|
495
|
+
# @overload cancel_ml_data_processing_job(params = {})
|
496
|
+
# @param [Hash] params ({})
|
497
|
+
def cancel_ml_data_processing_job(params = {}, options = {})
|
498
|
+
req = build_request(:cancel_ml_data_processing_job, params)
|
499
|
+
req.send_request(options)
|
500
|
+
end
|
501
|
+
|
502
|
+
# Cancels a Neptune ML model training job. See [Model training using the
|
503
|
+
# `modeltraining` command][1].
|
504
|
+
#
|
505
|
+
#
|
506
|
+
#
|
507
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-modeltraining.html
|
508
|
+
#
|
509
|
+
# @option params [required, String] :id
|
510
|
+
# The unique identifier of the model-training job to be canceled.
|
511
|
+
#
|
512
|
+
# @option params [String] :neptune_iam_role_arn
|
513
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
514
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
515
|
+
# group or an error will occur.
|
516
|
+
#
|
517
|
+
# @option params [Boolean] :clean
|
518
|
+
# If set to `TRUE`, this flag specifies that all Amazon S3 artifacts
|
519
|
+
# should be deleted when the job is stopped. The default is `FALSE`.
|
520
|
+
#
|
521
|
+
# @return [Types::CancelMLModelTrainingJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
522
|
+
#
|
523
|
+
# * {Types::CancelMLModelTrainingJobOutput#status #status} => String
|
524
|
+
#
|
525
|
+
# @example Request syntax with placeholder values
|
526
|
+
#
|
527
|
+
# resp = client.cancel_ml_model_training_job({
|
528
|
+
# id: "String", # required
|
529
|
+
# neptune_iam_role_arn: "String",
|
530
|
+
# clean: false,
|
531
|
+
# })
|
532
|
+
#
|
533
|
+
# @example Response structure
|
534
|
+
#
|
535
|
+
# resp.status #=> String
|
536
|
+
#
|
537
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/CancelMLModelTrainingJob AWS API Documentation
|
538
|
+
#
|
539
|
+
# @overload cancel_ml_model_training_job(params = {})
|
540
|
+
# @param [Hash] params ({})
|
541
|
+
def cancel_ml_model_training_job(params = {}, options = {})
|
542
|
+
req = build_request(:cancel_ml_model_training_job, params)
|
543
|
+
req.send_request(options)
|
544
|
+
end
|
545
|
+
|
546
|
+
# Cancels a specified model transform job. See [Use a trained model to
|
547
|
+
# generate new model artifacts][1].
|
548
|
+
#
|
549
|
+
#
|
550
|
+
#
|
551
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-model-transform.html
|
552
|
+
#
|
553
|
+
# @option params [required, String] :id
|
554
|
+
# The unique ID of the model transform job to be canceled.
|
555
|
+
#
|
556
|
+
# @option params [String] :neptune_iam_role_arn
|
557
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
558
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
559
|
+
# group or an error will occur.
|
560
|
+
#
|
561
|
+
# @option params [Boolean] :clean
|
562
|
+
# If this flag is set to `TRUE`, all Neptune ML S3 artifacts should be
|
563
|
+
# deleted when the job is stopped. The default is `FALSE`.
|
564
|
+
#
|
565
|
+
# @return [Types::CancelMLModelTransformJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
566
|
+
#
|
567
|
+
# * {Types::CancelMLModelTransformJobOutput#status #status} => String
|
568
|
+
#
|
569
|
+
# @example Request syntax with placeholder values
|
570
|
+
#
|
571
|
+
# resp = client.cancel_ml_model_transform_job({
|
572
|
+
# id: "String", # required
|
573
|
+
# neptune_iam_role_arn: "String",
|
574
|
+
# clean: false,
|
575
|
+
# })
|
576
|
+
#
|
577
|
+
# @example Response structure
|
578
|
+
#
|
579
|
+
# resp.status #=> String
|
580
|
+
#
|
581
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/CancelMLModelTransformJob AWS API Documentation
|
582
|
+
#
|
583
|
+
# @overload cancel_ml_model_transform_job(params = {})
|
584
|
+
# @param [Hash] params ({})
|
585
|
+
def cancel_ml_model_transform_job(params = {}, options = {})
|
586
|
+
req = build_request(:cancel_ml_model_transform_job, params)
|
587
|
+
req.send_request(options)
|
588
|
+
end
|
589
|
+
|
590
|
+
# Cancels a specified openCypher query. See [Neptune openCypher status
|
591
|
+
# endpoint][1] for more information.
|
592
|
+
#
|
593
|
+
#
|
594
|
+
#
|
595
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/access-graph-opencypher-status.html
|
596
|
+
#
|
597
|
+
# @option params [required, String] :query_id
|
598
|
+
# The unique ID of the openCypher query to cancel.
|
599
|
+
#
|
600
|
+
# @option params [Boolean] :silent
|
601
|
+
# If set to `TRUE`, causes the cancelation of the openCypher query to
|
602
|
+
# happen silently.
|
603
|
+
#
|
604
|
+
# @return [Types::CancelOpenCypherQueryOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
605
|
+
#
|
606
|
+
# * {Types::CancelOpenCypherQueryOutput#status #status} => String
|
607
|
+
# * {Types::CancelOpenCypherQueryOutput#payload #payload} => Boolean
|
608
|
+
#
|
609
|
+
# @example Request syntax with placeholder values
|
610
|
+
#
|
611
|
+
# resp = client.cancel_open_cypher_query({
|
612
|
+
# query_id: "String", # required
|
613
|
+
# silent: false,
|
614
|
+
# })
|
615
|
+
#
|
616
|
+
# @example Response structure
|
617
|
+
#
|
618
|
+
# resp.status #=> String
|
619
|
+
# resp.payload #=> Boolean
|
620
|
+
#
|
621
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/CancelOpenCypherQuery AWS API Documentation
|
622
|
+
#
|
623
|
+
# @overload cancel_open_cypher_query(params = {})
|
624
|
+
# @param [Hash] params ({})
|
625
|
+
def cancel_open_cypher_query(params = {}, options = {})
|
626
|
+
req = build_request(:cancel_open_cypher_query, params)
|
627
|
+
req.send_request(options)
|
628
|
+
end
|
629
|
+
|
630
|
+
# Creates a new Neptune ML inference endpoint that lets you query one
|
631
|
+
# specific model that the model-training process constructed. See
|
632
|
+
# [Managing inference endpoints using the endpoints command][1].
|
633
|
+
#
|
634
|
+
#
|
635
|
+
#
|
636
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-endpoints.html
|
637
|
+
#
|
638
|
+
# @option params [String] :id
|
639
|
+
# A unique identifier for the new inference endpoint. The default is an
|
640
|
+
# autogenerated timestamped name.
|
641
|
+
#
|
642
|
+
# @option params [String] :ml_model_training_job_id
|
643
|
+
# The job Id of the completed model-training job that has created the
|
644
|
+
# model that the inference endpoint will point to. You must supply
|
645
|
+
# either the `mlModelTrainingJobId` or the `mlModelTransformJobId`.
|
646
|
+
#
|
647
|
+
# @option params [String] :ml_model_transform_job_id
|
648
|
+
# The job Id of the completed model-transform job. You must supply
|
649
|
+
# either the `mlModelTrainingJobId` or the `mlModelTransformJobId`.
|
650
|
+
#
|
651
|
+
# @option params [Boolean] :update
|
652
|
+
# If set to `true`, `update` indicates that this is an update request.
|
653
|
+
# The default is `false`. You must supply either the
|
654
|
+
# `mlModelTrainingJobId` or the `mlModelTransformJobId`.
|
655
|
+
#
|
656
|
+
# @option params [String] :neptune_iam_role_arn
|
657
|
+
# The ARN of an IAM role providing Neptune access to SageMaker and
|
658
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
659
|
+
# group or an error will be thrown.
|
660
|
+
#
|
661
|
+
# @option params [String] :model_name
|
662
|
+
# Model type for training. By default the Neptune ML model is
|
663
|
+
# automatically based on the `modelType` used in data processing, but
|
664
|
+
# you can specify a different model type here. The default is `rgcn` for
|
665
|
+
# heterogeneous graphs and `kge` for knowledge graphs. The only valid
|
666
|
+
# value for heterogeneous graphs is `rgcn`. Valid values for knowledge
|
667
|
+
# graphs are: `kge`, `transe`, `distmult`, and `rotate`.
|
668
|
+
#
|
669
|
+
# @option params [String] :instance_type
|
670
|
+
# The type of Neptune ML instance to use for online servicing. The
|
671
|
+
# default is `ml.m5.xlarge`. Choosing the ML instance for an inference
|
672
|
+
# endpoint depends on the task type, the graph size, and your budget.
|
673
|
+
#
|
674
|
+
# @option params [Integer] :instance_count
|
675
|
+
# The minimum number of Amazon EC2 instances to deploy to an endpoint
|
676
|
+
# for prediction. The default is 1
|
677
|
+
#
|
678
|
+
# @option params [String] :volume_encryption_kms_key
|
679
|
+
# The Amazon Key Management Service (Amazon KMS) key that SageMaker uses
|
680
|
+
# to encrypt data on the storage volume attached to the ML compute
|
681
|
+
# instances that run the training job. The default is None.
|
682
|
+
#
|
683
|
+
# @return [Types::CreateMLEndpointOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
684
|
+
#
|
685
|
+
# * {Types::CreateMLEndpointOutput#id #id} => String
|
686
|
+
# * {Types::CreateMLEndpointOutput#arn #arn} => String
|
687
|
+
# * {Types::CreateMLEndpointOutput#creation_time_in_millis #creation_time_in_millis} => Integer
|
688
|
+
#
|
689
|
+
# @example Request syntax with placeholder values
|
690
|
+
#
|
691
|
+
# resp = client.create_ml_endpoint({
|
692
|
+
# id: "String",
|
693
|
+
# ml_model_training_job_id: "String",
|
694
|
+
# ml_model_transform_job_id: "String",
|
695
|
+
# update: false,
|
696
|
+
# neptune_iam_role_arn: "String",
|
697
|
+
# model_name: "String",
|
698
|
+
# instance_type: "String",
|
699
|
+
# instance_count: 1,
|
700
|
+
# volume_encryption_kms_key: "String",
|
701
|
+
# })
|
702
|
+
#
|
703
|
+
# @example Response structure
|
704
|
+
#
|
705
|
+
# resp.id #=> String
|
706
|
+
# resp.arn #=> String
|
707
|
+
# resp.creation_time_in_millis #=> Integer
|
708
|
+
#
|
709
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/CreateMLEndpoint AWS API Documentation
|
710
|
+
#
|
711
|
+
# @overload create_ml_endpoint(params = {})
|
712
|
+
# @param [Hash] params ({})
|
713
|
+
def create_ml_endpoint(params = {}, options = {})
|
714
|
+
req = build_request(:create_ml_endpoint, params)
|
715
|
+
req.send_request(options)
|
716
|
+
end
|
717
|
+
|
718
|
+
# Cancels the creation of a Neptune ML inference endpoint. See [Managing
|
719
|
+
# inference endpoints using the endpoints command][1].
|
720
|
+
#
|
721
|
+
#
|
722
|
+
#
|
723
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-endpoints.html
|
724
|
+
#
|
725
|
+
# @option params [required, String] :id
|
726
|
+
# The unique identifier of the inference endpoint.
|
727
|
+
#
|
728
|
+
# @option params [String] :neptune_iam_role_arn
|
729
|
+
# The ARN of an IAM role providing Neptune access to SageMaker and
|
730
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
731
|
+
# group or an error will be thrown.
|
732
|
+
#
|
733
|
+
# @option params [Boolean] :clean
|
734
|
+
# If this flag is set to `TRUE`, all Neptune ML S3 artifacts should be
|
735
|
+
# deleted when the job is stopped. The default is `FALSE`.
|
736
|
+
#
|
737
|
+
# @return [Types::DeleteMLEndpointOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
738
|
+
#
|
739
|
+
# * {Types::DeleteMLEndpointOutput#status #status} => String
|
740
|
+
#
|
741
|
+
# @example Request syntax with placeholder values
|
742
|
+
#
|
743
|
+
# resp = client.delete_ml_endpoint({
|
744
|
+
# id: "String", # required
|
745
|
+
# neptune_iam_role_arn: "String",
|
746
|
+
# clean: false,
|
747
|
+
# })
|
748
|
+
#
|
749
|
+
# @example Response structure
|
750
|
+
#
|
751
|
+
# resp.status #=> String
|
752
|
+
#
|
753
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/DeleteMLEndpoint AWS API Documentation
|
754
|
+
#
|
755
|
+
# @overload delete_ml_endpoint(params = {})
|
756
|
+
# @param [Hash] params ({})
|
757
|
+
def delete_ml_endpoint(params = {}, options = {})
|
758
|
+
req = build_request(:delete_ml_endpoint, params)
|
759
|
+
req.send_request(options)
|
760
|
+
end
|
761
|
+
|
762
|
+
# Deletes statistics for Gremlin and openCypher (property graph) data.
|
763
|
+
#
|
764
|
+
# @return [Types::DeletePropertygraphStatisticsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
765
|
+
#
|
766
|
+
# * {Types::DeletePropertygraphStatisticsOutput#status_code #status_code} => Integer
|
767
|
+
# * {Types::DeletePropertygraphStatisticsOutput#status #status} => String
|
768
|
+
# * {Types::DeletePropertygraphStatisticsOutput#payload #payload} => Types::DeleteStatisticsValueMap
|
769
|
+
#
|
770
|
+
# @example Response structure
|
771
|
+
#
|
772
|
+
# resp.status_code #=> Integer
|
773
|
+
# resp.status #=> String
|
774
|
+
# resp.payload.active #=> Boolean
|
775
|
+
# resp.payload.statistics_id #=> String
|
776
|
+
#
|
777
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/DeletePropertygraphStatistics AWS API Documentation
|
778
|
+
#
|
779
|
+
# @overload delete_propertygraph_statistics(params = {})
|
780
|
+
# @param [Hash] params ({})
|
781
|
+
def delete_propertygraph_statistics(params = {}, options = {})
|
782
|
+
req = build_request(:delete_propertygraph_statistics, params)
|
783
|
+
req.send_request(options)
|
784
|
+
end
|
785
|
+
|
786
|
+
# Deletes SPARQL statistics
|
787
|
+
#
|
788
|
+
# @return [Types::DeleteSparqlStatisticsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
789
|
+
#
|
790
|
+
# * {Types::DeleteSparqlStatisticsOutput#status_code #status_code} => Integer
|
791
|
+
# * {Types::DeleteSparqlStatisticsOutput#status #status} => String
|
792
|
+
# * {Types::DeleteSparqlStatisticsOutput#payload #payload} => Types::DeleteStatisticsValueMap
|
793
|
+
#
|
794
|
+
# @example Response structure
|
795
|
+
#
|
796
|
+
# resp.status_code #=> Integer
|
797
|
+
# resp.status #=> String
|
798
|
+
# resp.payload.active #=> Boolean
|
799
|
+
# resp.payload.statistics_id #=> String
|
800
|
+
#
|
801
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/DeleteSparqlStatistics AWS API Documentation
|
802
|
+
#
|
803
|
+
# @overload delete_sparql_statistics(params = {})
|
804
|
+
# @param [Hash] params ({})
|
805
|
+
def delete_sparql_statistics(params = {}, options = {})
|
806
|
+
req = build_request(:delete_sparql_statistics, params)
|
807
|
+
req.send_request(options)
|
808
|
+
end
|
809
|
+
|
810
|
+
# The fast reset REST API lets you reset a Neptune graph quicky and
|
811
|
+
# easily, removing all of its data.
|
812
|
+
#
|
813
|
+
# Neptune fast reset is a two-step process. First you call
|
814
|
+
# `ExecuteFastReset` with `action` set to `initiateDatabaseReset`. This
|
815
|
+
# returns a UUID token which you then include when calling
|
816
|
+
# `ExecuteFastReset` again with `action` set to `performDatabaseReset`.
|
817
|
+
# See [Empty an Amazon Neptune DB cluster using the fast reset API][1].
|
818
|
+
#
|
819
|
+
#
|
820
|
+
#
|
821
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/manage-console-fast-reset.html
|
822
|
+
#
|
823
|
+
# @option params [required, String] :action
|
824
|
+
# The fast reset action. One of the following values:
|
825
|
+
#
|
826
|
+
# * <b> <code>initiateDatabaseReset</code> </b> – This action
|
827
|
+
# generates a unique token needed to actually perform the fast reset.
|
828
|
+
#
|
829
|
+
# * <b> <code>performDatabaseReset</code> </b> – This action uses
|
830
|
+
# the token generated by the `initiateDatabaseReset` action to
|
831
|
+
# actually perform the fast reset.
|
832
|
+
#
|
833
|
+
# @option params [String] :token
|
834
|
+
# The fast-reset token to initiate the reset.
|
835
|
+
#
|
836
|
+
# @return [Types::ExecuteFastResetOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
837
|
+
#
|
838
|
+
# * {Types::ExecuteFastResetOutput#status #status} => String
|
839
|
+
# * {Types::ExecuteFastResetOutput#payload #payload} => Types::FastResetToken
|
840
|
+
#
|
841
|
+
# @example Request syntax with placeholder values
|
842
|
+
#
|
843
|
+
# resp = client.execute_fast_reset({
|
844
|
+
# action: "initiateDatabaseReset", # required, accepts initiateDatabaseReset, performDatabaseReset
|
845
|
+
# token: "String",
|
846
|
+
# })
|
847
|
+
#
|
848
|
+
# @example Response structure
|
849
|
+
#
|
850
|
+
# resp.status #=> String
|
851
|
+
# resp.payload.token #=> String
|
852
|
+
#
|
853
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ExecuteFastReset AWS API Documentation
|
854
|
+
#
|
855
|
+
# @overload execute_fast_reset(params = {})
|
856
|
+
# @param [Hash] params ({})
|
857
|
+
def execute_fast_reset(params = {}, options = {})
|
858
|
+
req = build_request(:execute_fast_reset, params)
|
859
|
+
req.send_request(options)
|
860
|
+
end
|
861
|
+
|
862
|
+
# Executes a Gremlin Explain query.
|
863
|
+
#
|
864
|
+
# Amazon Neptune has added a Gremlin feature named `explain` that
|
865
|
+
# provides is a self-service tool for understanding the execution
|
866
|
+
# approach being taken by the Neptune engine for the query. You invoke
|
867
|
+
# it by adding an `explain` parameter to an HTTP call that submits a
|
868
|
+
# Gremlin query.
|
869
|
+
#
|
870
|
+
# The explain feature provides information about the logical structure
|
871
|
+
# of query execution plans. You can use this information to identify
|
872
|
+
# potential evaluation and execution bottlenecks and to tune your query,
|
873
|
+
# as explained in [Tuning Gremlin queries][1]. You can also use query
|
874
|
+
# hints to improve query execution plans.
|
875
|
+
#
|
876
|
+
#
|
877
|
+
#
|
878
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/gremlin-traversal-tuning.html
|
879
|
+
#
|
880
|
+
# @option params [required, String] :gremlin_query
|
881
|
+
# The Gremlin explain query string.
|
882
|
+
#
|
883
|
+
# @return [Types::ExecuteGremlinExplainQueryOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
884
|
+
#
|
885
|
+
# * {Types::ExecuteGremlinExplainQueryOutput#output #output} => String
|
886
|
+
#
|
887
|
+
# @example Request syntax with placeholder values
|
888
|
+
#
|
889
|
+
# resp = client.execute_gremlin_explain_query({
|
890
|
+
# gremlin_query: "String", # required
|
891
|
+
# })
|
892
|
+
#
|
893
|
+
# @example Response structure
|
894
|
+
#
|
895
|
+
# resp.output #=> String
|
896
|
+
#
|
897
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ExecuteGremlinExplainQuery AWS API Documentation
|
898
|
+
#
|
899
|
+
# @overload execute_gremlin_explain_query(params = {})
|
900
|
+
# @param [Hash] params ({})
|
901
|
+
def execute_gremlin_explain_query(params = {}, options = {})
|
902
|
+
req = build_request(:execute_gremlin_explain_query, params)
|
903
|
+
req.send_request(options)
|
904
|
+
end
|
905
|
+
|
906
|
+
# Executes a Gremlin Profile query, which runs a specified traversal,
|
907
|
+
# collects various metrics about the run, and produces a profile report
|
908
|
+
# as output. See [Gremlin profile API in Neptune][1] for details.
|
909
|
+
#
|
910
|
+
#
|
911
|
+
#
|
912
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/gremlin-profile-api.html
|
913
|
+
#
|
914
|
+
# @option params [required, String] :gremlin_query
|
915
|
+
# The Gremlin query string to profile.
|
916
|
+
#
|
917
|
+
# @option params [Boolean] :results
|
918
|
+
# If this flag is set to `TRUE`, the query results are gathered and
|
919
|
+
# displayed as part of the profile report. If `FALSE`, only the result
|
920
|
+
# count is displayed.
|
921
|
+
#
|
922
|
+
# @option params [Integer] :chop
|
923
|
+
# If non-zero, causes the results string to be truncated at that number
|
924
|
+
# of characters. If set to zero, the string contains all the results.
|
925
|
+
#
|
926
|
+
# @option params [String] :serializer
|
927
|
+
# If non-null, the gathered results are returned in a serialized
|
928
|
+
# response message in the format specified by this parameter. See
|
929
|
+
# [Gremlin profile API in Neptune][1] for more information.
|
930
|
+
#
|
931
|
+
#
|
932
|
+
#
|
933
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/gremlin-profile-api.html
|
934
|
+
#
|
935
|
+
# @option params [Boolean] :index_ops
|
936
|
+
# If this flag is set to `TRUE`, the results include a detailed report
|
937
|
+
# of all index operations that took place during query execution and
|
938
|
+
# serialization.
|
939
|
+
#
|
940
|
+
# @return [Types::ExecuteGremlinProfileQueryOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
941
|
+
#
|
942
|
+
# * {Types::ExecuteGremlinProfileQueryOutput#output #output} => String
|
943
|
+
#
|
944
|
+
# @example Request syntax with placeholder values
|
945
|
+
#
|
946
|
+
# resp = client.execute_gremlin_profile_query({
|
947
|
+
# gremlin_query: "String", # required
|
948
|
+
# results: false,
|
949
|
+
# chop: 1,
|
950
|
+
# serializer: "String",
|
951
|
+
# index_ops: false,
|
952
|
+
# })
|
953
|
+
#
|
954
|
+
# @example Response structure
|
955
|
+
#
|
956
|
+
# resp.output #=> String
|
957
|
+
#
|
958
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ExecuteGremlinProfileQuery AWS API Documentation
|
959
|
+
#
|
960
|
+
# @overload execute_gremlin_profile_query(params = {})
|
961
|
+
# @param [Hash] params ({})
|
962
|
+
def execute_gremlin_profile_query(params = {}, options = {})
|
963
|
+
req = build_request(:execute_gremlin_profile_query, params)
|
964
|
+
req.send_request(options)
|
965
|
+
end
|
966
|
+
|
967
|
+
# This commands executes a Gremlin query. Amazon Neptune is compatible
|
968
|
+
# with Apache TinkerPop3 and Gremlin, so you can use the Gremlin
|
969
|
+
# traversal language to query the graph, as described under [The
|
970
|
+
# Graph][1] in the Apache TinkerPop3 documentation. More details can
|
971
|
+
# also be found in [Accessing a Neptune graph with Gremlin][2].
|
972
|
+
#
|
973
|
+
#
|
974
|
+
#
|
975
|
+
# [1]: https://tinkerpop.apache.org/docs/current/reference/#graph
|
976
|
+
# [2]: https://docs.aws.amazon.com/neptune/latest/userguide/access-graph-gremlin.html
|
977
|
+
#
|
978
|
+
# @option params [required, String] :gremlin_query
|
979
|
+
# Using this API, you can run Gremlin queries in string format much as
|
980
|
+
# you can using the HTTP endpoint. The interface is compatible with
|
981
|
+
# whatever Gremlin version your DB cluster is using (see the [Tinkerpop
|
982
|
+
# client section][1] to determine which Gremlin releases your engine
|
983
|
+
# version supports).
|
984
|
+
#
|
985
|
+
#
|
986
|
+
#
|
987
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/access-graph-gremlin-client.html#best-practices-gremlin-java-latest
|
988
|
+
#
|
989
|
+
# @option params [String] :serializer
|
990
|
+
# If non-null, the query results are returned in a serialized response
|
991
|
+
# message in the format specified by this parameter. See the
|
992
|
+
# [GraphSON][1] section in the TinkerPop documentation for a list of the
|
993
|
+
# formats that are currently supported.
|
994
|
+
#
|
995
|
+
#
|
996
|
+
#
|
997
|
+
# [1]: https://tinkerpop.apache.org/docs/current/reference/#_graphson
|
998
|
+
#
|
999
|
+
# @return [Types::ExecuteGremlinQueryOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1000
|
+
#
|
1001
|
+
# * {Types::ExecuteGremlinQueryOutput#request_id #request_id} => String
|
1002
|
+
# * {Types::ExecuteGremlinQueryOutput#status #status} => Types::GremlinQueryStatusAttributes
|
1003
|
+
# * {Types::ExecuteGremlinQueryOutput#result #result} => Hash,Array,String,Numeric,Boolean
|
1004
|
+
# * {Types::ExecuteGremlinQueryOutput#meta #meta} => Hash,Array,String,Numeric,Boolean
|
1005
|
+
#
|
1006
|
+
# @example Request syntax with placeholder values
|
1007
|
+
#
|
1008
|
+
# resp = client.execute_gremlin_query({
|
1009
|
+
# gremlin_query: "String", # required
|
1010
|
+
# serializer: "String",
|
1011
|
+
# })
|
1012
|
+
#
|
1013
|
+
# @example Response structure
|
1014
|
+
#
|
1015
|
+
# resp.request_id #=> String
|
1016
|
+
# resp.status.message #=> String
|
1017
|
+
# resp.status.code #=> Integer
|
1018
|
+
#
|
1019
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ExecuteGremlinQuery AWS API Documentation
|
1020
|
+
#
|
1021
|
+
# @overload execute_gremlin_query(params = {})
|
1022
|
+
# @param [Hash] params ({})
|
1023
|
+
def execute_gremlin_query(params = {}, options = {})
|
1024
|
+
req = build_request(:execute_gremlin_query, params)
|
1025
|
+
req.send_request(options)
|
1026
|
+
end
|
1027
|
+
|
1028
|
+
# Executes an openCypher `explain` request. See [The openCypher explain
|
1029
|
+
# feature][1] for more information.
|
1030
|
+
#
|
1031
|
+
#
|
1032
|
+
#
|
1033
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/access-graph-opencypher-explain.html
|
1034
|
+
#
|
1035
|
+
# @option params [required, String] :open_cypher_query
|
1036
|
+
# The openCypher query string.
|
1037
|
+
#
|
1038
|
+
# @option params [String] :parameters
|
1039
|
+
# The openCypher query parameters.
|
1040
|
+
#
|
1041
|
+
# @option params [required, String] :explain_mode
|
1042
|
+
# The openCypher `explain` mode. Can be one of: `static`, `dynamic`, or
|
1043
|
+
# `details`.
|
1044
|
+
#
|
1045
|
+
# @return [Types::ExecuteOpenCypherExplainQueryOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1046
|
+
#
|
1047
|
+
# * {Types::ExecuteOpenCypherExplainQueryOutput#results #results} => String
|
1048
|
+
#
|
1049
|
+
# @example Request syntax with placeholder values
|
1050
|
+
#
|
1051
|
+
# resp = client.execute_open_cypher_explain_query({
|
1052
|
+
# open_cypher_query: "String", # required
|
1053
|
+
# parameters: "String",
|
1054
|
+
# explain_mode: "static", # required, accepts static, dynamic, details
|
1055
|
+
# })
|
1056
|
+
#
|
1057
|
+
# @example Response structure
|
1058
|
+
#
|
1059
|
+
# resp.results #=> String
|
1060
|
+
#
|
1061
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ExecuteOpenCypherExplainQuery AWS API Documentation
|
1062
|
+
#
|
1063
|
+
# @overload execute_open_cypher_explain_query(params = {})
|
1064
|
+
# @param [Hash] params ({})
|
1065
|
+
def execute_open_cypher_explain_query(params = {}, options = {})
|
1066
|
+
req = build_request(:execute_open_cypher_explain_query, params)
|
1067
|
+
req.send_request(options)
|
1068
|
+
end
|
1069
|
+
|
1070
|
+
# Executes an openCypher query. See [Accessing the Neptune Graph with
|
1071
|
+
# openCypher][1] for more information.
|
1072
|
+
#
|
1073
|
+
# Neptune supports building graph applications using openCypher, which
|
1074
|
+
# is currently one of the most popular query languages among developers
|
1075
|
+
# working with graph databases. Developers, business analysts, and data
|
1076
|
+
# scientists like openCypher's declarative, SQL-inspired syntax because
|
1077
|
+
# it provides a familiar structure in which to querying property graphs.
|
1078
|
+
#
|
1079
|
+
# The openCypher language was originally developed by Neo4j, then
|
1080
|
+
# open-sourced in 2015 and contributed to the [openCypher project][2]
|
1081
|
+
# under an Apache 2 open-source license.
|
1082
|
+
#
|
1083
|
+
#
|
1084
|
+
#
|
1085
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/access-graph-opencypher.html
|
1086
|
+
# [2]: https://opencypher.org/
|
1087
|
+
#
|
1088
|
+
# @option params [required, String] :open_cypher_query
|
1089
|
+
# The openCypher query string to be executed.
|
1090
|
+
#
|
1091
|
+
# @option params [String] :parameters
|
1092
|
+
# The openCypher query parameters for query execution. See [Examples of
|
1093
|
+
# openCypher parameterized queries][1] for more information.
|
1094
|
+
#
|
1095
|
+
#
|
1096
|
+
#
|
1097
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/opencypher-parameterized-queries.html
|
1098
|
+
#
|
1099
|
+
# @return [Types::ExecuteOpenCypherQueryOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1100
|
+
#
|
1101
|
+
# * {Types::ExecuteOpenCypherQueryOutput#results #results} => Hash,Array,String,Numeric,Boolean
|
1102
|
+
#
|
1103
|
+
# @example Request syntax with placeholder values
|
1104
|
+
#
|
1105
|
+
# resp = client.execute_open_cypher_query({
|
1106
|
+
# open_cypher_query: "String", # required
|
1107
|
+
# parameters: "String",
|
1108
|
+
# })
|
1109
|
+
#
|
1110
|
+
# @example Response structure
|
1111
|
+
#
|
1112
|
+
#
|
1113
|
+
#
|
1114
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ExecuteOpenCypherQuery AWS API Documentation
|
1115
|
+
#
|
1116
|
+
# @overload execute_open_cypher_query(params = {})
|
1117
|
+
# @param [Hash] params ({})
|
1118
|
+
def execute_open_cypher_query(params = {}, options = {})
|
1119
|
+
req = build_request(:execute_open_cypher_query, params)
|
1120
|
+
req.send_request(options)
|
1121
|
+
end
|
1122
|
+
|
1123
|
+
# Check the status of the graph database on the host.
|
1124
|
+
#
|
1125
|
+
# @return [Types::GetEngineStatusOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1126
|
+
#
|
1127
|
+
# * {Types::GetEngineStatusOutput#status #status} => String
|
1128
|
+
# * {Types::GetEngineStatusOutput#start_time #start_time} => String
|
1129
|
+
# * {Types::GetEngineStatusOutput#db_engine_version #db_engine_version} => String
|
1130
|
+
# * {Types::GetEngineStatusOutput#role #role} => String
|
1131
|
+
# * {Types::GetEngineStatusOutput#dfe_query_engine #dfe_query_engine} => String
|
1132
|
+
# * {Types::GetEngineStatusOutput#gremlin #gremlin} => Types::QueryLanguageVersion
|
1133
|
+
# * {Types::GetEngineStatusOutput#sparql #sparql} => Types::QueryLanguageVersion
|
1134
|
+
# * {Types::GetEngineStatusOutput#opencypher #opencypher} => Types::QueryLanguageVersion
|
1135
|
+
# * {Types::GetEngineStatusOutput#lab_mode #lab_mode} => Hash<String,String>
|
1136
|
+
# * {Types::GetEngineStatusOutput#rolling_back_trx_count #rolling_back_trx_count} => Integer
|
1137
|
+
# * {Types::GetEngineStatusOutput#rolling_back_trx_earliest_start_time #rolling_back_trx_earliest_start_time} => String
|
1138
|
+
# * {Types::GetEngineStatusOutput#features #features} => Hash<String,Hash,Array,String,Numeric,Boolean>
|
1139
|
+
# * {Types::GetEngineStatusOutput#settings #settings} => Hash<String,String>
|
1140
|
+
#
|
1141
|
+
# @example Response structure
|
1142
|
+
#
|
1143
|
+
# resp.status #=> String
|
1144
|
+
# resp.start_time #=> String
|
1145
|
+
# resp.db_engine_version #=> String
|
1146
|
+
# resp.role #=> String
|
1147
|
+
# resp.dfe_query_engine #=> String
|
1148
|
+
# resp.gremlin.version #=> String
|
1149
|
+
# resp.sparql.version #=> String
|
1150
|
+
# resp.opencypher.version #=> String
|
1151
|
+
# resp.lab_mode #=> Hash
|
1152
|
+
# resp.lab_mode["String"] #=> String
|
1153
|
+
# resp.rolling_back_trx_count #=> Integer
|
1154
|
+
# resp.rolling_back_trx_earliest_start_time #=> String
|
1155
|
+
# resp.features #=> Hash
|
1156
|
+
# resp.settings #=> Hash
|
1157
|
+
# resp.settings["String"] #=> String
|
1158
|
+
#
|
1159
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetEngineStatus AWS API Documentation
|
1160
|
+
#
|
1161
|
+
# @overload get_engine_status(params = {})
|
1162
|
+
# @param [Hash] params ({})
|
1163
|
+
def get_engine_status(params = {}, options = {})
|
1164
|
+
req = build_request(:get_engine_status, params)
|
1165
|
+
req.send_request(options)
|
1166
|
+
end
|
1167
|
+
|
1168
|
+
# Gets the status of a specified Gremlin query.
|
1169
|
+
#
|
1170
|
+
# @option params [required, String] :query_id
|
1171
|
+
# The unique identifier that identifies the Gremlin query.
|
1172
|
+
#
|
1173
|
+
# @return [Types::GetGremlinQueryStatusOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1174
|
+
#
|
1175
|
+
# * {Types::GetGremlinQueryStatusOutput#query_id #query_id} => String
|
1176
|
+
# * {Types::GetGremlinQueryStatusOutput#query_string #query_string} => String
|
1177
|
+
# * {Types::GetGremlinQueryStatusOutput#query_eval_stats #query_eval_stats} => Types::QueryEvalStats
|
1178
|
+
#
|
1179
|
+
# @example Request syntax with placeholder values
|
1180
|
+
#
|
1181
|
+
# resp = client.get_gremlin_query_status({
|
1182
|
+
# query_id: "String", # required
|
1183
|
+
# })
|
1184
|
+
#
|
1185
|
+
# @example Response structure
|
1186
|
+
#
|
1187
|
+
# resp.query_id #=> String
|
1188
|
+
# resp.query_string #=> String
|
1189
|
+
# resp.query_eval_stats.waited #=> Integer
|
1190
|
+
# resp.query_eval_stats.elapsed #=> Integer
|
1191
|
+
# resp.query_eval_stats.cancelled #=> Boolean
|
1192
|
+
#
|
1193
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetGremlinQueryStatus AWS API Documentation
|
1194
|
+
#
|
1195
|
+
# @overload get_gremlin_query_status(params = {})
|
1196
|
+
# @param [Hash] params ({})
|
1197
|
+
def get_gremlin_query_status(params = {}, options = {})
|
1198
|
+
req = build_request(:get_gremlin_query_status, params)
|
1199
|
+
req.send_request(options)
|
1200
|
+
end
|
1201
|
+
|
1202
|
+
# Gets status information about a specified load job. Neptune keeps
|
1203
|
+
# track of the most recent 1,024 bulk load jobs, and stores the last
|
1204
|
+
# 10,000 error details per job.
|
1205
|
+
#
|
1206
|
+
# See [Neptune Loader Get-Status API][1] for more information.
|
1207
|
+
#
|
1208
|
+
#
|
1209
|
+
#
|
1210
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/load-api-reference-status.htm
|
1211
|
+
#
|
1212
|
+
# @option params [required, String] :load_id
|
1213
|
+
# The load ID of the load job to get the status of.
|
1214
|
+
#
|
1215
|
+
# @option params [Boolean] :details
|
1216
|
+
# Flag indicating whether or not to include details beyond the overall
|
1217
|
+
# status (`TRUE` or `FALSE`; the default is `FALSE`).
|
1218
|
+
#
|
1219
|
+
# @option params [Boolean] :errors
|
1220
|
+
# Flag indicating whether or not to include a list of errors encountered
|
1221
|
+
# (`TRUE` or `FALSE`; the default is `FALSE`).
|
1222
|
+
#
|
1223
|
+
# The list of errors is paged. The `page` and `errorsPerPage` parameters
|
1224
|
+
# allow you to page through all the errors.
|
1225
|
+
#
|
1226
|
+
# @option params [Integer] :page
|
1227
|
+
# The error page number (a positive integer; the default is `1`). Only
|
1228
|
+
# valid when the `errors` parameter is set to `TRUE`.
|
1229
|
+
#
|
1230
|
+
# @option params [Integer] :errors_per_page
|
1231
|
+
# The number of errors returned in each page (a positive integer; the
|
1232
|
+
# default is `10`). Only valid when the `errors` parameter set to
|
1233
|
+
# `TRUE`.
|
1234
|
+
#
|
1235
|
+
# @return [Types::GetLoaderJobStatusOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1236
|
+
#
|
1237
|
+
# * {Types::GetLoaderJobStatusOutput#status #status} => String
|
1238
|
+
# * {Types::GetLoaderJobStatusOutput#payload #payload} => Hash,Array,String,Numeric,Boolean
|
1239
|
+
#
|
1240
|
+
# @example Request syntax with placeholder values
|
1241
|
+
#
|
1242
|
+
# resp = client.get_loader_job_status({
|
1243
|
+
# load_id: "String", # required
|
1244
|
+
# details: false,
|
1245
|
+
# errors: false,
|
1246
|
+
# page: 1,
|
1247
|
+
# errors_per_page: 1,
|
1248
|
+
# })
|
1249
|
+
#
|
1250
|
+
# @example Response structure
|
1251
|
+
#
|
1252
|
+
# resp.status #=> String
|
1253
|
+
#
|
1254
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetLoaderJobStatus AWS API Documentation
|
1255
|
+
#
|
1256
|
+
# @overload get_loader_job_status(params = {})
|
1257
|
+
# @param [Hash] params ({})
|
1258
|
+
def get_loader_job_status(params = {}, options = {})
|
1259
|
+
req = build_request(:get_loader_job_status, params)
|
1260
|
+
req.send_request(options)
|
1261
|
+
end
|
1262
|
+
|
1263
|
+
# Retrieves information about a specified data processing job. See [The
|
1264
|
+
# `dataprocessing` command][1].
|
1265
|
+
#
|
1266
|
+
#
|
1267
|
+
#
|
1268
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-dataprocessing.html
|
1269
|
+
#
|
1270
|
+
# @option params [required, String] :id
|
1271
|
+
# The unique identifier of the data-processing job to be retrieved.
|
1272
|
+
#
|
1273
|
+
# @option params [String] :neptune_iam_role_arn
|
1274
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
1275
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
1276
|
+
# group or an error will occur.
|
1277
|
+
#
|
1278
|
+
# @return [Types::GetMLDataProcessingJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1279
|
+
#
|
1280
|
+
# * {Types::GetMLDataProcessingJobOutput#status #status} => String
|
1281
|
+
# * {Types::GetMLDataProcessingJobOutput#id #id} => String
|
1282
|
+
# * {Types::GetMLDataProcessingJobOutput#processing_job #processing_job} => Types::MlResourceDefinition
|
1283
|
+
#
|
1284
|
+
# @example Request syntax with placeholder values
|
1285
|
+
#
|
1286
|
+
# resp = client.get_ml_data_processing_job({
|
1287
|
+
# id: "String", # required
|
1288
|
+
# neptune_iam_role_arn: "String",
|
1289
|
+
# })
|
1290
|
+
#
|
1291
|
+
# @example Response structure
|
1292
|
+
#
|
1293
|
+
# resp.status #=> String
|
1294
|
+
# resp.id #=> String
|
1295
|
+
# resp.processing_job.name #=> String
|
1296
|
+
# resp.processing_job.arn #=> String
|
1297
|
+
# resp.processing_job.status #=> String
|
1298
|
+
# resp.processing_job.output_location #=> String
|
1299
|
+
# resp.processing_job.failure_reason #=> String
|
1300
|
+
# resp.processing_job.cloudwatch_log_url #=> String
|
1301
|
+
#
|
1302
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetMLDataProcessingJob AWS API Documentation
|
1303
|
+
#
|
1304
|
+
# @overload get_ml_data_processing_job(params = {})
|
1305
|
+
# @param [Hash] params ({})
|
1306
|
+
def get_ml_data_processing_job(params = {}, options = {})
|
1307
|
+
req = build_request(:get_ml_data_processing_job, params)
|
1308
|
+
req.send_request(options)
|
1309
|
+
end
|
1310
|
+
|
1311
|
+
# Retrieves details about an inference endpoint. See [Managing inference
|
1312
|
+
# endpoints using the endpoints command][1].
|
1313
|
+
#
|
1314
|
+
#
|
1315
|
+
#
|
1316
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-endpoints.html
|
1317
|
+
#
|
1318
|
+
# @option params [required, String] :id
|
1319
|
+
# The unique identifier of the inference endpoint.
|
1320
|
+
#
|
1321
|
+
# @option params [String] :neptune_iam_role_arn
|
1322
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
1323
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
1324
|
+
# group or an error will occur.
|
1325
|
+
#
|
1326
|
+
# @return [Types::GetMLEndpointOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1327
|
+
#
|
1328
|
+
# * {Types::GetMLEndpointOutput#status #status} => String
|
1329
|
+
# * {Types::GetMLEndpointOutput#id #id} => String
|
1330
|
+
# * {Types::GetMLEndpointOutput#endpoint #endpoint} => Types::MlResourceDefinition
|
1331
|
+
# * {Types::GetMLEndpointOutput#endpoint_config #endpoint_config} => Types::MlConfigDefinition
|
1332
|
+
#
|
1333
|
+
# @example Request syntax with placeholder values
|
1334
|
+
#
|
1335
|
+
# resp = client.get_ml_endpoint({
|
1336
|
+
# id: "String", # required
|
1337
|
+
# neptune_iam_role_arn: "String",
|
1338
|
+
# })
|
1339
|
+
#
|
1340
|
+
# @example Response structure
|
1341
|
+
#
|
1342
|
+
# resp.status #=> String
|
1343
|
+
# resp.id #=> String
|
1344
|
+
# resp.endpoint.name #=> String
|
1345
|
+
# resp.endpoint.arn #=> String
|
1346
|
+
# resp.endpoint.status #=> String
|
1347
|
+
# resp.endpoint.output_location #=> String
|
1348
|
+
# resp.endpoint.failure_reason #=> String
|
1349
|
+
# resp.endpoint.cloudwatch_log_url #=> String
|
1350
|
+
# resp.endpoint_config.name #=> String
|
1351
|
+
# resp.endpoint_config.arn #=> String
|
1352
|
+
#
|
1353
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetMLEndpoint AWS API Documentation
|
1354
|
+
#
|
1355
|
+
# @overload get_ml_endpoint(params = {})
|
1356
|
+
# @param [Hash] params ({})
|
1357
|
+
def get_ml_endpoint(params = {}, options = {})
|
1358
|
+
req = build_request(:get_ml_endpoint, params)
|
1359
|
+
req.send_request(options)
|
1360
|
+
end
|
1361
|
+
|
1362
|
+
# Retrieves information about a Neptune ML model training job. See
|
1363
|
+
# [Model training using the `modeltraining` command][1].
|
1364
|
+
#
|
1365
|
+
#
|
1366
|
+
#
|
1367
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-modeltraining.html
|
1368
|
+
#
|
1369
|
+
# @option params [required, String] :id
|
1370
|
+
# The unique identifier of the model-training job to retrieve.
|
1371
|
+
#
|
1372
|
+
# @option params [String] :neptune_iam_role_arn
|
1373
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
1374
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
1375
|
+
# group or an error will occur.
|
1376
|
+
#
|
1377
|
+
# @return [Types::GetMLModelTrainingJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1378
|
+
#
|
1379
|
+
# * {Types::GetMLModelTrainingJobOutput#status #status} => String
|
1380
|
+
# * {Types::GetMLModelTrainingJobOutput#id #id} => String
|
1381
|
+
# * {Types::GetMLModelTrainingJobOutput#processing_job #processing_job} => Types::MlResourceDefinition
|
1382
|
+
# * {Types::GetMLModelTrainingJobOutput#hpo_job #hpo_job} => Types::MlResourceDefinition
|
1383
|
+
# * {Types::GetMLModelTrainingJobOutput#model_transform_job #model_transform_job} => Types::MlResourceDefinition
|
1384
|
+
# * {Types::GetMLModelTrainingJobOutput#ml_models #ml_models} => Array<Types::MlConfigDefinition>
|
1385
|
+
#
|
1386
|
+
# @example Request syntax with placeholder values
|
1387
|
+
#
|
1388
|
+
# resp = client.get_ml_model_training_job({
|
1389
|
+
# id: "String", # required
|
1390
|
+
# neptune_iam_role_arn: "String",
|
1391
|
+
# })
|
1392
|
+
#
|
1393
|
+
# @example Response structure
|
1394
|
+
#
|
1395
|
+
# resp.status #=> String
|
1396
|
+
# resp.id #=> String
|
1397
|
+
# resp.processing_job.name #=> String
|
1398
|
+
# resp.processing_job.arn #=> String
|
1399
|
+
# resp.processing_job.status #=> String
|
1400
|
+
# resp.processing_job.output_location #=> String
|
1401
|
+
# resp.processing_job.failure_reason #=> String
|
1402
|
+
# resp.processing_job.cloudwatch_log_url #=> String
|
1403
|
+
# resp.hpo_job.name #=> String
|
1404
|
+
# resp.hpo_job.arn #=> String
|
1405
|
+
# resp.hpo_job.status #=> String
|
1406
|
+
# resp.hpo_job.output_location #=> String
|
1407
|
+
# resp.hpo_job.failure_reason #=> String
|
1408
|
+
# resp.hpo_job.cloudwatch_log_url #=> String
|
1409
|
+
# resp.model_transform_job.name #=> String
|
1410
|
+
# resp.model_transform_job.arn #=> String
|
1411
|
+
# resp.model_transform_job.status #=> String
|
1412
|
+
# resp.model_transform_job.output_location #=> String
|
1413
|
+
# resp.model_transform_job.failure_reason #=> String
|
1414
|
+
# resp.model_transform_job.cloudwatch_log_url #=> String
|
1415
|
+
# resp.ml_models #=> Array
|
1416
|
+
# resp.ml_models[0].name #=> String
|
1417
|
+
# resp.ml_models[0].arn #=> String
|
1418
|
+
#
|
1419
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetMLModelTrainingJob AWS API Documentation
|
1420
|
+
#
|
1421
|
+
# @overload get_ml_model_training_job(params = {})
|
1422
|
+
# @param [Hash] params ({})
|
1423
|
+
def get_ml_model_training_job(params = {}, options = {})
|
1424
|
+
req = build_request(:get_ml_model_training_job, params)
|
1425
|
+
req.send_request(options)
|
1426
|
+
end
|
1427
|
+
|
1428
|
+
# Gets information about a specified model transform job. See [Use a
|
1429
|
+
# trained model to generate new model artifacts][1].
|
1430
|
+
#
|
1431
|
+
#
|
1432
|
+
#
|
1433
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-model-transform.html
|
1434
|
+
#
|
1435
|
+
# @option params [required, String] :id
|
1436
|
+
# The unique identifier of the model-transform job to be reetrieved.
|
1437
|
+
#
|
1438
|
+
# @option params [String] :neptune_iam_role_arn
|
1439
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
1440
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
1441
|
+
# group or an error will occur.
|
1442
|
+
#
|
1443
|
+
# @return [Types::GetMLModelTransformJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1444
|
+
#
|
1445
|
+
# * {Types::GetMLModelTransformJobOutput#status #status} => String
|
1446
|
+
# * {Types::GetMLModelTransformJobOutput#id #id} => String
|
1447
|
+
# * {Types::GetMLModelTransformJobOutput#base_processing_job #base_processing_job} => Types::MlResourceDefinition
|
1448
|
+
# * {Types::GetMLModelTransformJobOutput#remote_model_transform_job #remote_model_transform_job} => Types::MlResourceDefinition
|
1449
|
+
# * {Types::GetMLModelTransformJobOutput#models #models} => Array<Types::MlConfigDefinition>
|
1450
|
+
#
|
1451
|
+
# @example Request syntax with placeholder values
|
1452
|
+
#
|
1453
|
+
# resp = client.get_ml_model_transform_job({
|
1454
|
+
# id: "String", # required
|
1455
|
+
# neptune_iam_role_arn: "String",
|
1456
|
+
# })
|
1457
|
+
#
|
1458
|
+
# @example Response structure
|
1459
|
+
#
|
1460
|
+
# resp.status #=> String
|
1461
|
+
# resp.id #=> String
|
1462
|
+
# resp.base_processing_job.name #=> String
|
1463
|
+
# resp.base_processing_job.arn #=> String
|
1464
|
+
# resp.base_processing_job.status #=> String
|
1465
|
+
# resp.base_processing_job.output_location #=> String
|
1466
|
+
# resp.base_processing_job.failure_reason #=> String
|
1467
|
+
# resp.base_processing_job.cloudwatch_log_url #=> String
|
1468
|
+
# resp.remote_model_transform_job.name #=> String
|
1469
|
+
# resp.remote_model_transform_job.arn #=> String
|
1470
|
+
# resp.remote_model_transform_job.status #=> String
|
1471
|
+
# resp.remote_model_transform_job.output_location #=> String
|
1472
|
+
# resp.remote_model_transform_job.failure_reason #=> String
|
1473
|
+
# resp.remote_model_transform_job.cloudwatch_log_url #=> String
|
1474
|
+
# resp.models #=> Array
|
1475
|
+
# resp.models[0].name #=> String
|
1476
|
+
# resp.models[0].arn #=> String
|
1477
|
+
#
|
1478
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetMLModelTransformJob AWS API Documentation
|
1479
|
+
#
|
1480
|
+
# @overload get_ml_model_transform_job(params = {})
|
1481
|
+
# @param [Hash] params ({})
|
1482
|
+
def get_ml_model_transform_job(params = {}, options = {})
|
1483
|
+
req = build_request(:get_ml_model_transform_job, params)
|
1484
|
+
req.send_request(options)
|
1485
|
+
end
|
1486
|
+
|
1487
|
+
# Retrieves the status of a specified openCypher query.
|
1488
|
+
#
|
1489
|
+
# @option params [required, String] :query_id
|
1490
|
+
# The unique ID of the openCypher query for which to retrieve the query
|
1491
|
+
# status.
|
1492
|
+
#
|
1493
|
+
# @return [Types::GetOpenCypherQueryStatusOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1494
|
+
#
|
1495
|
+
# * {Types::GetOpenCypherQueryStatusOutput#query_id #query_id} => String
|
1496
|
+
# * {Types::GetOpenCypherQueryStatusOutput#query_string #query_string} => String
|
1497
|
+
# * {Types::GetOpenCypherQueryStatusOutput#query_eval_stats #query_eval_stats} => Types::QueryEvalStats
|
1498
|
+
#
|
1499
|
+
# @example Request syntax with placeholder values
|
1500
|
+
#
|
1501
|
+
# resp = client.get_open_cypher_query_status({
|
1502
|
+
# query_id: "String", # required
|
1503
|
+
# })
|
1504
|
+
#
|
1505
|
+
# @example Response structure
|
1506
|
+
#
|
1507
|
+
# resp.query_id #=> String
|
1508
|
+
# resp.query_string #=> String
|
1509
|
+
# resp.query_eval_stats.waited #=> Integer
|
1510
|
+
# resp.query_eval_stats.elapsed #=> Integer
|
1511
|
+
# resp.query_eval_stats.cancelled #=> Boolean
|
1512
|
+
#
|
1513
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetOpenCypherQueryStatus AWS API Documentation
|
1514
|
+
#
|
1515
|
+
# @overload get_open_cypher_query_status(params = {})
|
1516
|
+
# @param [Hash] params ({})
|
1517
|
+
def get_open_cypher_query_status(params = {}, options = {})
|
1518
|
+
req = build_request(:get_open_cypher_query_status, params)
|
1519
|
+
req.send_request(options)
|
1520
|
+
end
|
1521
|
+
|
1522
|
+
# Gets property graph statistics (Gremlin and openCypher).
|
1523
|
+
#
|
1524
|
+
# @return [Types::GetPropertygraphStatisticsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1525
|
+
#
|
1526
|
+
# * {Types::GetPropertygraphStatisticsOutput#status #status} => String
|
1527
|
+
# * {Types::GetPropertygraphStatisticsOutput#payload #payload} => Types::Statistics
|
1528
|
+
#
|
1529
|
+
# @example Response structure
|
1530
|
+
#
|
1531
|
+
# resp.status #=> String
|
1532
|
+
# resp.payload.auto_compute #=> Boolean
|
1533
|
+
# resp.payload.active #=> Boolean
|
1534
|
+
# resp.payload.statistics_id #=> String
|
1535
|
+
# resp.payload.date #=> Time
|
1536
|
+
# resp.payload.note #=> String
|
1537
|
+
# resp.payload.signature_info.signature_count #=> Integer
|
1538
|
+
# resp.payload.signature_info.instance_count #=> Integer
|
1539
|
+
# resp.payload.signature_info.predicate_count #=> Integer
|
1540
|
+
#
|
1541
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetPropertygraphStatistics AWS API Documentation
|
1542
|
+
#
|
1543
|
+
# @overload get_propertygraph_statistics(params = {})
|
1544
|
+
# @param [Hash] params ({})
|
1545
|
+
def get_propertygraph_statistics(params = {}, options = {})
|
1546
|
+
req = build_request(:get_propertygraph_statistics, params)
|
1547
|
+
req.send_request(options)
|
1548
|
+
end
|
1549
|
+
|
1550
|
+
# Gets a stream for a property graph.
|
1551
|
+
#
|
1552
|
+
# With the Neptune Streams feature, you can generate a complete sequence
|
1553
|
+
# of change-log entries that record every change made to your graph data
|
1554
|
+
# as it happens. `GetPropertygraphStream` lets you collect these
|
1555
|
+
# change-log entries for a property graph.
|
1556
|
+
#
|
1557
|
+
# The Neptune streams feature needs to be enabled on your Neptune
|
1558
|
+
# DBcluster. To enable streams, set the [neptune\_streams][1] DB cluster
|
1559
|
+
# parameter to `1`.
|
1560
|
+
#
|
1561
|
+
# See [Capturing graph changes in real time using Neptune streams][2].
|
1562
|
+
#
|
1563
|
+
#
|
1564
|
+
#
|
1565
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/parameters.html#parameters-db-cluster-parameters-neptune_streams
|
1566
|
+
# [2]: https://docs.aws.amazon.com/neptune/latest/userguide/streams.html
|
1567
|
+
#
|
1568
|
+
# @option params [Integer] :limit
|
1569
|
+
# Specifies the maximum number of records to return. There is also a
|
1570
|
+
# size limit of 10 MB on the response that can't be modified and that
|
1571
|
+
# takes precedence over the number of records specified in the `limit`
|
1572
|
+
# parameter. The response does include a threshold-breaching record if
|
1573
|
+
# the 10 MB limit was reached.
|
1574
|
+
#
|
1575
|
+
# The range for `limit` is 1 to 100,000, with a default of 10.
|
1576
|
+
#
|
1577
|
+
# @option params [String] :iterator_type
|
1578
|
+
# Can be one of:
|
1579
|
+
#
|
1580
|
+
# * `AT_SEQUENCE_NUMBER` – Indicates that reading should start from
|
1581
|
+
# the event sequence number specified jointly by the `commitNum` and
|
1582
|
+
# `opNum` parameters.
|
1583
|
+
#
|
1584
|
+
# * `AFTER_SEQUENCE_NUMBER` – Indicates that reading should start
|
1585
|
+
# right after the event sequence number specified jointly by the
|
1586
|
+
# `commitNum` and `opNum` parameters.
|
1587
|
+
#
|
1588
|
+
# * `TRIM_HORIZON` – Indicates that reading should start at the last
|
1589
|
+
# untrimmed record in the system, which is the oldest unexpired (not
|
1590
|
+
# yet deleted) record in the change-log stream.
|
1591
|
+
#
|
1592
|
+
# * `LATEST` – Indicates that reading should start at the most
|
1593
|
+
# recent record in the system, which is the latest unexpired (not yet
|
1594
|
+
# deleted) record in the change-log stream.
|
1595
|
+
#
|
1596
|
+
# @option params [Integer] :commit_num
|
1597
|
+
# The commit number of the starting record to read from the change-log
|
1598
|
+
# stream. This parameter is required when `iteratorType`
|
1599
|
+
# is`AT_SEQUENCE_NUMBER` or `AFTER_SEQUENCE_NUMBER`, and ignored when
|
1600
|
+
# `iteratorType` is `TRIM_HORIZON` or `LATEST`.
|
1601
|
+
#
|
1602
|
+
# @option params [Integer] :op_num
|
1603
|
+
# The operation sequence number within the specified commit to start
|
1604
|
+
# reading from in the change-log stream data. The default is `1`.
|
1605
|
+
#
|
1606
|
+
# @option params [String] :encoding
|
1607
|
+
# If set to TRUE, Neptune compresses the response using gzip encoding.
|
1608
|
+
#
|
1609
|
+
# @return [Types::GetPropertygraphStreamOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1610
|
+
#
|
1611
|
+
# * {Types::GetPropertygraphStreamOutput#last_event_id #last_event_id} => Hash<String,String>
|
1612
|
+
# * {Types::GetPropertygraphStreamOutput#last_trx_timestamp_in_millis #last_trx_timestamp_in_millis} => Integer
|
1613
|
+
# * {Types::GetPropertygraphStreamOutput#format #format} => String
|
1614
|
+
# * {Types::GetPropertygraphStreamOutput#records #records} => Array<Types::PropertygraphRecord>
|
1615
|
+
# * {Types::GetPropertygraphStreamOutput#total_records #total_records} => Integer
|
1616
|
+
#
|
1617
|
+
# @example Request syntax with placeholder values
|
1618
|
+
#
|
1619
|
+
# resp = client.get_propertygraph_stream({
|
1620
|
+
# limit: 1,
|
1621
|
+
# iterator_type: "AT_SEQUENCE_NUMBER", # accepts AT_SEQUENCE_NUMBER, AFTER_SEQUENCE_NUMBER, TRIM_HORIZON, LATEST
|
1622
|
+
# commit_num: 1,
|
1623
|
+
# op_num: 1,
|
1624
|
+
# encoding: "gzip", # accepts gzip
|
1625
|
+
# })
|
1626
|
+
#
|
1627
|
+
# @example Response structure
|
1628
|
+
#
|
1629
|
+
# resp.last_event_id #=> Hash
|
1630
|
+
# resp.last_event_id["String"] #=> String
|
1631
|
+
# resp.last_trx_timestamp_in_millis #=> Integer
|
1632
|
+
# resp.format #=> String
|
1633
|
+
# resp.records #=> Array
|
1634
|
+
# resp.records[0].commit_timestamp_in_millis #=> Integer
|
1635
|
+
# resp.records[0].event_id #=> Hash
|
1636
|
+
# resp.records[0].event_id["String"] #=> String
|
1637
|
+
# resp.records[0].data.id #=> String
|
1638
|
+
# resp.records[0].data.type #=> String
|
1639
|
+
# resp.records[0].data.key #=> String
|
1640
|
+
# resp.records[0].data.from #=> String
|
1641
|
+
# resp.records[0].data.to #=> String
|
1642
|
+
# resp.records[0].op #=> String
|
1643
|
+
# resp.records[0].is_last_op #=> Boolean
|
1644
|
+
# resp.total_records #=> Integer
|
1645
|
+
#
|
1646
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetPropertygraphStream AWS API Documentation
|
1647
|
+
#
|
1648
|
+
# @overload get_propertygraph_stream(params = {})
|
1649
|
+
# @param [Hash] params ({})
|
1650
|
+
def get_propertygraph_stream(params = {}, options = {})
|
1651
|
+
req = build_request(:get_propertygraph_stream, params)
|
1652
|
+
req.send_request(options)
|
1653
|
+
end
|
1654
|
+
|
1655
|
+
# Gets a graph summary for a property graph.
|
1656
|
+
#
|
1657
|
+
# @option params [String] :mode
|
1658
|
+
# Mode can take one of two values: `BASIC` (the default), and
|
1659
|
+
# `DETAILED`.
|
1660
|
+
#
|
1661
|
+
# @return [Types::GetPropertygraphSummaryOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1662
|
+
#
|
1663
|
+
# * {Types::GetPropertygraphSummaryOutput#status_code #status_code} => Integer
|
1664
|
+
# * {Types::GetPropertygraphSummaryOutput#payload #payload} => Types::PropertygraphSummaryValueMap
|
1665
|
+
#
|
1666
|
+
# @example Request syntax with placeholder values
|
1667
|
+
#
|
1668
|
+
# resp = client.get_propertygraph_summary({
|
1669
|
+
# mode: "basic", # accepts basic, detailed
|
1670
|
+
# })
|
1671
|
+
#
|
1672
|
+
# @example Response structure
|
1673
|
+
#
|
1674
|
+
# resp.status_code #=> Integer
|
1675
|
+
# resp.payload.version #=> String
|
1676
|
+
# resp.payload.last_statistics_computation_time #=> Time
|
1677
|
+
# resp.payload.graph_summary.num_nodes #=> Integer
|
1678
|
+
# resp.payload.graph_summary.num_edges #=> Integer
|
1679
|
+
# resp.payload.graph_summary.num_node_labels #=> Integer
|
1680
|
+
# resp.payload.graph_summary.num_edge_labels #=> Integer
|
1681
|
+
# resp.payload.graph_summary.node_labels #=> Array
|
1682
|
+
# resp.payload.graph_summary.node_labels[0] #=> String
|
1683
|
+
# resp.payload.graph_summary.edge_labels #=> Array
|
1684
|
+
# resp.payload.graph_summary.edge_labels[0] #=> String
|
1685
|
+
# resp.payload.graph_summary.num_node_properties #=> Integer
|
1686
|
+
# resp.payload.graph_summary.num_edge_properties #=> Integer
|
1687
|
+
# resp.payload.graph_summary.node_properties #=> Array
|
1688
|
+
# resp.payload.graph_summary.node_properties[0] #=> Hash
|
1689
|
+
# resp.payload.graph_summary.node_properties[0]["String"] #=> Integer
|
1690
|
+
# resp.payload.graph_summary.edge_properties #=> Array
|
1691
|
+
# resp.payload.graph_summary.edge_properties[0] #=> Hash
|
1692
|
+
# resp.payload.graph_summary.edge_properties[0]["String"] #=> Integer
|
1693
|
+
# resp.payload.graph_summary.total_node_property_values #=> Integer
|
1694
|
+
# resp.payload.graph_summary.total_edge_property_values #=> Integer
|
1695
|
+
# resp.payload.graph_summary.node_structures #=> Array
|
1696
|
+
# resp.payload.graph_summary.node_structures[0].count #=> Integer
|
1697
|
+
# resp.payload.graph_summary.node_structures[0].node_properties #=> Array
|
1698
|
+
# resp.payload.graph_summary.node_structures[0].node_properties[0] #=> String
|
1699
|
+
# resp.payload.graph_summary.node_structures[0].distinct_outgoing_edge_labels #=> Array
|
1700
|
+
# resp.payload.graph_summary.node_structures[0].distinct_outgoing_edge_labels[0] #=> String
|
1701
|
+
# resp.payload.graph_summary.edge_structures #=> Array
|
1702
|
+
# resp.payload.graph_summary.edge_structures[0].count #=> Integer
|
1703
|
+
# resp.payload.graph_summary.edge_structures[0].edge_properties #=> Array
|
1704
|
+
# resp.payload.graph_summary.edge_structures[0].edge_properties[0] #=> String
|
1705
|
+
#
|
1706
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetPropertygraphSummary AWS API Documentation
|
1707
|
+
#
|
1708
|
+
# @overload get_propertygraph_summary(params = {})
|
1709
|
+
# @param [Hash] params ({})
|
1710
|
+
def get_propertygraph_summary(params = {}, options = {})
|
1711
|
+
req = build_request(:get_propertygraph_summary, params)
|
1712
|
+
req.send_request(options)
|
1713
|
+
end
|
1714
|
+
|
1715
|
+
# Gets a graph summary for an RDF graph.
|
1716
|
+
#
|
1717
|
+
# @option params [String] :mode
|
1718
|
+
# Mode can take one of two values: `BASIC` (the default), and
|
1719
|
+
# `DETAILED`.
|
1720
|
+
#
|
1721
|
+
# @return [Types::GetRDFGraphSummaryOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1722
|
+
#
|
1723
|
+
# * {Types::GetRDFGraphSummaryOutput#status_code #status_code} => Integer
|
1724
|
+
# * {Types::GetRDFGraphSummaryOutput#payload #payload} => Types::RDFGraphSummaryValueMap
|
1725
|
+
#
|
1726
|
+
# @example Request syntax with placeholder values
|
1727
|
+
#
|
1728
|
+
# resp = client.get_rdf_graph_summary({
|
1729
|
+
# mode: "basic", # accepts basic, detailed
|
1730
|
+
# })
|
1731
|
+
#
|
1732
|
+
# @example Response structure
|
1733
|
+
#
|
1734
|
+
# resp.status_code #=> Integer
|
1735
|
+
# resp.payload.version #=> String
|
1736
|
+
# resp.payload.last_statistics_computation_time #=> Time
|
1737
|
+
# resp.payload.graph_summary.num_distinct_subjects #=> Integer
|
1738
|
+
# resp.payload.graph_summary.num_distinct_predicates #=> Integer
|
1739
|
+
# resp.payload.graph_summary.num_quads #=> Integer
|
1740
|
+
# resp.payload.graph_summary.num_classes #=> Integer
|
1741
|
+
# resp.payload.graph_summary.classes #=> Array
|
1742
|
+
# resp.payload.graph_summary.classes[0] #=> String
|
1743
|
+
# resp.payload.graph_summary.predicates #=> Array
|
1744
|
+
# resp.payload.graph_summary.predicates[0] #=> Hash
|
1745
|
+
# resp.payload.graph_summary.predicates[0]["String"] #=> Integer
|
1746
|
+
# resp.payload.graph_summary.subject_structures #=> Array
|
1747
|
+
# resp.payload.graph_summary.subject_structures[0].count #=> Integer
|
1748
|
+
# resp.payload.graph_summary.subject_structures[0].predicates #=> Array
|
1749
|
+
# resp.payload.graph_summary.subject_structures[0].predicates[0] #=> String
|
1750
|
+
#
|
1751
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetRDFGraphSummary AWS API Documentation
|
1752
|
+
#
|
1753
|
+
# @overload get_rdf_graph_summary(params = {})
|
1754
|
+
# @param [Hash] params ({})
|
1755
|
+
def get_rdf_graph_summary(params = {}, options = {})
|
1756
|
+
req = build_request(:get_rdf_graph_summary, params)
|
1757
|
+
req.send_request(options)
|
1758
|
+
end
|
1759
|
+
|
1760
|
+
# Gets RDF statistics (SPARQL).
|
1761
|
+
#
|
1762
|
+
# @return [Types::GetSparqlStatisticsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1763
|
+
#
|
1764
|
+
# * {Types::GetSparqlStatisticsOutput#status #status} => String
|
1765
|
+
# * {Types::GetSparqlStatisticsOutput#payload #payload} => Types::Statistics
|
1766
|
+
#
|
1767
|
+
# @example Response structure
|
1768
|
+
#
|
1769
|
+
# resp.status #=> String
|
1770
|
+
# resp.payload.auto_compute #=> Boolean
|
1771
|
+
# resp.payload.active #=> Boolean
|
1772
|
+
# resp.payload.statistics_id #=> String
|
1773
|
+
# resp.payload.date #=> Time
|
1774
|
+
# resp.payload.note #=> String
|
1775
|
+
# resp.payload.signature_info.signature_count #=> Integer
|
1776
|
+
# resp.payload.signature_info.instance_count #=> Integer
|
1777
|
+
# resp.payload.signature_info.predicate_count #=> Integer
|
1778
|
+
#
|
1779
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetSparqlStatistics AWS API Documentation
|
1780
|
+
#
|
1781
|
+
# @overload get_sparql_statistics(params = {})
|
1782
|
+
# @param [Hash] params ({})
|
1783
|
+
def get_sparql_statistics(params = {}, options = {})
|
1784
|
+
req = build_request(:get_sparql_statistics, params)
|
1785
|
+
req.send_request(options)
|
1786
|
+
end
|
1787
|
+
|
1788
|
+
# Gets a stream for an RDF graph.
|
1789
|
+
#
|
1790
|
+
# With the Neptune Streams feature, you can generate a complete sequence
|
1791
|
+
# of change-log entries that record every change made to your graph data
|
1792
|
+
# as it happens. `GetSparqlStream` lets you collect these change-log
|
1793
|
+
# entries for an RDF graph.
|
1794
|
+
#
|
1795
|
+
# The Neptune streams feature needs to be enabled on your Neptune
|
1796
|
+
# DBcluster. To enable streams, set the [neptune\_streams][1] DB cluster
|
1797
|
+
# parameter to `1`.
|
1798
|
+
#
|
1799
|
+
# See [Capturing graph changes in real time using Neptune streams][2].
|
1800
|
+
#
|
1801
|
+
#
|
1802
|
+
#
|
1803
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/parameters.html#parameters-db-cluster-parameters-neptune_streams
|
1804
|
+
# [2]: https://docs.aws.amazon.com/neptune/latest/userguide/streams.html
|
1805
|
+
#
|
1806
|
+
# @option params [Integer] :limit
|
1807
|
+
# Specifies the maximum number of records to return. There is also a
|
1808
|
+
# size limit of 10 MB on the response that can't be modified and that
|
1809
|
+
# takes precedence over the number of records specified in the `limit`
|
1810
|
+
# parameter. The response does include a threshold-breaching record if
|
1811
|
+
# the 10 MB limit was reached.
|
1812
|
+
#
|
1813
|
+
# The range for `limit` is 1 to 100,000, with a default of 10.
|
1814
|
+
#
|
1815
|
+
# @option params [String] :iterator_type
|
1816
|
+
# Can be one of:
|
1817
|
+
#
|
1818
|
+
# * `AT_SEQUENCE_NUMBER` – Indicates that reading should start from
|
1819
|
+
# the event sequence number specified jointly by the `commitNum` and
|
1820
|
+
# `opNum` parameters.
|
1821
|
+
#
|
1822
|
+
# * `AFTER_SEQUENCE_NUMBER` – Indicates that reading should start
|
1823
|
+
# right after the event sequence number specified jointly by the
|
1824
|
+
# `commitNum` and `opNum` parameters.
|
1825
|
+
#
|
1826
|
+
# * `TRIM_HORIZON` – Indicates that reading should start at the last
|
1827
|
+
# untrimmed record in the system, which is the oldest unexpired (not
|
1828
|
+
# yet deleted) record in the change-log stream.
|
1829
|
+
#
|
1830
|
+
# * `LATEST` – Indicates that reading should start at the most
|
1831
|
+
# recent record in the system, which is the latest unexpired (not yet
|
1832
|
+
# deleted) record in the change-log stream.
|
1833
|
+
#
|
1834
|
+
# @option params [Integer] :commit_num
|
1835
|
+
# The commit number of the starting record to read from the change-log
|
1836
|
+
# stream. This parameter is required when `iteratorType`
|
1837
|
+
# is`AT_SEQUENCE_NUMBER` or `AFTER_SEQUENCE_NUMBER`, and ignored when
|
1838
|
+
# `iteratorType` is `TRIM_HORIZON` or `LATEST`.
|
1839
|
+
#
|
1840
|
+
# @option params [Integer] :op_num
|
1841
|
+
# The operation sequence number within the specified commit to start
|
1842
|
+
# reading from in the change-log stream data. The default is `1`.
|
1843
|
+
#
|
1844
|
+
# @option params [String] :encoding
|
1845
|
+
# If set to TRUE, Neptune compresses the response using gzip encoding.
|
1846
|
+
#
|
1847
|
+
# @return [Types::GetSparqlStreamOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1848
|
+
#
|
1849
|
+
# * {Types::GetSparqlStreamOutput#last_event_id #last_event_id} => Hash<String,String>
|
1850
|
+
# * {Types::GetSparqlStreamOutput#last_trx_timestamp_in_millis #last_trx_timestamp_in_millis} => Integer
|
1851
|
+
# * {Types::GetSparqlStreamOutput#format #format} => String
|
1852
|
+
# * {Types::GetSparqlStreamOutput#records #records} => Array<Types::SparqlRecord>
|
1853
|
+
# * {Types::GetSparqlStreamOutput#total_records #total_records} => Integer
|
1854
|
+
#
|
1855
|
+
# @example Request syntax with placeholder values
|
1856
|
+
#
|
1857
|
+
# resp = client.get_sparql_stream({
|
1858
|
+
# limit: 1,
|
1859
|
+
# iterator_type: "AT_SEQUENCE_NUMBER", # accepts AT_SEQUENCE_NUMBER, AFTER_SEQUENCE_NUMBER, TRIM_HORIZON, LATEST
|
1860
|
+
# commit_num: 1,
|
1861
|
+
# op_num: 1,
|
1862
|
+
# encoding: "gzip", # accepts gzip
|
1863
|
+
# })
|
1864
|
+
#
|
1865
|
+
# @example Response structure
|
1866
|
+
#
|
1867
|
+
# resp.last_event_id #=> Hash
|
1868
|
+
# resp.last_event_id["String"] #=> String
|
1869
|
+
# resp.last_trx_timestamp_in_millis #=> Integer
|
1870
|
+
# resp.format #=> String
|
1871
|
+
# resp.records #=> Array
|
1872
|
+
# resp.records[0].commit_timestamp_in_millis #=> Integer
|
1873
|
+
# resp.records[0].event_id #=> Hash
|
1874
|
+
# resp.records[0].event_id["String"] #=> String
|
1875
|
+
# resp.records[0].data.stmt #=> String
|
1876
|
+
# resp.records[0].op #=> String
|
1877
|
+
# resp.records[0].is_last_op #=> Boolean
|
1878
|
+
# resp.total_records #=> Integer
|
1879
|
+
#
|
1880
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/GetSparqlStream AWS API Documentation
|
1881
|
+
#
|
1882
|
+
# @overload get_sparql_stream(params = {})
|
1883
|
+
# @param [Hash] params ({})
|
1884
|
+
def get_sparql_stream(params = {}, options = {})
|
1885
|
+
req = build_request(:get_sparql_stream, params)
|
1886
|
+
req.send_request(options)
|
1887
|
+
end
|
1888
|
+
|
1889
|
+
# Lists active Gremlin queries. See [Gremlin query status API][1] for
|
1890
|
+
# details about the output.
|
1891
|
+
#
|
1892
|
+
#
|
1893
|
+
#
|
1894
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/gremlin-api-status.html
|
1895
|
+
#
|
1896
|
+
# @option params [Boolean] :include_waiting
|
1897
|
+
# If set to `TRUE`, the list returned includes waiting queries. The
|
1898
|
+
# default is `FALSE`;
|
1899
|
+
#
|
1900
|
+
# @return [Types::ListGremlinQueriesOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1901
|
+
#
|
1902
|
+
# * {Types::ListGremlinQueriesOutput#accepted_query_count #accepted_query_count} => Integer
|
1903
|
+
# * {Types::ListGremlinQueriesOutput#running_query_count #running_query_count} => Integer
|
1904
|
+
# * {Types::ListGremlinQueriesOutput#queries #queries} => Array<Types::GremlinQueryStatus>
|
1905
|
+
#
|
1906
|
+
# @example Request syntax with placeholder values
|
1907
|
+
#
|
1908
|
+
# resp = client.list_gremlin_queries({
|
1909
|
+
# include_waiting: false,
|
1910
|
+
# })
|
1911
|
+
#
|
1912
|
+
# @example Response structure
|
1913
|
+
#
|
1914
|
+
# resp.accepted_query_count #=> Integer
|
1915
|
+
# resp.running_query_count #=> Integer
|
1916
|
+
# resp.queries #=> Array
|
1917
|
+
# resp.queries[0].query_id #=> String
|
1918
|
+
# resp.queries[0].query_string #=> String
|
1919
|
+
# resp.queries[0].query_eval_stats.waited #=> Integer
|
1920
|
+
# resp.queries[0].query_eval_stats.elapsed #=> Integer
|
1921
|
+
# resp.queries[0].query_eval_stats.cancelled #=> Boolean
|
1922
|
+
#
|
1923
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ListGremlinQueries AWS API Documentation
|
1924
|
+
#
|
1925
|
+
# @overload list_gremlin_queries(params = {})
|
1926
|
+
# @param [Hash] params ({})
|
1927
|
+
def list_gremlin_queries(params = {}, options = {})
|
1928
|
+
req = build_request(:list_gremlin_queries, params)
|
1929
|
+
req.send_request(options)
|
1930
|
+
end
|
1931
|
+
|
1932
|
+
# Retrieves a list of the `loadIds` for all active loader jobs.
|
1933
|
+
#
|
1934
|
+
# @option params [Integer] :limit
|
1935
|
+
# The number of load IDs to list. Must be a positive integer greater
|
1936
|
+
# than zero and not more than `100` (which is the default).
|
1937
|
+
#
|
1938
|
+
# @option params [Boolean] :include_queued_loads
|
1939
|
+
# An optional parameter that can be used to exclude the load IDs of
|
1940
|
+
# queued load requests when requesting a list of load IDs by setting the
|
1941
|
+
# parameter to `FALSE`. The default value is `TRUE`.
|
1942
|
+
#
|
1943
|
+
# @return [Types::ListLoaderJobsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1944
|
+
#
|
1945
|
+
# * {Types::ListLoaderJobsOutput#status #status} => String
|
1946
|
+
# * {Types::ListLoaderJobsOutput#payload #payload} => Types::LoaderIdResult
|
1947
|
+
#
|
1948
|
+
# @example Request syntax with placeholder values
|
1949
|
+
#
|
1950
|
+
# resp = client.list_loader_jobs({
|
1951
|
+
# limit: 1,
|
1952
|
+
# include_queued_loads: false,
|
1953
|
+
# })
|
1954
|
+
#
|
1955
|
+
# @example Response structure
|
1956
|
+
#
|
1957
|
+
# resp.status #=> String
|
1958
|
+
# resp.payload.load_ids #=> Array
|
1959
|
+
# resp.payload.load_ids[0] #=> String
|
1960
|
+
#
|
1961
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ListLoaderJobs AWS API Documentation
|
1962
|
+
#
|
1963
|
+
# @overload list_loader_jobs(params = {})
|
1964
|
+
# @param [Hash] params ({})
|
1965
|
+
def list_loader_jobs(params = {}, options = {})
|
1966
|
+
req = build_request(:list_loader_jobs, params)
|
1967
|
+
req.send_request(options)
|
1968
|
+
end
|
1969
|
+
|
1970
|
+
# Returns a list of Neptune ML data processing jobs. See [Listing active
|
1971
|
+
# data-processing jobs using the Neptune ML dataprocessing command][1].
|
1972
|
+
#
|
1973
|
+
#
|
1974
|
+
#
|
1975
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-dataprocessing.html#machine-learning-api-dataprocessing-list-jobs
|
1976
|
+
#
|
1977
|
+
# @option params [Integer] :max_items
|
1978
|
+
# The maximum number of items to return (from 1 to 1024; the default is
|
1979
|
+
# 10).
|
1980
|
+
#
|
1981
|
+
# @option params [String] :neptune_iam_role_arn
|
1982
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
1983
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
1984
|
+
# group or an error will occur.
|
1985
|
+
#
|
1986
|
+
# @return [Types::ListMLDataProcessingJobsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
1987
|
+
#
|
1988
|
+
# * {Types::ListMLDataProcessingJobsOutput#ids #ids} => Array<String>
|
1989
|
+
#
|
1990
|
+
# @example Request syntax with placeholder values
|
1991
|
+
#
|
1992
|
+
# resp = client.list_ml_data_processing_jobs({
|
1993
|
+
# max_items: 1,
|
1994
|
+
# neptune_iam_role_arn: "String",
|
1995
|
+
# })
|
1996
|
+
#
|
1997
|
+
# @example Response structure
|
1998
|
+
#
|
1999
|
+
# resp.ids #=> Array
|
2000
|
+
# resp.ids[0] #=> String
|
2001
|
+
#
|
2002
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ListMLDataProcessingJobs AWS API Documentation
|
2003
|
+
#
|
2004
|
+
# @overload list_ml_data_processing_jobs(params = {})
|
2005
|
+
# @param [Hash] params ({})
|
2006
|
+
def list_ml_data_processing_jobs(params = {}, options = {})
|
2007
|
+
req = build_request(:list_ml_data_processing_jobs, params)
|
2008
|
+
req.send_request(options)
|
2009
|
+
end
|
2010
|
+
|
2011
|
+
# Lists existing inference endpoints. See [Managing inference endpoints
|
2012
|
+
# using the endpoints command][1].
|
2013
|
+
#
|
2014
|
+
#
|
2015
|
+
#
|
2016
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-endpoints.html
|
2017
|
+
#
|
2018
|
+
# @option params [Integer] :max_items
|
2019
|
+
# The maximum number of items to return (from 1 to 1024; the default is
|
2020
|
+
# 10.
|
2021
|
+
#
|
2022
|
+
# @option params [String] :neptune_iam_role_arn
|
2023
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
2024
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
2025
|
+
# group or an error will occur.
|
2026
|
+
#
|
2027
|
+
# @return [Types::ListMLEndpointsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2028
|
+
#
|
2029
|
+
# * {Types::ListMLEndpointsOutput#ids #ids} => Array<String>
|
2030
|
+
#
|
2031
|
+
# @example Request syntax with placeholder values
|
2032
|
+
#
|
2033
|
+
# resp = client.list_ml_endpoints({
|
2034
|
+
# max_items: 1,
|
2035
|
+
# neptune_iam_role_arn: "String",
|
2036
|
+
# })
|
2037
|
+
#
|
2038
|
+
# @example Response structure
|
2039
|
+
#
|
2040
|
+
# resp.ids #=> Array
|
2041
|
+
# resp.ids[0] #=> String
|
2042
|
+
#
|
2043
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ListMLEndpoints AWS API Documentation
|
2044
|
+
#
|
2045
|
+
# @overload list_ml_endpoints(params = {})
|
2046
|
+
# @param [Hash] params ({})
|
2047
|
+
def list_ml_endpoints(params = {}, options = {})
|
2048
|
+
req = build_request(:list_ml_endpoints, params)
|
2049
|
+
req.send_request(options)
|
2050
|
+
end
|
2051
|
+
|
2052
|
+
# Lists Neptune ML model-training jobs. See [Model training using the
|
2053
|
+
# `modeltraining` command][1].
|
2054
|
+
#
|
2055
|
+
#
|
2056
|
+
#
|
2057
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-modeltraining.html
|
2058
|
+
#
|
2059
|
+
# @option params [Integer] :max_items
|
2060
|
+
# The maximum number of items to return (from 1 to 1024; the default is
|
2061
|
+
# 10).
|
2062
|
+
#
|
2063
|
+
# @option params [String] :neptune_iam_role_arn
|
2064
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
2065
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
2066
|
+
# group or an error will occur.
|
2067
|
+
#
|
2068
|
+
# @return [Types::ListMLModelTrainingJobsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2069
|
+
#
|
2070
|
+
# * {Types::ListMLModelTrainingJobsOutput#ids #ids} => Array<String>
|
2071
|
+
#
|
2072
|
+
# @example Request syntax with placeholder values
|
2073
|
+
#
|
2074
|
+
# resp = client.list_ml_model_training_jobs({
|
2075
|
+
# max_items: 1,
|
2076
|
+
# neptune_iam_role_arn: "String",
|
2077
|
+
# })
|
2078
|
+
#
|
2079
|
+
# @example Response structure
|
2080
|
+
#
|
2081
|
+
# resp.ids #=> Array
|
2082
|
+
# resp.ids[0] #=> String
|
2083
|
+
#
|
2084
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ListMLModelTrainingJobs AWS API Documentation
|
2085
|
+
#
|
2086
|
+
# @overload list_ml_model_training_jobs(params = {})
|
2087
|
+
# @param [Hash] params ({})
|
2088
|
+
def list_ml_model_training_jobs(params = {}, options = {})
|
2089
|
+
req = build_request(:list_ml_model_training_jobs, params)
|
2090
|
+
req.send_request(options)
|
2091
|
+
end
|
2092
|
+
|
2093
|
+
# Returns a list of model transform job IDs. See [Use a trained model to
|
2094
|
+
# generate new model artifacts][1].
|
2095
|
+
#
|
2096
|
+
#
|
2097
|
+
#
|
2098
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-model-transform.html
|
2099
|
+
#
|
2100
|
+
# @option params [Integer] :max_items
|
2101
|
+
# The maximum number of items to return (from 1 to 1024; the default is
|
2102
|
+
# 10).
|
2103
|
+
#
|
2104
|
+
# @option params [String] :neptune_iam_role_arn
|
2105
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
2106
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
2107
|
+
# group or an error will occur.
|
2108
|
+
#
|
2109
|
+
# @return [Types::ListMLModelTransformJobsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2110
|
+
#
|
2111
|
+
# * {Types::ListMLModelTransformJobsOutput#ids #ids} => Array<String>
|
2112
|
+
#
|
2113
|
+
# @example Request syntax with placeholder values
|
2114
|
+
#
|
2115
|
+
# resp = client.list_ml_model_transform_jobs({
|
2116
|
+
# max_items: 1,
|
2117
|
+
# neptune_iam_role_arn: "String",
|
2118
|
+
# })
|
2119
|
+
#
|
2120
|
+
# @example Response structure
|
2121
|
+
#
|
2122
|
+
# resp.ids #=> Array
|
2123
|
+
# resp.ids[0] #=> String
|
2124
|
+
#
|
2125
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ListMLModelTransformJobs AWS API Documentation
|
2126
|
+
#
|
2127
|
+
# @overload list_ml_model_transform_jobs(params = {})
|
2128
|
+
# @param [Hash] params ({})
|
2129
|
+
def list_ml_model_transform_jobs(params = {}, options = {})
|
2130
|
+
req = build_request(:list_ml_model_transform_jobs, params)
|
2131
|
+
req.send_request(options)
|
2132
|
+
end
|
2133
|
+
|
2134
|
+
# Lists active openCypher queries. See [Neptune openCypher status
|
2135
|
+
# endpoint][1] for more information.
|
2136
|
+
#
|
2137
|
+
#
|
2138
|
+
#
|
2139
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/access-graph-opencypher-status.html
|
2140
|
+
#
|
2141
|
+
# @option params [Boolean] :include_waiting
|
2142
|
+
# When set to `TRUE` and other parameters are not present, causes status
|
2143
|
+
# information to be returned for waiting queries as well as for running
|
2144
|
+
# queries.
|
2145
|
+
#
|
2146
|
+
# @return [Types::ListOpenCypherQueriesOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2147
|
+
#
|
2148
|
+
# * {Types::ListOpenCypherQueriesOutput#accepted_query_count #accepted_query_count} => Integer
|
2149
|
+
# * {Types::ListOpenCypherQueriesOutput#running_query_count #running_query_count} => Integer
|
2150
|
+
# * {Types::ListOpenCypherQueriesOutput#queries #queries} => Array<Types::GremlinQueryStatus>
|
2151
|
+
#
|
2152
|
+
# @example Request syntax with placeholder values
|
2153
|
+
#
|
2154
|
+
# resp = client.list_open_cypher_queries({
|
2155
|
+
# include_waiting: false,
|
2156
|
+
# })
|
2157
|
+
#
|
2158
|
+
# @example Response structure
|
2159
|
+
#
|
2160
|
+
# resp.accepted_query_count #=> Integer
|
2161
|
+
# resp.running_query_count #=> Integer
|
2162
|
+
# resp.queries #=> Array
|
2163
|
+
# resp.queries[0].query_id #=> String
|
2164
|
+
# resp.queries[0].query_string #=> String
|
2165
|
+
# resp.queries[0].query_eval_stats.waited #=> Integer
|
2166
|
+
# resp.queries[0].query_eval_stats.elapsed #=> Integer
|
2167
|
+
# resp.queries[0].query_eval_stats.cancelled #=> Boolean
|
2168
|
+
#
|
2169
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ListOpenCypherQueries AWS API Documentation
|
2170
|
+
#
|
2171
|
+
# @overload list_open_cypher_queries(params = {})
|
2172
|
+
# @param [Hash] params ({})
|
2173
|
+
def list_open_cypher_queries(params = {}, options = {})
|
2174
|
+
req = build_request(:list_open_cypher_queries, params)
|
2175
|
+
req.send_request(options)
|
2176
|
+
end
|
2177
|
+
|
2178
|
+
# Manages the generation and use of property graph statistics.
|
2179
|
+
#
|
2180
|
+
# @option params [String] :mode
|
2181
|
+
# The statistics generation mode. One of: `DISABLE_AUTOCOMPUTE`,
|
2182
|
+
# `ENABLE_AUTOCOMPUTE`, or `REFRESH`, the last of which manually
|
2183
|
+
# triggers DFE statistics generation.
|
2184
|
+
#
|
2185
|
+
# @return [Types::ManagePropertygraphStatisticsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2186
|
+
#
|
2187
|
+
# * {Types::ManagePropertygraphStatisticsOutput#status #status} => String
|
2188
|
+
# * {Types::ManagePropertygraphStatisticsOutput#payload #payload} => Types::RefreshStatisticsIdMap
|
2189
|
+
#
|
2190
|
+
# @example Request syntax with placeholder values
|
2191
|
+
#
|
2192
|
+
# resp = client.manage_propertygraph_statistics({
|
2193
|
+
# mode: "disableAutoCompute", # accepts disableAutoCompute, enableAutoCompute, refresh
|
2194
|
+
# })
|
2195
|
+
#
|
2196
|
+
# @example Response structure
|
2197
|
+
#
|
2198
|
+
# resp.status #=> String
|
2199
|
+
# resp.payload.statistics_id #=> String
|
2200
|
+
#
|
2201
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ManagePropertygraphStatistics AWS API Documentation
|
2202
|
+
#
|
2203
|
+
# @overload manage_propertygraph_statistics(params = {})
|
2204
|
+
# @param [Hash] params ({})
|
2205
|
+
def manage_propertygraph_statistics(params = {}, options = {})
|
2206
|
+
req = build_request(:manage_propertygraph_statistics, params)
|
2207
|
+
req.send_request(options)
|
2208
|
+
end
|
2209
|
+
|
2210
|
+
# Manages the generation and use of RDF graph statistics.
|
2211
|
+
#
|
2212
|
+
# @option params [String] :mode
|
2213
|
+
# The statistics generation mode. One of: `DISABLE_AUTOCOMPUTE`,
|
2214
|
+
# `ENABLE_AUTOCOMPUTE`, or `REFRESH`, the last of which manually
|
2215
|
+
# triggers DFE statistics generation.
|
2216
|
+
#
|
2217
|
+
# @return [Types::ManageSparqlStatisticsOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2218
|
+
#
|
2219
|
+
# * {Types::ManageSparqlStatisticsOutput#status #status} => String
|
2220
|
+
# * {Types::ManageSparqlStatisticsOutput#payload #payload} => Types::RefreshStatisticsIdMap
|
2221
|
+
#
|
2222
|
+
# @example Request syntax with placeholder values
|
2223
|
+
#
|
2224
|
+
# resp = client.manage_sparql_statistics({
|
2225
|
+
# mode: "disableAutoCompute", # accepts disableAutoCompute, enableAutoCompute, refresh
|
2226
|
+
# })
|
2227
|
+
#
|
2228
|
+
# @example Response structure
|
2229
|
+
#
|
2230
|
+
# resp.status #=> String
|
2231
|
+
# resp.payload.statistics_id #=> String
|
2232
|
+
#
|
2233
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/ManageSparqlStatistics AWS API Documentation
|
2234
|
+
#
|
2235
|
+
# @overload manage_sparql_statistics(params = {})
|
2236
|
+
# @param [Hash] params ({})
|
2237
|
+
def manage_sparql_statistics(params = {}, options = {})
|
2238
|
+
req = build_request(:manage_sparql_statistics, params)
|
2239
|
+
req.send_request(options)
|
2240
|
+
end
|
2241
|
+
|
2242
|
+
# Starts a Neptune bulk loader job to load data from an Amazon S3 bucket
|
2243
|
+
# into a Neptune DB instance. See [Using the Amazon Neptune Bulk Loader
|
2244
|
+
# to Ingest Data][1].
|
2245
|
+
#
|
2246
|
+
#
|
2247
|
+
#
|
2248
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/bulk-load.html
|
2249
|
+
#
|
2250
|
+
# @option params [required, String] :source
|
2251
|
+
# The `source` parameter accepts an S3 URI that identifies a single
|
2252
|
+
# file, multiple files, a folder, or multiple folders. Neptune loads
|
2253
|
+
# every data file in any folder that is specified.
|
2254
|
+
#
|
2255
|
+
# The URI can be in any of the following formats.
|
2256
|
+
#
|
2257
|
+
# * `s3://(bucket_name)/(object-key-name)`
|
2258
|
+
#
|
2259
|
+
# * `https://s3.amazonaws.com/(bucket_name)/(object-key-name)`
|
2260
|
+
#
|
2261
|
+
# * `https://s3.us-east-1.amazonaws.com/(bucket_name)/(object-key-name)`
|
2262
|
+
#
|
2263
|
+
# The `object-key-name` element of the URI is equivalent to the
|
2264
|
+
# [prefix][1] parameter in an S3 [ListObjects][2] API call. It
|
2265
|
+
# identifies all the objects in the specified S3 bucket whose names
|
2266
|
+
# begin with that prefix. That can be a single file or folder, or
|
2267
|
+
# multiple files and/or folders.
|
2268
|
+
#
|
2269
|
+
# The specified folder or folders can contain multiple vertex files and
|
2270
|
+
# multiple edge files.
|
2271
|
+
#
|
2272
|
+
#
|
2273
|
+
#
|
2274
|
+
# [1]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_ListObjects.html#API_ListObjects_RequestParameters
|
2275
|
+
# [2]: https://docs.aws.amazon.com/AmazonS3/latest/API/API_ListObjects.html
|
2276
|
+
#
|
2277
|
+
# @option params [required, String] :format
|
2278
|
+
# The format of the data. For more information about data formats for
|
2279
|
+
# the Neptune `Loader` command, see [Load Data Formats][1].
|
2280
|
+
#
|
2281
|
+
# **Allowed values**
|
2282
|
+
#
|
2283
|
+
# * <b> <code>csv</code> </b> for the [Gremlin CSV data format][2].
|
2284
|
+
#
|
2285
|
+
# * <b> <code>opencypher</code> </b> for the [openCypher CSV data
|
2286
|
+
# format][3].
|
2287
|
+
#
|
2288
|
+
# * <b> <code>ntriples</code> </b> for the [N-Triples RDF data
|
2289
|
+
# format][4].
|
2290
|
+
#
|
2291
|
+
# * <b> <code>nquads</code> </b> for the [N-Quads RDF data format][5].
|
2292
|
+
#
|
2293
|
+
# * <b> <code>rdfxml</code> </b> for the [RDF\\XML RDF data format][6].
|
2294
|
+
#
|
2295
|
+
# * <b> <code>turtle</code> </b> for the [Turtle RDF data format][7].
|
2296
|
+
#
|
2297
|
+
#
|
2298
|
+
#
|
2299
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/bulk-load-tutorial-format.html
|
2300
|
+
# [2]: https://docs.aws.amazon.com/neptune/latest/userguide/bulk-load-tutorial-format-gremlin.html
|
2301
|
+
# [3]: https://docs.aws.amazon.com/neptune/latest/userguide/bulk-load-tutorial-format-opencypher.html
|
2302
|
+
# [4]: https://www.w3.org/TR/n-triples/
|
2303
|
+
# [5]: https://www.w3.org/TR/n-quads/
|
2304
|
+
# [6]: https://www.w3.org/TR/rdf-syntax-grammar/
|
2305
|
+
# [7]: https://www.w3.org/TR/turtle/
|
2306
|
+
#
|
2307
|
+
# @option params [required, String] :s3_bucket_region
|
2308
|
+
# The Amazon region of the S3 bucket. This must match the Amazon Region
|
2309
|
+
# of the DB cluster.
|
2310
|
+
#
|
2311
|
+
# @option params [required, String] :iam_role_arn
|
2312
|
+
# The Amazon Resource Name (ARN) for an IAM role to be assumed by the
|
2313
|
+
# Neptune DB instance for access to the S3 bucket. The IAM role ARN
|
2314
|
+
# provided here should be attached to the DB cluster (see [Adding the
|
2315
|
+
# IAM Role to an Amazon Neptune Cluster][1].
|
2316
|
+
#
|
2317
|
+
#
|
2318
|
+
#
|
2319
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/bulk-load-tutorial-IAM-add-role-cluster.html
|
2320
|
+
#
|
2321
|
+
# @option params [String] :mode
|
2322
|
+
# The load job mode.
|
2323
|
+
#
|
2324
|
+
# *Allowed values*: `RESUME`, `NEW`, `AUTO`.
|
2325
|
+
#
|
2326
|
+
# *Default value*: `AUTO`.
|
2327
|
+
#
|
2328
|
+
# ****
|
2329
|
+
#
|
2330
|
+
# * `RESUME` – In RESUME mode, the loader looks for a previous load
|
2331
|
+
# from this source, and if it finds one, resumes that load job. If no
|
2332
|
+
# previous load job is found, the loader stops.
|
2333
|
+
#
|
2334
|
+
# The loader avoids reloading files that were successfully loaded in a
|
2335
|
+
# previous job. It only tries to process failed files. If you dropped
|
2336
|
+
# previously loaded data from your Neptune cluster, that data is not
|
2337
|
+
# reloaded in this mode. If a previous load job loaded all files from
|
2338
|
+
# the same source successfully, nothing is reloaded, and the loader
|
2339
|
+
# returns success.
|
2340
|
+
#
|
2341
|
+
# * `NEW` – In NEW mode, the creates a new load request regardless
|
2342
|
+
# of any previous loads. You can use this mode to reload all the data
|
2343
|
+
# from a source after dropping previously loaded data from your
|
2344
|
+
# Neptune cluster, or to load new data available at the same source.
|
2345
|
+
#
|
2346
|
+
# * `AUTO` – In AUTO mode, the loader looks for a previous load job
|
2347
|
+
# from the same source, and if it finds one, resumes that job, just as
|
2348
|
+
# in `RESUME` mode.
|
2349
|
+
#
|
2350
|
+
# If the loader doesn't find a previous load job from the same
|
2351
|
+
# source, it loads all data from the source, just as in `NEW` mode.
|
2352
|
+
#
|
2353
|
+
# @option params [Boolean] :fail_on_error
|
2354
|
+
# <b> <code>failOnError</code> </b> – A flag to toggle a complete
|
2355
|
+
# stop on an error.
|
2356
|
+
#
|
2357
|
+
# *Allowed values*: `"TRUE"`, `"FALSE"`.
|
2358
|
+
#
|
2359
|
+
# *Default value*: `"TRUE"`.
|
2360
|
+
#
|
2361
|
+
# When this parameter is set to `"FALSE"`, the loader tries to load all
|
2362
|
+
# the data in the location specified, skipping any entries with errors.
|
2363
|
+
#
|
2364
|
+
# When this parameter is set to `"TRUE"`, the loader stops as soon as it
|
2365
|
+
# encounters an error. Data loaded up to that point persists.
|
2366
|
+
#
|
2367
|
+
# @option params [String] :parallelism
|
2368
|
+
# The optional `parallelism` parameter can be set to reduce the number
|
2369
|
+
# of threads used by the bulk load process.
|
2370
|
+
#
|
2371
|
+
# *Allowed values*:
|
2372
|
+
#
|
2373
|
+
# * `LOW` – The number of threads used is the number of available
|
2374
|
+
# vCPUs divided by 8.
|
2375
|
+
#
|
2376
|
+
# * `MEDIUM` – The number of threads used is the number of available
|
2377
|
+
# vCPUs divided by 2.
|
2378
|
+
#
|
2379
|
+
# * `HIGH` – The number of threads used is the same as the number of
|
2380
|
+
# available vCPUs.
|
2381
|
+
#
|
2382
|
+
# * `OVERSUBSCRIBE` – The number of threads used is the number of
|
2383
|
+
# available vCPUs multiplied by 2. If this value is used, the bulk
|
2384
|
+
# loader takes up all available resources.
|
2385
|
+
#
|
2386
|
+
# This does not mean, however, that the `OVERSUBSCRIBE` setting
|
2387
|
+
# results in 100% CPU utilization. Because the load operation is I/O
|
2388
|
+
# bound, the highest CPU utilization to expect is in the 60% to 70%
|
2389
|
+
# range.
|
2390
|
+
#
|
2391
|
+
# *Default value*: `HIGH`
|
2392
|
+
#
|
2393
|
+
# The `parallelism` setting can sometimes result in a deadlock between
|
2394
|
+
# threads when loading openCypher data. When this happens, Neptune
|
2395
|
+
# returns the `LOAD_DATA_DEADLOCK` error. You can generally fix the
|
2396
|
+
# issue by setting `parallelism` to a lower setting and retrying the
|
2397
|
+
# load command.
|
2398
|
+
#
|
2399
|
+
# @option params [Hash<String,String>] :parser_configuration
|
2400
|
+
# <b> <code>parserConfiguration</code> </b> – An optional object
|
2401
|
+
# with additional parser configuration values. Each of the child
|
2402
|
+
# parameters is also optional:
|
2403
|
+
#
|
2404
|
+
# ****
|
2405
|
+
#
|
2406
|
+
# * <b> <code>namedGraphUri</code> </b> – The default graph for all
|
2407
|
+
# RDF formats when no graph is specified (for non-quads formats and
|
2408
|
+
# NQUAD entries with no graph).
|
2409
|
+
#
|
2410
|
+
# The default is
|
2411
|
+
# `https://aws.amazon.com/neptune/vocab/v01/DefaultNamedGraph`.
|
2412
|
+
#
|
2413
|
+
# * <b> <code>baseUri</code> </b> – The base URI for RDF/XML and
|
2414
|
+
# Turtle formats.
|
2415
|
+
#
|
2416
|
+
# The default is `https://aws.amazon.com/neptune/default`.
|
2417
|
+
#
|
2418
|
+
# * <b> <code>allowEmptyStrings</code> </b> – Gremlin users need to
|
2419
|
+
# be able to pass empty string values("") as node and edge
|
2420
|
+
# properties when loading CSV data. If `allowEmptyStrings` is set to
|
2421
|
+
# `false` (the default), such empty strings are treated as nulls and
|
2422
|
+
# are not loaded.
|
2423
|
+
#
|
2424
|
+
# If `allowEmptyStrings` is set to `true`, the loader treats empty
|
2425
|
+
# strings as valid property values and loads them accordingly.
|
2426
|
+
#
|
2427
|
+
# @option params [Boolean] :update_single_cardinality_properties
|
2428
|
+
# `updateSingleCardinalityProperties` is an optional parameter that
|
2429
|
+
# controls how the bulk loader treats a new value for single-cardinality
|
2430
|
+
# vertex or edge properties. This is not supported for loading
|
2431
|
+
# openCypher data.
|
2432
|
+
#
|
2433
|
+
# *Allowed values*: `"TRUE"`, `"FALSE"`.
|
2434
|
+
#
|
2435
|
+
# *Default value*: `"FALSE"`.
|
2436
|
+
#
|
2437
|
+
# By default, or when `updateSingleCardinalityProperties` is explicitly
|
2438
|
+
# set to `"FALSE"`, the loader treats a new value as an error, because
|
2439
|
+
# it violates single cardinality.
|
2440
|
+
#
|
2441
|
+
# When `updateSingleCardinalityProperties` is set to `"TRUE"`, on the
|
2442
|
+
# other hand, the bulk loader replaces the existing value with the new
|
2443
|
+
# one. If multiple edge or single-cardinality vertex property values are
|
2444
|
+
# provided in the source file(s) being loaded, the final value at the
|
2445
|
+
# end of the bulk load could be any one of those new values. The loader
|
2446
|
+
# only guarantees that the existing value has been replaced by one of
|
2447
|
+
# the new ones.
|
2448
|
+
#
|
2449
|
+
# @option params [Boolean] :queue_request
|
2450
|
+
# This is an optional flag parameter that indicates whether the load
|
2451
|
+
# request can be queued up or not.
|
2452
|
+
#
|
2453
|
+
# You don't have to wait for one load job to complete before issuing
|
2454
|
+
# the next one, because Neptune can queue up as many as 64 jobs at a
|
2455
|
+
# time, provided that their `queueRequest` parameters are all set to
|
2456
|
+
# `"TRUE"`.
|
2457
|
+
#
|
2458
|
+
# If the `queueRequest` parameter is omitted or set to `"FALSE"`, the
|
2459
|
+
# load request will fail if another load job is already running.
|
2460
|
+
#
|
2461
|
+
# *Allowed values*: `"TRUE"`, `"FALSE"`.
|
2462
|
+
#
|
2463
|
+
# *Default value*: `"FALSE"`.
|
2464
|
+
#
|
2465
|
+
# @option params [Array<String>] :dependencies
|
2466
|
+
# This is an optional parameter that can make a queued load request
|
2467
|
+
# contingent on the successful completion of one or more previous jobs
|
2468
|
+
# in the queue.
|
2469
|
+
#
|
2470
|
+
# Neptune can queue up as many as 64 load requests at a time, if their
|
2471
|
+
# `queueRequest` parameters are set to `"TRUE"`. The `dependencies`
|
2472
|
+
# parameter lets you make execution of such a queued request dependent
|
2473
|
+
# on the successful completion of one or more specified previous
|
2474
|
+
# requests in the queue.
|
2475
|
+
#
|
2476
|
+
# For example, if load `Job-A` and `Job-B` are independent of each
|
2477
|
+
# other, but load `Job-C` needs `Job-A` and `Job-B` to be finished
|
2478
|
+
# before it begins, proceed as follows:
|
2479
|
+
#
|
2480
|
+
# 1. Submit `load-job-A` and `load-job-B` one after another in any
|
2481
|
+
# order, and save their load-ids.
|
2482
|
+
#
|
2483
|
+
# 2. Submit `load-job-C` with the load-ids of the two jobs in its
|
2484
|
+
# `dependencies` field:
|
2485
|
+
#
|
2486
|
+
# Because of the `dependencies` parameter, the bulk loader will not
|
2487
|
+
# start `Job-C` until `Job-A` and `Job-B` have completed successfully.
|
2488
|
+
# If either one of them fails, Job-C will not be executed, and its
|
2489
|
+
# status will be set to `LOAD_FAILED_BECAUSE_DEPENDENCY_NOT_SATISFIED`.
|
2490
|
+
#
|
2491
|
+
# You can set up multiple levels of dependency in this way, so that the
|
2492
|
+
# failure of one job will cause all requests that are directly or
|
2493
|
+
# indirectly dependent on it to be cancelled.
|
2494
|
+
#
|
2495
|
+
# @option params [Boolean] :user_provided_edge_ids
|
2496
|
+
# This parameter is required only when loading openCypher data that
|
2497
|
+
# contains relationship IDs. It must be included and set to `True` when
|
2498
|
+
# openCypher relationship IDs are explicitly provided in the load data
|
2499
|
+
# (recommended).
|
2500
|
+
#
|
2501
|
+
# When `userProvidedEdgeIds` is absent or set to `True`, an `:ID` column
|
2502
|
+
# must be present in every relationship file in the load.
|
2503
|
+
#
|
2504
|
+
# When `userProvidedEdgeIds` is present and set to `False`, relationship
|
2505
|
+
# files in the load **must not** contain an `:ID` column. Instead, the
|
2506
|
+
# Neptune loader automatically generates an ID for each relationship.
|
2507
|
+
#
|
2508
|
+
# It's useful to provide relationship IDs explicitly so that the loader
|
2509
|
+
# can resume loading after error in the CSV data have been fixed,
|
2510
|
+
# without having to reload any relationships that have already been
|
2511
|
+
# loaded. If relationship IDs have not been explicitly assigned, the
|
2512
|
+
# loader cannot resume a failed load if any relationship file has had to
|
2513
|
+
# be corrected, and must instead reload all the relationships.
|
2514
|
+
#
|
2515
|
+
# @return [Types::StartLoaderJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2516
|
+
#
|
2517
|
+
# * {Types::StartLoaderJobOutput#status #status} => String
|
2518
|
+
# * {Types::StartLoaderJobOutput#payload #payload} => Hash<String,String>
|
2519
|
+
#
|
2520
|
+
# @example Request syntax with placeholder values
|
2521
|
+
#
|
2522
|
+
# resp = client.start_loader_job({
|
2523
|
+
# source: "String", # required
|
2524
|
+
# format: "csv", # required, accepts csv, opencypher, ntriples, nquads, rdfxml, turtle
|
2525
|
+
# s3_bucket_region: "us-east-1", # required, accepts us-east-1, us-east-2, us-west-1, us-west-2, ca-central-1, sa-east-1, eu-north-1, eu-west-1, eu-west-2, eu-west-3, eu-central-1, me-south-1, af-south-1, ap-east-1, ap-northeast-1, ap-northeast-2, ap-southeast-1, ap-southeast-2, ap-south-1, cn-north-1, cn-northwest-1, us-gov-west-1, us-gov-east-1
|
2526
|
+
# iam_role_arn: "String", # required
|
2527
|
+
# mode: "RESUME", # accepts RESUME, NEW, AUTO
|
2528
|
+
# fail_on_error: false,
|
2529
|
+
# parallelism: "LOW", # accepts LOW, MEDIUM, HIGH, OVERSUBSCRIBE
|
2530
|
+
# parser_configuration: {
|
2531
|
+
# "String" => "String",
|
2532
|
+
# },
|
2533
|
+
# update_single_cardinality_properties: false,
|
2534
|
+
# queue_request: false,
|
2535
|
+
# dependencies: ["String"],
|
2536
|
+
# user_provided_edge_ids: false,
|
2537
|
+
# })
|
2538
|
+
#
|
2539
|
+
# @example Response structure
|
2540
|
+
#
|
2541
|
+
# resp.status #=> String
|
2542
|
+
# resp.payload #=> Hash
|
2543
|
+
# resp.payload["String"] #=> String
|
2544
|
+
#
|
2545
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/StartLoaderJob AWS API Documentation
|
2546
|
+
#
|
2547
|
+
# @overload start_loader_job(params = {})
|
2548
|
+
# @param [Hash] params ({})
|
2549
|
+
def start_loader_job(params = {}, options = {})
|
2550
|
+
req = build_request(:start_loader_job, params)
|
2551
|
+
req.send_request(options)
|
2552
|
+
end
|
2553
|
+
|
2554
|
+
# Creates a new Neptune ML data processing job for processing the graph
|
2555
|
+
# data exported from Neptune for training. See [The `dataprocessing`
|
2556
|
+
# command][1].
|
2557
|
+
#
|
2558
|
+
#
|
2559
|
+
#
|
2560
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-dataprocessing.html
|
2561
|
+
#
|
2562
|
+
# @option params [String] :id
|
2563
|
+
# A unique identifier for the new job. The default is an autogenerated
|
2564
|
+
# UUID.
|
2565
|
+
#
|
2566
|
+
# @option params [String] :previous_data_processing_job_id
|
2567
|
+
# The job ID of a completed data processing job run on an earlier
|
2568
|
+
# version of the data.
|
2569
|
+
#
|
2570
|
+
# @option params [required, String] :input_data_s3_location
|
2571
|
+
# The URI of the Amazon S3 location where you want SageMaker to download
|
2572
|
+
# the data needed to run the data processing job.
|
2573
|
+
#
|
2574
|
+
# @option params [required, String] :processed_data_s3_location
|
2575
|
+
# The URI of the Amazon S3 location where you want SageMaker to save the
|
2576
|
+
# results of a data processing job.
|
2577
|
+
#
|
2578
|
+
# @option params [String] :sagemaker_iam_role_arn
|
2579
|
+
# The ARN of an IAM role for SageMaker execution. This must be listed in
|
2580
|
+
# your DB cluster parameter group or an error will occur.
|
2581
|
+
#
|
2582
|
+
# @option params [String] :neptune_iam_role_arn
|
2583
|
+
# The Amazon Resource Name (ARN) of an IAM role that SageMaker can
|
2584
|
+
# assume to perform tasks on your behalf. This must be listed in your DB
|
2585
|
+
# cluster parameter group or an error will occur.
|
2586
|
+
#
|
2587
|
+
# @option params [String] :processing_instance_type
|
2588
|
+
# The type of ML instance used during data processing. Its memory should
|
2589
|
+
# be large enough to hold the processed dataset. The default is the
|
2590
|
+
# smallest ml.r5 type whose memory is ten times larger than the size of
|
2591
|
+
# the exported graph data on disk.
|
2592
|
+
#
|
2593
|
+
# @option params [Integer] :processing_instance_volume_size_in_gb
|
2594
|
+
# The disk volume size of the processing instance. Both input data and
|
2595
|
+
# processed data are stored on disk, so the volume size must be large
|
2596
|
+
# enough to hold both data sets. The default is 0. If not specified or
|
2597
|
+
# 0, Neptune ML chooses the volume size automatically based on the data
|
2598
|
+
# size.
|
2599
|
+
#
|
2600
|
+
# @option params [Integer] :processing_time_out_in_seconds
|
2601
|
+
# Timeout in seconds for the data processing job. The default is 86,400
|
2602
|
+
# (1 day).
|
2603
|
+
#
|
2604
|
+
# @option params [String] :model_type
|
2605
|
+
# One of the two model types that Neptune ML currently supports:
|
2606
|
+
# heterogeneous graph models (`heterogeneous`), and knowledge graph
|
2607
|
+
# (`kge`). The default is none. If not specified, Neptune ML chooses the
|
2608
|
+
# model type automatically based on the data.
|
2609
|
+
#
|
2610
|
+
# @option params [String] :config_file_name
|
2611
|
+
# A data specification file that describes how to load the exported
|
2612
|
+
# graph data for training. The file is automatically generated by the
|
2613
|
+
# Neptune export toolkit. The default is
|
2614
|
+
# `training-data-configuration.json`.
|
2615
|
+
#
|
2616
|
+
# @option params [Array<String>] :subnets
|
2617
|
+
# The IDs of the subnets in the Neptune VPC. The default is None.
|
2618
|
+
#
|
2619
|
+
# @option params [Array<String>] :security_group_ids
|
2620
|
+
# The VPC security group IDs. The default is None.
|
2621
|
+
#
|
2622
|
+
# @option params [String] :volume_encryption_kms_key
|
2623
|
+
# The Amazon Key Management Service (Amazon KMS) key that SageMaker uses
|
2624
|
+
# to encrypt data on the storage volume attached to the ML compute
|
2625
|
+
# instances that run the training job. The default is None.
|
2626
|
+
#
|
2627
|
+
# @option params [String] :s3_output_encryption_kms_key
|
2628
|
+
# The Amazon Key Management Service (Amazon KMS) key that SageMaker uses
|
2629
|
+
# to encrypt the output of the processing job. The default is none.
|
2630
|
+
#
|
2631
|
+
# @return [Types::StartMLDataProcessingJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2632
|
+
#
|
2633
|
+
# * {Types::StartMLDataProcessingJobOutput#id #id} => String
|
2634
|
+
# * {Types::StartMLDataProcessingJobOutput#arn #arn} => String
|
2635
|
+
# * {Types::StartMLDataProcessingJobOutput#creation_time_in_millis #creation_time_in_millis} => Integer
|
2636
|
+
#
|
2637
|
+
# @example Request syntax with placeholder values
|
2638
|
+
#
|
2639
|
+
# resp = client.start_ml_data_processing_job({
|
2640
|
+
# id: "String",
|
2641
|
+
# previous_data_processing_job_id: "String",
|
2642
|
+
# input_data_s3_location: "String", # required
|
2643
|
+
# processed_data_s3_location: "String", # required
|
2644
|
+
# sagemaker_iam_role_arn: "String",
|
2645
|
+
# neptune_iam_role_arn: "String",
|
2646
|
+
# processing_instance_type: "String",
|
2647
|
+
# processing_instance_volume_size_in_gb: 1,
|
2648
|
+
# processing_time_out_in_seconds: 1,
|
2649
|
+
# model_type: "String",
|
2650
|
+
# config_file_name: "String",
|
2651
|
+
# subnets: ["String"],
|
2652
|
+
# security_group_ids: ["String"],
|
2653
|
+
# volume_encryption_kms_key: "String",
|
2654
|
+
# s3_output_encryption_kms_key: "String",
|
2655
|
+
# })
|
2656
|
+
#
|
2657
|
+
# @example Response structure
|
2658
|
+
#
|
2659
|
+
# resp.id #=> String
|
2660
|
+
# resp.arn #=> String
|
2661
|
+
# resp.creation_time_in_millis #=> Integer
|
2662
|
+
#
|
2663
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/StartMLDataProcessingJob AWS API Documentation
|
2664
|
+
#
|
2665
|
+
# @overload start_ml_data_processing_job(params = {})
|
2666
|
+
# @param [Hash] params ({})
|
2667
|
+
def start_ml_data_processing_job(params = {}, options = {})
|
2668
|
+
req = build_request(:start_ml_data_processing_job, params)
|
2669
|
+
req.send_request(options)
|
2670
|
+
end
|
2671
|
+
|
2672
|
+
# Creates a new Neptune ML model training job. See [Model training using
|
2673
|
+
# the `modeltraining` command][1].
|
2674
|
+
#
|
2675
|
+
#
|
2676
|
+
#
|
2677
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-api-modeltraining.html
|
2678
|
+
#
|
2679
|
+
# @option params [String] :id
|
2680
|
+
# A unique identifier for the new job. The default is An autogenerated
|
2681
|
+
# UUID.
|
2682
|
+
#
|
2683
|
+
# @option params [String] :previous_model_training_job_id
|
2684
|
+
# The job ID of a completed model-training job that you want to update
|
2685
|
+
# incrementally based on updated data.
|
2686
|
+
#
|
2687
|
+
# @option params [required, String] :data_processing_job_id
|
2688
|
+
# The job ID of the completed data-processing job that has created the
|
2689
|
+
# data that the training will work with.
|
2690
|
+
#
|
2691
|
+
# @option params [required, String] :train_model_s3_location
|
2692
|
+
# The location in Amazon S3 where the model artifacts are to be stored.
|
2693
|
+
#
|
2694
|
+
# @option params [String] :sagemaker_iam_role_arn
|
2695
|
+
# The ARN of an IAM role for SageMaker execution.This must be listed in
|
2696
|
+
# your DB cluster parameter group or an error will occur.
|
2697
|
+
#
|
2698
|
+
# @option params [String] :neptune_iam_role_arn
|
2699
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
2700
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
2701
|
+
# group or an error will occur.
|
2702
|
+
#
|
2703
|
+
# @option params [String] :base_processing_instance_type
|
2704
|
+
# The type of ML instance used in preparing and managing training of ML
|
2705
|
+
# models. This is a CPU instance chosen based on memory requirements for
|
2706
|
+
# processing the training data and model.
|
2707
|
+
#
|
2708
|
+
# @option params [String] :training_instance_type
|
2709
|
+
# The type of ML instance used for model training. All Neptune ML models
|
2710
|
+
# support CPU, GPU, and multiGPU training. The default is
|
2711
|
+
# `ml.p3.2xlarge`. Choosing the right instance type for training depends
|
2712
|
+
# on the task type, graph size, and your budget.
|
2713
|
+
#
|
2714
|
+
# @option params [Integer] :training_instance_volume_size_in_gb
|
2715
|
+
# The disk volume size of the training instance. Both input data and the
|
2716
|
+
# output model are stored on disk, so the volume size must be large
|
2717
|
+
# enough to hold both data sets. The default is 0. If not specified or
|
2718
|
+
# 0, Neptune ML selects a disk volume size based on the recommendation
|
2719
|
+
# generated in the data processing step.
|
2720
|
+
#
|
2721
|
+
# @option params [Integer] :training_time_out_in_seconds
|
2722
|
+
# Timeout in seconds for the training job. The default is 86,400 (1
|
2723
|
+
# day).
|
2724
|
+
#
|
2725
|
+
# @option params [Integer] :max_hpo_number_of_training_jobs
|
2726
|
+
# Maximum total number of training jobs to start for the hyperparameter
|
2727
|
+
# tuning job. The default is 2. Neptune ML automatically tunes the
|
2728
|
+
# hyperparameters of the machine learning model. To obtain a model that
|
2729
|
+
# performs well, use at least 10 jobs (in other words, set
|
2730
|
+
# `maxHPONumberOfTrainingJobs` to 10). In general, the more tuning runs,
|
2731
|
+
# the better the results.
|
2732
|
+
#
|
2733
|
+
# @option params [Integer] :max_hpo_parallel_training_jobs
|
2734
|
+
# Maximum number of parallel training jobs to start for the
|
2735
|
+
# hyperparameter tuning job. The default is 2. The number of parallel
|
2736
|
+
# jobs you can run is limited by the available resources on your
|
2737
|
+
# training instance.
|
2738
|
+
#
|
2739
|
+
# @option params [Array<String>] :subnets
|
2740
|
+
# The IDs of the subnets in the Neptune VPC. The default is None.
|
2741
|
+
#
|
2742
|
+
# @option params [Array<String>] :security_group_ids
|
2743
|
+
# The VPC security group IDs. The default is None.
|
2744
|
+
#
|
2745
|
+
# @option params [String] :volume_encryption_kms_key
|
2746
|
+
# The Amazon Key Management Service (KMS) key that SageMaker uses to
|
2747
|
+
# encrypt data on the storage volume attached to the ML compute
|
2748
|
+
# instances that run the training job. The default is None.
|
2749
|
+
#
|
2750
|
+
# @option params [String] :s3_output_encryption_kms_key
|
2751
|
+
# The Amazon Key Management Service (KMS) key that SageMaker uses to
|
2752
|
+
# encrypt the output of the processing job. The default is none.
|
2753
|
+
#
|
2754
|
+
# @option params [Boolean] :enable_managed_spot_training
|
2755
|
+
# Optimizes the cost of training machine-learning models by using Amazon
|
2756
|
+
# Elastic Compute Cloud spot instances. The default is `False`.
|
2757
|
+
#
|
2758
|
+
# @option params [Types::CustomModelTrainingParameters] :custom_model_training_parameters
|
2759
|
+
# The configuration for custom model training. This is a JSON object.
|
2760
|
+
#
|
2761
|
+
# @return [Types::StartMLModelTrainingJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2762
|
+
#
|
2763
|
+
# * {Types::StartMLModelTrainingJobOutput#id #id} => String
|
2764
|
+
# * {Types::StartMLModelTrainingJobOutput#arn #arn} => String
|
2765
|
+
# * {Types::StartMLModelTrainingJobOutput#creation_time_in_millis #creation_time_in_millis} => Integer
|
2766
|
+
#
|
2767
|
+
# @example Request syntax with placeholder values
|
2768
|
+
#
|
2769
|
+
# resp = client.start_ml_model_training_job({
|
2770
|
+
# id: "String",
|
2771
|
+
# previous_model_training_job_id: "String",
|
2772
|
+
# data_processing_job_id: "String", # required
|
2773
|
+
# train_model_s3_location: "String", # required
|
2774
|
+
# sagemaker_iam_role_arn: "String",
|
2775
|
+
# neptune_iam_role_arn: "String",
|
2776
|
+
# base_processing_instance_type: "String",
|
2777
|
+
# training_instance_type: "String",
|
2778
|
+
# training_instance_volume_size_in_gb: 1,
|
2779
|
+
# training_time_out_in_seconds: 1,
|
2780
|
+
# max_hpo_number_of_training_jobs: 1,
|
2781
|
+
# max_hpo_parallel_training_jobs: 1,
|
2782
|
+
# subnets: ["String"],
|
2783
|
+
# security_group_ids: ["String"],
|
2784
|
+
# volume_encryption_kms_key: "String",
|
2785
|
+
# s3_output_encryption_kms_key: "String",
|
2786
|
+
# enable_managed_spot_training: false,
|
2787
|
+
# custom_model_training_parameters: {
|
2788
|
+
# source_s3_directory_path: "String", # required
|
2789
|
+
# training_entry_point_script: "String",
|
2790
|
+
# transform_entry_point_script: "String",
|
2791
|
+
# },
|
2792
|
+
# })
|
2793
|
+
#
|
2794
|
+
# @example Response structure
|
2795
|
+
#
|
2796
|
+
# resp.id #=> String
|
2797
|
+
# resp.arn #=> String
|
2798
|
+
# resp.creation_time_in_millis #=> Integer
|
2799
|
+
#
|
2800
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/StartMLModelTrainingJob AWS API Documentation
|
2801
|
+
#
|
2802
|
+
# @overload start_ml_model_training_job(params = {})
|
2803
|
+
# @param [Hash] params ({})
|
2804
|
+
def start_ml_model_training_job(params = {}, options = {})
|
2805
|
+
req = build_request(:start_ml_model_training_job, params)
|
2806
|
+
req.send_request(options)
|
2807
|
+
end
|
2808
|
+
|
2809
|
+
# Creates a new model transform job. See [Use a trained model to
|
2810
|
+
# generate new model artifacts][1].
|
2811
|
+
#
|
2812
|
+
#
|
2813
|
+
#
|
2814
|
+
# [1]: https://docs.aws.amazon.com/neptune/latest/userguide/machine-learning-model-transform.html
|
2815
|
+
#
|
2816
|
+
# @option params [String] :id
|
2817
|
+
# A unique identifier for the new job. The default is an autogenerated
|
2818
|
+
# UUID.
|
2819
|
+
#
|
2820
|
+
# @option params [String] :data_processing_job_id
|
2821
|
+
# The job ID of a completed data-processing job. You must include either
|
2822
|
+
# `dataProcessingJobId` and a `mlModelTrainingJobId`, or a
|
2823
|
+
# `trainingJobName`.
|
2824
|
+
#
|
2825
|
+
# @option params [String] :ml_model_training_job_id
|
2826
|
+
# The job ID of a completed model-training job. You must include either
|
2827
|
+
# `dataProcessingJobId` and a `mlModelTrainingJobId`, or a
|
2828
|
+
# `trainingJobName`.
|
2829
|
+
#
|
2830
|
+
# @option params [String] :training_job_name
|
2831
|
+
# The name of a completed SageMaker training job. You must include
|
2832
|
+
# either `dataProcessingJobId` and a `mlModelTrainingJobId`, or a
|
2833
|
+
# `trainingJobName`.
|
2834
|
+
#
|
2835
|
+
# @option params [required, String] :model_transform_output_s3_location
|
2836
|
+
# The location in Amazon S3 where the model artifacts are to be stored.
|
2837
|
+
#
|
2838
|
+
# @option params [String] :sagemaker_iam_role_arn
|
2839
|
+
# The ARN of an IAM role for SageMaker execution. This must be listed in
|
2840
|
+
# your DB cluster parameter group or an error will occur.
|
2841
|
+
#
|
2842
|
+
# @option params [String] :neptune_iam_role_arn
|
2843
|
+
# The ARN of an IAM role that provides Neptune access to SageMaker and
|
2844
|
+
# Amazon S3 resources. This must be listed in your DB cluster parameter
|
2845
|
+
# group or an error will occur.
|
2846
|
+
#
|
2847
|
+
# @option params [Types::CustomModelTransformParameters] :custom_model_transform_parameters
|
2848
|
+
# Configuration information for a model transform using a custom model.
|
2849
|
+
# The `customModelTransformParameters` object contains the following
|
2850
|
+
# fields, which must have values compatible with the saved model
|
2851
|
+
# parameters from the training job:
|
2852
|
+
#
|
2853
|
+
# @option params [String] :base_processing_instance_type
|
2854
|
+
# The type of ML instance used in preparing and managing training of ML
|
2855
|
+
# models. This is an ML compute instance chosen based on memory
|
2856
|
+
# requirements for processing the training data and model.
|
2857
|
+
#
|
2858
|
+
# @option params [Integer] :base_processing_instance_volume_size_in_gb
|
2859
|
+
# The disk volume size of the training instance in gigabytes. The
|
2860
|
+
# default is 0. Both input data and the output model are stored on disk,
|
2861
|
+
# so the volume size must be large enough to hold both data sets. If not
|
2862
|
+
# specified or 0, Neptune ML selects a disk volume size based on the
|
2863
|
+
# recommendation generated in the data processing step.
|
2864
|
+
#
|
2865
|
+
# @option params [Array<String>] :subnets
|
2866
|
+
# The IDs of the subnets in the Neptune VPC. The default is None.
|
2867
|
+
#
|
2868
|
+
# @option params [Array<String>] :security_group_ids
|
2869
|
+
# The VPC security group IDs. The default is None.
|
2870
|
+
#
|
2871
|
+
# @option params [String] :volume_encryption_kms_key
|
2872
|
+
# The Amazon Key Management Service (KMS) key that SageMaker uses to
|
2873
|
+
# encrypt data on the storage volume attached to the ML compute
|
2874
|
+
# instances that run the training job. The default is None.
|
2875
|
+
#
|
2876
|
+
# @option params [String] :s3_output_encryption_kms_key
|
2877
|
+
# The Amazon Key Management Service (KMS) key that SageMaker uses to
|
2878
|
+
# encrypt the output of the processing job. The default is none.
|
2879
|
+
#
|
2880
|
+
# @return [Types::StartMLModelTransformJobOutput] Returns a {Seahorse::Client::Response response} object which responds to the following methods:
|
2881
|
+
#
|
2882
|
+
# * {Types::StartMLModelTransformJobOutput#id #id} => String
|
2883
|
+
# * {Types::StartMLModelTransformJobOutput#arn #arn} => String
|
2884
|
+
# * {Types::StartMLModelTransformJobOutput#creation_time_in_millis #creation_time_in_millis} => Integer
|
2885
|
+
#
|
2886
|
+
# @example Request syntax with placeholder values
|
2887
|
+
#
|
2888
|
+
# resp = client.start_ml_model_transform_job({
|
2889
|
+
# id: "String",
|
2890
|
+
# data_processing_job_id: "String",
|
2891
|
+
# ml_model_training_job_id: "String",
|
2892
|
+
# training_job_name: "String",
|
2893
|
+
# model_transform_output_s3_location: "String", # required
|
2894
|
+
# sagemaker_iam_role_arn: "String",
|
2895
|
+
# neptune_iam_role_arn: "String",
|
2896
|
+
# custom_model_transform_parameters: {
|
2897
|
+
# source_s3_directory_path: "String", # required
|
2898
|
+
# transform_entry_point_script: "String",
|
2899
|
+
# },
|
2900
|
+
# base_processing_instance_type: "String",
|
2901
|
+
# base_processing_instance_volume_size_in_gb: 1,
|
2902
|
+
# subnets: ["String"],
|
2903
|
+
# security_group_ids: ["String"],
|
2904
|
+
# volume_encryption_kms_key: "String",
|
2905
|
+
# s3_output_encryption_kms_key: "String",
|
2906
|
+
# })
|
2907
|
+
#
|
2908
|
+
# @example Response structure
|
2909
|
+
#
|
2910
|
+
# resp.id #=> String
|
2911
|
+
# resp.arn #=> String
|
2912
|
+
# resp.creation_time_in_millis #=> Integer
|
2913
|
+
#
|
2914
|
+
# @see http://docs.aws.amazon.com/goto/WebAPI/neptunedata-2023-08-01/StartMLModelTransformJob AWS API Documentation
|
2915
|
+
#
|
2916
|
+
# @overload start_ml_model_transform_job(params = {})
|
2917
|
+
# @param [Hash] params ({})
|
2918
|
+
def start_ml_model_transform_job(params = {}, options = {})
|
2919
|
+
req = build_request(:start_ml_model_transform_job, params)
|
2920
|
+
req.send_request(options)
|
2921
|
+
end
|
2922
|
+
|
2923
|
+
# @!endgroup
|
2924
|
+
|
2925
|
+
# @param params ({})
|
2926
|
+
# @api private
|
2927
|
+
def build_request(operation_name, params = {})
|
2928
|
+
handlers = @handlers.for(operation_name)
|
2929
|
+
context = Seahorse::Client::RequestContext.new(
|
2930
|
+
operation_name: operation_name,
|
2931
|
+
operation: config.api.operation(operation_name),
|
2932
|
+
client: self,
|
2933
|
+
params: params,
|
2934
|
+
config: config)
|
2935
|
+
context[:gem_name] = 'aws-sdk-neptunedata'
|
2936
|
+
context[:gem_version] = '1.0.0'
|
2937
|
+
Seahorse::Client::Request.new(handlers, context)
|
2938
|
+
end
|
2939
|
+
|
2940
|
+
# @api private
|
2941
|
+
# @deprecated
|
2942
|
+
def waiter_names
|
2943
|
+
[]
|
2944
|
+
end
|
2945
|
+
|
2946
|
+
class << self
|
2947
|
+
|
2948
|
+
# @api private
|
2949
|
+
attr_reader :identifier
|
2950
|
+
|
2951
|
+
# @api private
|
2952
|
+
def errors_module
|
2953
|
+
Errors
|
2954
|
+
end
|
2955
|
+
|
2956
|
+
end
|
2957
|
+
end
|
2958
|
+
end
|