@aws-sdk/client-sagemaker 3.1028.0 → 3.1029.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +7 -0
- package/dist-cjs/index.js +16 -0
- package/dist-cjs/schemas/schemas_0.js +159 -99
- package/dist-es/SageMaker.js +2 -0
- package/dist-es/commands/StartClusterHealthCheckCommand.js +16 -0
- package/dist-es/commands/index.js +1 -0
- package/dist-es/models/enums.js +2 -0
- package/dist-es/schemas/schemas_0.js +124 -64
- package/dist-types/SageMaker.d.ts +7 -0
- package/dist-types/SageMakerClient.d.ts +3 -2
- package/dist-types/commands/BatchAddClusterNodesCommand.d.ts +19 -1
- package/dist-types/commands/CreateClusterCommand.d.ts +7 -2
- package/dist-types/commands/CreateCodeRepositoryCommand.d.ts +2 -1
- package/dist-types/commands/DescribeClusterCommand.d.ts +17 -2
- package/dist-types/commands/DescribeClusterNodeCommand.d.ts +2 -2
- package/dist-types/commands/DescribePipelineCommand.d.ts +2 -1
- package/dist-types/commands/DescribePipelineDefinitionForExecutionCommand.d.ts +1 -1
- package/dist-types/commands/DescribePipelineExecutionCommand.d.ts +1 -2
- package/dist-types/commands/ListPipelineExecutionStepsCommand.d.ts +2 -1
- package/dist-types/commands/ListPipelineParametersForExecutionCommand.d.ts +1 -2
- package/dist-types/commands/StartClusterHealthCheckCommand.d.ts +88 -0
- package/dist-types/commands/UpdateClusterCommand.d.ts +7 -2
- package/dist-types/commands/index.d.ts +1 -0
- package/dist-types/models/enums.d.ts +2 -0
- package/dist-types/models/models_0.d.ts +97 -61
- package/dist-types/models/models_1.d.ts +59 -35
- package/dist-types/models/models_2.d.ts +57 -118
- package/dist-types/models/models_3.d.ts +117 -114
- package/dist-types/models/models_4.d.ts +141 -4
- package/dist-types/schemas/schemas_0.d.ts +7 -0
- package/dist-types/ts3.4/SageMaker.d.ts +17 -0
- package/dist-types/ts3.4/SageMakerClient.d.ts +6 -0
- package/dist-types/ts3.4/commands/CreateCodeRepositoryCommand.d.ts +2 -4
- package/dist-types/ts3.4/commands/DescribePipelineCommand.d.ts +2 -4
- package/dist-types/ts3.4/commands/DescribePipelineDefinitionForExecutionCommand.d.ts +1 -1
- package/dist-types/ts3.4/commands/DescribePipelineExecutionCommand.d.ts +4 -2
- package/dist-types/ts3.4/commands/ListPipelineExecutionStepsCommand.d.ts +2 -4
- package/dist-types/ts3.4/commands/ListPipelineParametersForExecutionCommand.d.ts +4 -2
- package/dist-types/ts3.4/commands/StartClusterHealthCheckCommand.d.ts +51 -0
- package/dist-types/ts3.4/commands/index.d.ts +1 -0
- package/dist-types/ts3.4/models/enums.d.ts +2 -0
- package/dist-types/ts3.4/models/models_0.d.ts +24 -18
- package/dist-types/ts3.4/models/models_1.d.ts +18 -11
- package/dist-types/ts3.4/models/models_2.d.ts +16 -33
- package/dist-types/ts3.4/models/models_3.d.ts +28 -28
- package/dist-types/ts3.4/models/models_4.d.ts +40 -1
- package/dist-types/ts3.4/schemas/schemas_0.d.ts +7 -0
- package/package.json +1 -1
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
import type { ActionStatus, ActivationState, ActiveClusterOperationName, AdditionalS3DataSourceDataType, AggregationTransformationValue, AlgorithmStatus, AppInstanceType, AppStatus, AppType, ArtifactSourceIdType, AssemblyType, AssociationEdgeType, AsyncNotificationTopicTypes, AthenaResultCompressionType, AthenaResultFormat, AutoMLAlgorithm, AutoMLChannelType, AutoMLJobObjectiveType, AutoMLJobSecondaryStatus, AutoMLJobStatus, AutoMLMetricEnum, AutoMLMetricExtendedEnum, AutoMLMode, AutoMLProcessingUnit, AutoMLS3DataType, AutotuneMode, BatchAddClusterNodesErrorCode, BatchDeleteClusterNodesErrorCode, BatchRebootClusterNodesErrorCode, BatchReplaceClusterNodesErrorCode, BatchStrategy, CandidateStatus, CandidateStepType, CapacityReservationType, CapacitySizeType, CaptureMode, ClarifyFeatureType, ClarifyTextGranularity, ClarifyTextLanguage, ClusterAutoScalerType, ClusterAutoScalingMode, ClusterAutoScalingStatus, ClusterCapacityType, ClusterConfigMode, ClusterEventResourceType, ClusterInstanceStatus, ClusterInstanceType, ClusterKubernetesTaintEffect, ClusterNodeProvisioningMode, ClusterNodeRecovery, ClusterSlurmConfigStrategy, ClusterSlurmNodeType, ClusterStatus, CompilationJobStatus, CompleteOnConvergence, CompressionType, ConditionOutcome, ContainerMode, DataSourceName, DeepHealthCheckType, DetailedAlgorithmStatus, FairShare, FeatureStatus, FileSystemAccessMode, FileSystemType, FillingType,
|
|
1
|
+
import type { ActionStatus, ActivationState, ActiveClusterOperationName, AdditionalS3DataSourceDataType, AggregationTransformationValue, AlgorithmStatus, AppInstanceType, AppStatus, AppType, ArtifactSourceIdType, AssemblyType, AssociationEdgeType, AsyncNotificationTopicTypes, AthenaResultCompressionType, AthenaResultFormat, AutoMLAlgorithm, AutoMLChannelType, AutoMLJobObjectiveType, AutoMLJobSecondaryStatus, AutoMLJobStatus, AutoMLMetricEnum, AutoMLMetricExtendedEnum, AutoMLMode, AutoMLProcessingUnit, AutoMLS3DataType, AutotuneMode, BatchAddClusterNodesErrorCode, BatchDeleteClusterNodesErrorCode, BatchRebootClusterNodesErrorCode, BatchReplaceClusterNodesErrorCode, BatchStrategy, CandidateStatus, CandidateStepType, CapacityReservationType, CapacitySizeType, CaptureMode, ClarifyFeatureType, ClarifyTextGranularity, ClarifyTextLanguage, ClusterAutoScalerType, ClusterAutoScalingMode, ClusterAutoScalingStatus, ClusterCapacityType, ClusterConfigMode, ClusterEventResourceType, ClusterInstanceStatus, ClusterInstanceType, ClusterKubernetesTaintEffect, ClusterNodeProvisioningMode, ClusterNodeRecovery, ClusterSlurmConfigStrategy, ClusterSlurmNodeType, ClusterStatus, CompilationJobStatus, CompleteOnConvergence, CompressionType, ConditionOutcome, ContainerMode, DataSourceName, DeepHealthCheckType, DetailedAlgorithmStatus, FairShare, FeatureStatus, FileSystemAccessMode, FileSystemType, FillingType, HyperParameterScalingType, HyperParameterTuningJobObjectiveType, IdleResourceSharing, InstanceGroupStatus, LifecycleManagement, MetricSetSource, MIGProfileType, ModelApprovalStatus, ModelCacheSetting, ModelCompressionType, ModelPackageRegistrationType, ModelPackageStatus, NodeUnavailabilityType, ObjectiveStatus, OutputCompressionType, ParameterType, PreemptTeamTasks, ProblemType, ProcessingS3DataDistributionType, ProcessingS3InputMode, ProductionVariantInstanceType, RecordWrapper, RepositoryAccessMode, ResourceSharingStrategy, S3DataDistribution, S3DataType, S3ModelDataType, SchedulerResourceStatus, SoftwareUpdateStatus, SplitType, TargetDevice, TargetPlatformAccelerator, TargetPlatformArch, TargetPlatformOs, TrafficRoutingConfigType, TrainingInputMode, TrainingInstanceType, TrainingRepositoryAccessMode, TransformInstanceType, VolumeAttachmentStatus } from "./enums";
|
|
2
2
|
/**
|
|
3
3
|
* <p>Configuration for allocating accelerator partitions.</p>
|
|
4
4
|
* @public
|
|
@@ -163,6 +163,16 @@ export interface AddClusterNodeSpecification {
|
|
|
163
163
|
* @public
|
|
164
164
|
*/
|
|
165
165
|
IncrementTargetCountBy: number | undefined;
|
|
166
|
+
/**
|
|
167
|
+
* <p>The availability zones in which to add nodes. Use this to target node placement in specific availability zones within a flexible instance group.</p>
|
|
168
|
+
* @public
|
|
169
|
+
*/
|
|
170
|
+
AvailabilityZones?: string[] | undefined;
|
|
171
|
+
/**
|
|
172
|
+
* <p>The instance types to use when adding nodes. Use this to target specific instance types within a flexible instance group.</p>
|
|
173
|
+
* @public
|
|
174
|
+
*/
|
|
175
|
+
InstanceTypes?: ClusterInstanceType[] | undefined;
|
|
166
176
|
}
|
|
167
177
|
/**
|
|
168
178
|
* <p>Information about additional Elastic Network Interfaces (ENIs) associated with an instance.</p>
|
|
@@ -2965,6 +2975,16 @@ export interface BatchAddClusterNodesError {
|
|
|
2965
2975
|
* @public
|
|
2966
2976
|
*/
|
|
2967
2977
|
FailedCount: number | undefined;
|
|
2978
|
+
/**
|
|
2979
|
+
* <p>The availability zones associated with the failed node addition request.</p>
|
|
2980
|
+
* @public
|
|
2981
|
+
*/
|
|
2982
|
+
AvailabilityZones?: string[] | undefined;
|
|
2983
|
+
/**
|
|
2984
|
+
* <p>The instance types associated with the failed node addition request.</p>
|
|
2985
|
+
* @public
|
|
2986
|
+
*/
|
|
2987
|
+
InstanceTypes?: ClusterInstanceType[] | undefined;
|
|
2968
2988
|
/**
|
|
2969
2989
|
* <p>A descriptive message providing additional details about the error.</p>
|
|
2970
2990
|
* @public
|
|
@@ -2991,6 +3011,16 @@ export interface NodeAdditionResult {
|
|
|
2991
3011
|
* @public
|
|
2992
3012
|
*/
|
|
2993
3013
|
Status: ClusterInstanceStatus | undefined;
|
|
3014
|
+
/**
|
|
3015
|
+
* <p>The availability zones associated with the successfully added node.</p>
|
|
3016
|
+
* @public
|
|
3017
|
+
*/
|
|
3018
|
+
AvailabilityZones?: string[] | undefined;
|
|
3019
|
+
/**
|
|
3020
|
+
* <p>The instance types associated with the successfully added node.</p>
|
|
3021
|
+
* @public
|
|
3022
|
+
*/
|
|
3023
|
+
InstanceTypes?: ClusterInstanceType[] | undefined;
|
|
2994
3024
|
}
|
|
2995
3025
|
/**
|
|
2996
3026
|
* @public
|
|
@@ -4863,6 +4893,22 @@ export interface DeploymentConfiguration {
|
|
|
4863
4893
|
*/
|
|
4864
4894
|
AutoRollbackConfiguration?: AlarmDetails[] | undefined;
|
|
4865
4895
|
}
|
|
4896
|
+
/**
|
|
4897
|
+
* <p>The instance requirement details for a flexible instance group, including the current and desired instance types.</p>
|
|
4898
|
+
* @public
|
|
4899
|
+
*/
|
|
4900
|
+
export interface ClusterInstanceRequirementDetails {
|
|
4901
|
+
/**
|
|
4902
|
+
* <p>The instance types currently in use by the instance group.</p>
|
|
4903
|
+
* @public
|
|
4904
|
+
*/
|
|
4905
|
+
CurrentInstanceTypes?: ClusterInstanceType[] | undefined;
|
|
4906
|
+
/**
|
|
4907
|
+
* <p>The desired instance types for the instance group, as specified in the most recent update request.</p>
|
|
4908
|
+
* @public
|
|
4909
|
+
*/
|
|
4910
|
+
DesiredInstanceTypes?: ClusterInstanceType[] | undefined;
|
|
4911
|
+
}
|
|
4866
4912
|
/**
|
|
4867
4913
|
* <p>Defines the configuration for attaching additional storage to the instances in the SageMaker HyperPod cluster instance group. To learn more, see <a href="https://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-hyperpod-release-notes.html#sagemaker-hyperpod-release-notes-20240620">SageMaker HyperPod release notes: June 20, 2024</a>.</p>
|
|
4868
4914
|
* @public
|
|
@@ -4922,6 +4968,27 @@ export declare namespace ClusterInstanceStorageConfig {
|
|
|
4922
4968
|
_: (name: string, value: any) => T;
|
|
4923
4969
|
}
|
|
4924
4970
|
}
|
|
4971
|
+
/**
|
|
4972
|
+
* <p>Details about a specific instance type within a flexible instance group, including the count and configuration.</p>
|
|
4973
|
+
* @public
|
|
4974
|
+
*/
|
|
4975
|
+
export interface ClusterInstanceTypeDetail {
|
|
4976
|
+
/**
|
|
4977
|
+
* <p>The instance type.</p>
|
|
4978
|
+
* @public
|
|
4979
|
+
*/
|
|
4980
|
+
InstanceType?: ClusterInstanceType | undefined;
|
|
4981
|
+
/**
|
|
4982
|
+
* <p>The number of instances of this type currently running in the instance group.</p>
|
|
4983
|
+
* @public
|
|
4984
|
+
*/
|
|
4985
|
+
CurrentCount?: number | undefined;
|
|
4986
|
+
/**
|
|
4987
|
+
* <p>The number of threads per CPU core for this instance type.</p>
|
|
4988
|
+
* @public
|
|
4989
|
+
*/
|
|
4990
|
+
ThreadsPerCore?: number | undefined;
|
|
4991
|
+
}
|
|
4925
4992
|
/**
|
|
4926
4993
|
* <p>A Kubernetes taint that can be applied to cluster nodes.</p>
|
|
4927
4994
|
* @public
|
|
@@ -4978,12 +5045,12 @@ export interface ClusterLifeCycleConfig {
|
|
|
4978
5045
|
* <p>An Amazon S3 bucket path where your lifecycle scripts are stored.</p> <important> <p>Make sure that the S3 bucket path starts with <code>s3://sagemaker-</code>. The <a href="https://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-hyperpod-prerequisites.html#sagemaker-hyperpod-prerequisites-iam-role-for-hyperpod">IAM role for SageMaker HyperPod</a> has the managed <a href="https://docs.aws.amazon.com/sagemaker/latest/dg/security-iam-awsmanpol-cluster.html"> <code>AmazonSageMakerClusterInstanceRolePolicy</code> </a> attached, which allows access to S3 buckets with the specific prefix <code>sagemaker-</code>.</p> </important>
|
|
4979
5046
|
* @public
|
|
4980
5047
|
*/
|
|
4981
|
-
SourceS3Uri
|
|
5048
|
+
SourceS3Uri?: string | undefined;
|
|
4982
5049
|
/**
|
|
4983
5050
|
* <p>The file name of the entrypoint script of lifecycle scripts under <code>SourceS3Uri</code>. This entrypoint script runs during cluster creation.</p>
|
|
4984
5051
|
* @public
|
|
4985
5052
|
*/
|
|
4986
|
-
OnCreate
|
|
5053
|
+
OnCreate?: string | undefined;
|
|
4987
5054
|
}
|
|
4988
5055
|
/**
|
|
4989
5056
|
* <p>The configuration object of the schedule that SageMaker follows when updating the AMI.</p>
|
|
@@ -5047,6 +5114,16 @@ export interface ClusterInstanceGroupDetails {
|
|
|
5047
5114
|
* @public
|
|
5048
5115
|
*/
|
|
5049
5116
|
InstanceType?: ClusterInstanceType | undefined;
|
|
5117
|
+
/**
|
|
5118
|
+
* <p>The instance requirements for the instance group, including the current and desired instance types. This field is present for flexible instance groups that support multiple instance types.</p>
|
|
5119
|
+
* @public
|
|
5120
|
+
*/
|
|
5121
|
+
InstanceRequirements?: ClusterInstanceRequirementDetails | undefined;
|
|
5122
|
+
/**
|
|
5123
|
+
* <p>Details about the instance types in the instance group, including the count and configuration of each instance type. This field is present for flexible instance groups that support multiple instance types.</p>
|
|
5124
|
+
* @public
|
|
5125
|
+
*/
|
|
5126
|
+
InstanceTypeDetails?: ClusterInstanceTypeDetail[] | undefined;
|
|
5050
5127
|
/**
|
|
5051
5128
|
* <p>Details of LifeCycle configuration for the instance group.</p>
|
|
5052
5129
|
* @public
|
|
@@ -5143,6 +5220,17 @@ export interface ClusterInstanceGroupDetails {
|
|
|
5143
5220
|
*/
|
|
5144
5221
|
SlurmConfig?: ClusterSlurmConfigDetails | undefined;
|
|
5145
5222
|
}
|
|
5223
|
+
/**
|
|
5224
|
+
* <p>The instance requirements for a flexible instance group. Use this to specify multiple instance types that the instance group can use. The order of instance types in the list determines the priority for instance provisioning.</p>
|
|
5225
|
+
* @public
|
|
5226
|
+
*/
|
|
5227
|
+
export interface ClusterInstanceRequirements {
|
|
5228
|
+
/**
|
|
5229
|
+
* <p>The list of instance types that the instance group can use. The order of instance types determines the priority—HyperPod attempts to provision instances using the first instance type in the list and falls back to subsequent types if capacity is unavailable.</p>
|
|
5230
|
+
* @public
|
|
5231
|
+
*/
|
|
5232
|
+
InstanceTypes: ClusterInstanceType[] | undefined;
|
|
5233
|
+
}
|
|
5146
5234
|
/**
|
|
5147
5235
|
* <p>Kubernetes configuration that specifies labels and taints to be applied to cluster nodes in an instance group. </p>
|
|
5148
5236
|
* @public
|
|
@@ -5200,11 +5288,16 @@ export interface ClusterInstanceGroupSpecification {
|
|
|
5200
5288
|
* @public
|
|
5201
5289
|
*/
|
|
5202
5290
|
InstanceType?: ClusterInstanceType | undefined;
|
|
5291
|
+
/**
|
|
5292
|
+
* <p>The instance requirements for the instance group, including the instance types to use. Use this to create a flexible instance group that supports multiple instance types. The <code>InstanceType</code> and <code>InstanceRequirements</code> properties are mutually exclusive.</p>
|
|
5293
|
+
* @public
|
|
5294
|
+
*/
|
|
5295
|
+
InstanceRequirements?: ClusterInstanceRequirements | undefined;
|
|
5203
5296
|
/**
|
|
5204
5297
|
* <p>Specifies the LifeCycle configuration for the instance group.</p>
|
|
5205
5298
|
* @public
|
|
5206
5299
|
*/
|
|
5207
|
-
LifeCycleConfig
|
|
5300
|
+
LifeCycleConfig?: ClusterLifeCycleConfig | undefined;
|
|
5208
5301
|
/**
|
|
5209
5302
|
* <p>Specifies an IAM execution role to be assumed by the instance group.</p>
|
|
5210
5303
|
* @public
|
|
@@ -7154,60 +7247,3 @@ export interface CreateCodeRepositoryInput {
|
|
|
7154
7247
|
*/
|
|
7155
7248
|
Tags?: Tag[] | undefined;
|
|
7156
7249
|
}
|
|
7157
|
-
/**
|
|
7158
|
-
* @public
|
|
7159
|
-
*/
|
|
7160
|
-
export interface CreateCodeRepositoryOutput {
|
|
7161
|
-
/**
|
|
7162
|
-
* <p>The Amazon Resource Name (ARN) of the new repository.</p>
|
|
7163
|
-
* @public
|
|
7164
|
-
*/
|
|
7165
|
-
CodeRepositoryArn: string | undefined;
|
|
7166
|
-
}
|
|
7167
|
-
/**
|
|
7168
|
-
* <p>Contains information about the location of input model artifacts, the name and shape of the expected data inputs, and the framework in which the model was trained.</p>
|
|
7169
|
-
* @public
|
|
7170
|
-
*/
|
|
7171
|
-
export interface InputConfig {
|
|
7172
|
-
/**
|
|
7173
|
-
* <p>The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).</p>
|
|
7174
|
-
* @public
|
|
7175
|
-
*/
|
|
7176
|
-
S3Uri: string | undefined;
|
|
7177
|
-
/**
|
|
7178
|
-
* <p>Specifies the name and shape of the expected data inputs for your trained model with a JSON dictionary form. The data inputs are <code>Framework</code> specific. </p> <ul> <li> <p> <code>TensorFlow</code>: You must specify the name and shape (NHWC format) of the expected data inputs using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.</p> <ul> <li> <p>Examples for one input:</p> <ul> <li> <p>If using the console, <code>\{"input":[1,1024,1024,3]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"input\":[1,1024,1024,3]\}</code> </p> </li> </ul> </li> <li> <p>Examples for two inputs:</p> <ul> <li> <p>If using the console, <code>\{"data1": [1,28,28,1], "data2":[1,28,28,1]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"data1\": [1,28,28,1], \"data2\":[1,28,28,1]\}</code> </p> </li> </ul> </li> </ul> </li> <li> <p> <code>KERAS</code>: You must specify the name and shape (NCHW format) of expected data inputs using a dictionary format for your trained model. Note that while Keras model artifacts should be uploaded in NHWC (channel-last) format, <code>DataInputConfig</code> should be specified in NCHW (channel-first) format. The dictionary formats required for the console and CLI are different.</p> <ul> <li> <p>Examples for one input:</p> <ul> <li> <p>If using the console, <code>\{"input_1":[1,3,224,224]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"input_1\":[1,3,224,224]\}</code> </p> </li> </ul> </li> <li> <p>Examples for two inputs:</p> <ul> <li> <p>If using the console, <code>\{"input_1": [1,3,224,224], "input_2":[1,3,224,224]\} </code> </p> </li> <li> <p>If using the CLI, <code>\{\"input_1\": [1,3,224,224], \"input_2\":[1,3,224,224]\}</code> </p> </li> </ul> </li> </ul> </li> <li> <p> <code>MXNET/ONNX/DARKNET</code>: You must specify the name and shape (NCHW format) of the expected data inputs in order using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.</p> <ul> <li> <p>Examples for one input:</p> <ul> <li> <p>If using the console, <code>\{"data":[1,3,1024,1024]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"data\":[1,3,1024,1024]\}</code> </p> </li> </ul> </li> <li> <p>Examples for two inputs:</p> <ul> <li> <p>If using the console, <code>\{"var1": [1,1,28,28], "var2":[1,1,28,28]\} </code> </p> </li> <li> <p>If using the CLI, <code>\{\"var1\": [1,1,28,28], \"var2\":[1,1,28,28]\}</code> </p> </li> </ul> </li> </ul> </li> <li> <p> <code>PyTorch</code>: You can either specify the name and shape (NCHW format) of expected data inputs in order using a dictionary format for your trained model or you can specify the shape only using a list format. The dictionary formats required for the console and CLI are different. The list formats for the console and CLI are the same.</p> <ul> <li> <p>Examples for one input in dictionary format:</p> <ul> <li> <p>If using the console, <code>\{"input0":[1,3,224,224]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"input0\":[1,3,224,224]\}</code> </p> </li> </ul> </li> <li> <p>Example for one input in list format: <code>[[1,3,224,224]]</code> </p> </li> <li> <p>Examples for two inputs in dictionary format:</p> <ul> <li> <p>If using the console, <code>\{"input0":[1,3,224,224], "input1":[1,3,224,224]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"input0\":[1,3,224,224], \"input1\":[1,3,224,224]\} </code> </p> </li> </ul> </li> <li> <p>Example for two inputs in list format: <code>[[1,3,224,224], [1,3,224,224]]</code> </p> </li> </ul> </li> <li> <p> <code>XGBOOST</code>: input data name and shape are not needed.</p> </li> </ul> <p> <code>DataInputConfig</code> supports the following parameters for <code>CoreML</code> <code>TargetDevice</code> (ML Model format):</p> <ul> <li> <p> <code>shape</code>: Input shape, for example <code>\{"input_1": \{"shape": [1,224,224,3]\}\}</code>. In addition to static input shapes, CoreML converter supports Flexible input shapes:</p> <ul> <li> <p>Range Dimension. You can use the Range Dimension feature if you know the input shape will be within some specific interval in that dimension, for example: <code>\{"input_1": \{"shape": ["1..10", 224, 224, 3]\}\}</code> </p> </li> <li> <p>Enumerated shapes. Sometimes, the models are trained to work only on a select set of inputs. You can enumerate all supported input shapes, for example: <code>\{"input_1": \{"shape": [[1, 224, 224, 3], [1, 160, 160, 3]]\}\}</code> </p> </li> </ul> </li> <li> <p> <code>default_shape</code>: Default input shape. You can set a default shape during conversion for both Range Dimension and Enumerated Shapes. For example <code>\{"input_1": \{"shape": ["1..10", 224, 224, 3], "default_shape": [1, 224, 224, 3]\}\}</code> </p> </li> <li> <p> <code>type</code>: Input type. Allowed values: <code>Image</code> and <code>Tensor</code>. By default, the converter generates an ML Model with inputs of type Tensor (MultiArray). User can set input type to be Image. Image input type requires additional input parameters such as <code>bias</code> and <code>scale</code>.</p> </li> <li> <p> <code>bias</code>: If the input type is an Image, you need to provide the bias vector.</p> </li> <li> <p> <code>scale</code>: If the input type is an Image, you need to provide a scale factor.</p> </li> </ul> <p>CoreML <code>ClassifierConfig</code> parameters can be specified using <a href="https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_OutputConfig.html">OutputConfig</a> <code>CompilerOptions</code>. CoreML converter supports Tensorflow and PyTorch models. CoreML conversion examples:</p> <ul> <li> <p>Tensor type input:</p> <ul> <li> <p> <code>"DataInputConfig": \{"input_1": \{"shape": [[1,224,224,3], [1,160,160,3]], "default_shape": [1,224,224,3]\}\}</code> </p> </li> </ul> </li> <li> <p>Tensor type input without input name (PyTorch):</p> <ul> <li> <p> <code>"DataInputConfig": [\{"shape": [[1,3,224,224], [1,3,160,160]], "default_shape": [1,3,224,224]\}]</code> </p> </li> </ul> </li> <li> <p>Image type input:</p> <ul> <li> <p> <code>"DataInputConfig": \{"input_1": \{"shape": [[1,224,224,3], [1,160,160,3]], "default_shape": [1,224,224,3], "type": "Image", "bias": [-1,-1,-1], "scale": 0.007843137255\}\}</code> </p> </li> <li> <p> <code>"CompilerOptions": \{"class_labels": "imagenet_labels_1000.txt"\}</code> </p> </li> </ul> </li> <li> <p>Image type input without input name (PyTorch):</p> <ul> <li> <p> <code>"DataInputConfig": [\{"shape": [[1,3,224,224], [1,3,160,160]], "default_shape": [1,3,224,224], "type": "Image", "bias": [-1,-1,-1], "scale": 0.007843137255\}]</code> </p> </li> <li> <p> <code>"CompilerOptions": \{"class_labels": "imagenet_labels_1000.txt"\}</code> </p> </li> </ul> </li> </ul> <p>Depending on the model format, <code>DataInputConfig</code> requires the following parameters for <code>ml_eia2</code> <a href="https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_OutputConfig.html#sagemaker-Type-OutputConfig-TargetDevice">OutputConfig:TargetDevice</a>.</p> <ul> <li> <p>For TensorFlow models saved in the SavedModel format, specify the input names from <code>signature_def_key</code> and the input model shapes for <code>DataInputConfig</code>. Specify the <code>signature_def_key</code> in <a href="https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_OutputConfig.html#sagemaker-Type-OutputConfig-CompilerOptions"> <code>OutputConfig:CompilerOptions</code> </a> if the model does not use TensorFlow's default signature def key. For example:</p> <ul> <li> <p> <code>"DataInputConfig": \{"inputs": [1, 224, 224, 3]\}</code> </p> </li> <li> <p> <code>"CompilerOptions": \{"signature_def_key": "serving_custom"\}</code> </p> </li> </ul> </li> <li> <p>For TensorFlow models saved as a frozen graph, specify the input tensor names and shapes in <code>DataInputConfig</code> and the output tensor names for <code>output_names</code> in <a href="https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_OutputConfig.html#sagemaker-Type-OutputConfig-CompilerOptions"> <code>OutputConfig:CompilerOptions</code> </a>. For example:</p> <ul> <li> <p> <code>"DataInputConfig": \{"input_tensor:0": [1, 224, 224, 3]\}</code> </p> </li> <li> <p> <code>"CompilerOptions": \{"output_names": ["output_tensor:0"]\}</code> </p> </li> </ul> </li> </ul>
|
|
7179
|
-
* @public
|
|
7180
|
-
*/
|
|
7181
|
-
DataInputConfig?: string | undefined;
|
|
7182
|
-
/**
|
|
7183
|
-
* <p>Identifies the framework in which the model was trained. For example: TENSORFLOW.</p>
|
|
7184
|
-
* @public
|
|
7185
|
-
*/
|
|
7186
|
-
Framework: Framework | undefined;
|
|
7187
|
-
/**
|
|
7188
|
-
* <p>Specifies the framework version to use. This API field is only supported for the MXNet, PyTorch, TensorFlow and TensorFlow Lite frameworks.</p> <p>For information about framework versions supported for cloud targets and edge devices, see <a href="https://docs.aws.amazon.com/sagemaker/latest/dg/neo-supported-cloud.html">Cloud Supported Instance Types and Frameworks</a> and <a href="https://docs.aws.amazon.com/sagemaker/latest/dg/neo-supported-devices-edge-frameworks.html">Edge Supported Frameworks</a>.</p>
|
|
7189
|
-
* @public
|
|
7190
|
-
*/
|
|
7191
|
-
FrameworkVersion?: string | undefined;
|
|
7192
|
-
}
|
|
7193
|
-
/**
|
|
7194
|
-
* <p>Contains information about a target platform that you want your model to run on, such as OS, architecture, and accelerators. It is an alternative of <code>TargetDevice</code>.</p>
|
|
7195
|
-
* @public
|
|
7196
|
-
*/
|
|
7197
|
-
export interface TargetPlatform {
|
|
7198
|
-
/**
|
|
7199
|
-
* <p>Specifies a target platform OS.</p> <ul> <li> <p> <code>LINUX</code>: Linux-based operating systems.</p> </li> <li> <p> <code>ANDROID</code>: Android operating systems. Android API level can be specified using the <code>ANDROID_PLATFORM</code> compiler option. For example, <code>"CompilerOptions": \{'ANDROID_PLATFORM': 28\}</code> </p> </li> </ul>
|
|
7200
|
-
* @public
|
|
7201
|
-
*/
|
|
7202
|
-
Os: TargetPlatformOs | undefined;
|
|
7203
|
-
/**
|
|
7204
|
-
* <p>Specifies a target platform architecture.</p> <ul> <li> <p> <code>X86_64</code>: 64-bit version of the x86 instruction set.</p> </li> <li> <p> <code>X86</code>: 32-bit version of the x86 instruction set.</p> </li> <li> <p> <code>ARM64</code>: ARMv8 64-bit CPU.</p> </li> <li> <p> <code>ARM_EABIHF</code>: ARMv7 32-bit, Hard Float.</p> </li> <li> <p> <code>ARM_EABI</code>: ARMv7 32-bit, Soft Float. Used by Android 32-bit ARM platform.</p> </li> </ul>
|
|
7205
|
-
* @public
|
|
7206
|
-
*/
|
|
7207
|
-
Arch: TargetPlatformArch | undefined;
|
|
7208
|
-
/**
|
|
7209
|
-
* <p>Specifies a target platform accelerator (optional).</p> <ul> <li> <p> <code>NVIDIA</code>: Nvidia graphics processing unit. It also requires <code>gpu-code</code>, <code>trt-ver</code>, <code>cuda-ver</code> compiler options</p> </li> <li> <p> <code>MALI</code>: ARM Mali graphics processor</p> </li> <li> <p> <code>INTEL_GRAPHICS</code>: Integrated Intel graphics</p> </li> </ul>
|
|
7210
|
-
* @public
|
|
7211
|
-
*/
|
|
7212
|
-
Accelerator?: TargetPlatformAccelerator | undefined;
|
|
7213
|
-
}
|
|
@@ -1,6 +1,63 @@
|
|
|
1
1
|
import { AutomaticJsonStringConversion as __AutomaticJsonStringConversion } from "@smithy/smithy-client";
|
|
2
|
-
import type { _InstanceType, AccountDefaultStatus, ActivationState, AppInstanceType, AppNetworkAccessType, AppSecurityGroupManagement, AppType, AuthMode, AutoMountHomeEFS, AvailabilityZoneBalanceEnforcementMode, AwsManagedHumanLoopRequestSource, CapacityReservationPreference, CollectionType, ContentClassifier, DataDistributionType, DeviceSubsetType, DirectInternetAccess, EdgePresetDeploymentType, ExecutionRoleIdentityConfig, FailureHandlingPolicy, FeatureStatus, FeatureType, FlatInvocations, HubContentType, HyperParameterScalingType, HyperParameterTuningAllocationStrategy, HyperParameterTuningJobStrategyType, HyperParameterTuningJobWarmStartType, InferenceComponentPlacementStrategy, InferenceExecutionMode, InferenceExperimentType, InputMode, IPAddressType, JobType, ManagedInstanceScalingScaleInStrategy, ManagedInstanceScalingStatus, MetricPublishFrequencyInSeconds, MlTools, ModelApprovalStatus, ModelCardStatus, ModelInfrastructureType, ModelPackageRegistrationType, ModelRegistrationMode, ModelSpeculativeDecodingS3DataType, ModelSpeculativeDecodingTechnique, MonitoringProblemType, MonitoringType, NotebookInstanceAcceleratorType, NotebookOutputOption, OptimizationJobDeploymentInstanceType, PartnerAppAuthType, PartnerAppType, ProcessingInstanceType, ProcessingS3CompressionType, ProcessingS3DataDistributionType, ProcessingS3DataType, ProcessingS3InputMode, ProcessingS3UploadMode, Processor, ProductionVariantAcceleratorType, ProductionVariantInferenceAmiVersion, ProductionVariantInstanceType, RecommendationJobSupportedEndpointType, RecommendationJobType, RedshiftResultCompressionType, RedshiftResultFormat, RootAccess, RoutingStrategy, RStudioServerProAccessStatus, RStudioServerProUserGroup, SageMakerImageName, SkipModelValidation, StorageType, StudioWebPortal, TableFormat, TagPropagation, TargetDevice, ThroughputMode, TrackingServerSize, TrafficType, TrainingInputMode, TrainingInstanceType, TrainingJobEarlyStoppingType, TtlDurationUnit, VendorGuidance } from "./enums";
|
|
3
|
-
import type { AdditionalInferenceSpecificationDefinition, AmazonQSettings, AnnotationConsolidationConfig, AppLifecycleManagement, AppSpecification, AsyncInferenceConfig, AthenaDatasetDefinition, AuthorizedUrl, AutoParameter, AutoRollbackConfig, Autotune, BatchTransformInput, BestObjectiveNotImproving, Bias, BlueGreenUpdatePolicy, CanvasAppSettings, CapacitySize, CaptureContentTypeHeader, CaptureOption, CategoricalParameter, CategoricalParameterRange, CfnCreateTemplateProvider, Channel, CheckpointConfig, ClarifyExplainerConfig, CodeEditorAppSettings, CodeRepository, CollectionConfig, ComputeQuotaConfig, ComputeQuotaTarget, ContainerDefinition, ContextSource, ContinuousParameterRange, ConvergenceDetected, CustomImage, HyperParameterTuningJobObjective, InferenceSpecification,
|
|
2
|
+
import type { _InstanceType, AccountDefaultStatus, ActivationState, AppInstanceType, AppNetworkAccessType, AppSecurityGroupManagement, AppType, AuthMode, AutoMountHomeEFS, AvailabilityZoneBalanceEnforcementMode, AwsManagedHumanLoopRequestSource, CapacityReservationPreference, CollectionType, ContentClassifier, DataDistributionType, DeviceSubsetType, DirectInternetAccess, EdgePresetDeploymentType, ExecutionRoleIdentityConfig, FailureHandlingPolicy, FeatureStatus, FeatureType, FlatInvocations, Framework, HubContentType, HyperParameterScalingType, HyperParameterTuningAllocationStrategy, HyperParameterTuningJobStrategyType, HyperParameterTuningJobWarmStartType, InferenceComponentPlacementStrategy, InferenceExecutionMode, InferenceExperimentType, InputMode, IPAddressType, JobType, ManagedInstanceScalingScaleInStrategy, ManagedInstanceScalingStatus, MetricPublishFrequencyInSeconds, MlTools, ModelApprovalStatus, ModelCardStatus, ModelInfrastructureType, ModelPackageRegistrationType, ModelRegistrationMode, ModelSpeculativeDecodingS3DataType, ModelSpeculativeDecodingTechnique, MonitoringProblemType, MonitoringType, NotebookInstanceAcceleratorType, NotebookOutputOption, OptimizationJobDeploymentInstanceType, PartnerAppAuthType, PartnerAppType, ProcessingInstanceType, ProcessingS3CompressionType, ProcessingS3DataDistributionType, ProcessingS3DataType, ProcessingS3InputMode, ProcessingS3UploadMode, Processor, ProductionVariantAcceleratorType, ProductionVariantInferenceAmiVersion, ProductionVariantInstanceType, RecommendationJobSupportedEndpointType, RecommendationJobType, RedshiftResultCompressionType, RedshiftResultFormat, RootAccess, RoutingStrategy, RStudioServerProAccessStatus, RStudioServerProUserGroup, SageMakerImageName, SkipModelValidation, StorageType, StudioWebPortal, TableFormat, TagPropagation, TargetDevice, TargetPlatformAccelerator, TargetPlatformArch, TargetPlatformOs, ThroughputMode, TrackingServerSize, TrafficType, TrainingInputMode, TrainingInstanceType, TrainingJobEarlyStoppingType, TtlDurationUnit, VendorGuidance } from "./enums";
|
|
3
|
+
import type { AdditionalInferenceSpecificationDefinition, AmazonQSettings, AnnotationConsolidationConfig, AppLifecycleManagement, AppSpecification, AsyncInferenceConfig, AthenaDatasetDefinition, AuthorizedUrl, AutoParameter, AutoRollbackConfig, Autotune, BatchTransformInput, BestObjectiveNotImproving, Bias, BlueGreenUpdatePolicy, CanvasAppSettings, CapacitySize, CaptureContentTypeHeader, CaptureOption, CategoricalParameter, CategoricalParameterRange, CfnCreateTemplateProvider, Channel, CheckpointConfig, ClarifyExplainerConfig, CodeEditorAppSettings, CodeRepository, CollectionConfig, ComputeQuotaConfig, ComputeQuotaTarget, ContainerDefinition, ContextSource, ContinuousParameterRange, ConvergenceDetected, CustomImage, HyperParameterTuningJobObjective, InferenceSpecification, MetadataProperties, MetricDefinition, MetricsSource, ModelDataSource, OutputDataConfig, ResourceConfig, ResourceSpec, StoppingCondition, Tag, TransformJobDefinition, VpcConfig } from "./models_0";
|
|
4
|
+
/**
|
|
5
|
+
* @public
|
|
6
|
+
*/
|
|
7
|
+
export interface CreateCodeRepositoryOutput {
|
|
8
|
+
/**
|
|
9
|
+
* <p>The Amazon Resource Name (ARN) of the new repository.</p>
|
|
10
|
+
* @public
|
|
11
|
+
*/
|
|
12
|
+
CodeRepositoryArn: string | undefined;
|
|
13
|
+
}
|
|
14
|
+
/**
|
|
15
|
+
* <p>Contains information about the location of input model artifacts, the name and shape of the expected data inputs, and the framework in which the model was trained.</p>
|
|
16
|
+
* @public
|
|
17
|
+
*/
|
|
18
|
+
export interface InputConfig {
|
|
19
|
+
/**
|
|
20
|
+
* <p>The S3 path where the model artifacts, which result from model training, are stored. This path must point to a single gzip compressed tar archive (.tar.gz suffix).</p>
|
|
21
|
+
* @public
|
|
22
|
+
*/
|
|
23
|
+
S3Uri: string | undefined;
|
|
24
|
+
/**
|
|
25
|
+
* <p>Specifies the name and shape of the expected data inputs for your trained model with a JSON dictionary form. The data inputs are <code>Framework</code> specific. </p> <ul> <li> <p> <code>TensorFlow</code>: You must specify the name and shape (NHWC format) of the expected data inputs using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.</p> <ul> <li> <p>Examples for one input:</p> <ul> <li> <p>If using the console, <code>\{"input":[1,1024,1024,3]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"input\":[1,1024,1024,3]\}</code> </p> </li> </ul> </li> <li> <p>Examples for two inputs:</p> <ul> <li> <p>If using the console, <code>\{"data1": [1,28,28,1], "data2":[1,28,28,1]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"data1\": [1,28,28,1], \"data2\":[1,28,28,1]\}</code> </p> </li> </ul> </li> </ul> </li> <li> <p> <code>KERAS</code>: You must specify the name and shape (NCHW format) of expected data inputs using a dictionary format for your trained model. Note that while Keras model artifacts should be uploaded in NHWC (channel-last) format, <code>DataInputConfig</code> should be specified in NCHW (channel-first) format. The dictionary formats required for the console and CLI are different.</p> <ul> <li> <p>Examples for one input:</p> <ul> <li> <p>If using the console, <code>\{"input_1":[1,3,224,224]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"input_1\":[1,3,224,224]\}</code> </p> </li> </ul> </li> <li> <p>Examples for two inputs:</p> <ul> <li> <p>If using the console, <code>\{"input_1": [1,3,224,224], "input_2":[1,3,224,224]\} </code> </p> </li> <li> <p>If using the CLI, <code>\{\"input_1\": [1,3,224,224], \"input_2\":[1,3,224,224]\}</code> </p> </li> </ul> </li> </ul> </li> <li> <p> <code>MXNET/ONNX/DARKNET</code>: You must specify the name and shape (NCHW format) of the expected data inputs in order using a dictionary format for your trained model. The dictionary formats required for the console and CLI are different.</p> <ul> <li> <p>Examples for one input:</p> <ul> <li> <p>If using the console, <code>\{"data":[1,3,1024,1024]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"data\":[1,3,1024,1024]\}</code> </p> </li> </ul> </li> <li> <p>Examples for two inputs:</p> <ul> <li> <p>If using the console, <code>\{"var1": [1,1,28,28], "var2":[1,1,28,28]\} </code> </p> </li> <li> <p>If using the CLI, <code>\{\"var1\": [1,1,28,28], \"var2\":[1,1,28,28]\}</code> </p> </li> </ul> </li> </ul> </li> <li> <p> <code>PyTorch</code>: You can either specify the name and shape (NCHW format) of expected data inputs in order using a dictionary format for your trained model or you can specify the shape only using a list format. The dictionary formats required for the console and CLI are different. The list formats for the console and CLI are the same.</p> <ul> <li> <p>Examples for one input in dictionary format:</p> <ul> <li> <p>If using the console, <code>\{"input0":[1,3,224,224]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"input0\":[1,3,224,224]\}</code> </p> </li> </ul> </li> <li> <p>Example for one input in list format: <code>[[1,3,224,224]]</code> </p> </li> <li> <p>Examples for two inputs in dictionary format:</p> <ul> <li> <p>If using the console, <code>\{"input0":[1,3,224,224], "input1":[1,3,224,224]\}</code> </p> </li> <li> <p>If using the CLI, <code>\{\"input0\":[1,3,224,224], \"input1\":[1,3,224,224]\} </code> </p> </li> </ul> </li> <li> <p>Example for two inputs in list format: <code>[[1,3,224,224], [1,3,224,224]]</code> </p> </li> </ul> </li> <li> <p> <code>XGBOOST</code>: input data name and shape are not needed.</p> </li> </ul> <p> <code>DataInputConfig</code> supports the following parameters for <code>CoreML</code> <code>TargetDevice</code> (ML Model format):</p> <ul> <li> <p> <code>shape</code>: Input shape, for example <code>\{"input_1": \{"shape": [1,224,224,3]\}\}</code>. In addition to static input shapes, CoreML converter supports Flexible input shapes:</p> <ul> <li> <p>Range Dimension. You can use the Range Dimension feature if you know the input shape will be within some specific interval in that dimension, for example: <code>\{"input_1": \{"shape": ["1..10", 224, 224, 3]\}\}</code> </p> </li> <li> <p>Enumerated shapes. Sometimes, the models are trained to work only on a select set of inputs. You can enumerate all supported input shapes, for example: <code>\{"input_1": \{"shape": [[1, 224, 224, 3], [1, 160, 160, 3]]\}\}</code> </p> </li> </ul> </li> <li> <p> <code>default_shape</code>: Default input shape. You can set a default shape during conversion for both Range Dimension and Enumerated Shapes. For example <code>\{"input_1": \{"shape": ["1..10", 224, 224, 3], "default_shape": [1, 224, 224, 3]\}\}</code> </p> </li> <li> <p> <code>type</code>: Input type. Allowed values: <code>Image</code> and <code>Tensor</code>. By default, the converter generates an ML Model with inputs of type Tensor (MultiArray). User can set input type to be Image. Image input type requires additional input parameters such as <code>bias</code> and <code>scale</code>.</p> </li> <li> <p> <code>bias</code>: If the input type is an Image, you need to provide the bias vector.</p> </li> <li> <p> <code>scale</code>: If the input type is an Image, you need to provide a scale factor.</p> </li> </ul> <p>CoreML <code>ClassifierConfig</code> parameters can be specified using <a href="https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_OutputConfig.html">OutputConfig</a> <code>CompilerOptions</code>. CoreML converter supports Tensorflow and PyTorch models. CoreML conversion examples:</p> <ul> <li> <p>Tensor type input:</p> <ul> <li> <p> <code>"DataInputConfig": \{"input_1": \{"shape": [[1,224,224,3], [1,160,160,3]], "default_shape": [1,224,224,3]\}\}</code> </p> </li> </ul> </li> <li> <p>Tensor type input without input name (PyTorch):</p> <ul> <li> <p> <code>"DataInputConfig": [\{"shape": [[1,3,224,224], [1,3,160,160]], "default_shape": [1,3,224,224]\}]</code> </p> </li> </ul> </li> <li> <p>Image type input:</p> <ul> <li> <p> <code>"DataInputConfig": \{"input_1": \{"shape": [[1,224,224,3], [1,160,160,3]], "default_shape": [1,224,224,3], "type": "Image", "bias": [-1,-1,-1], "scale": 0.007843137255\}\}</code> </p> </li> <li> <p> <code>"CompilerOptions": \{"class_labels": "imagenet_labels_1000.txt"\}</code> </p> </li> </ul> </li> <li> <p>Image type input without input name (PyTorch):</p> <ul> <li> <p> <code>"DataInputConfig": [\{"shape": [[1,3,224,224], [1,3,160,160]], "default_shape": [1,3,224,224], "type": "Image", "bias": [-1,-1,-1], "scale": 0.007843137255\}]</code> </p> </li> <li> <p> <code>"CompilerOptions": \{"class_labels": "imagenet_labels_1000.txt"\}</code> </p> </li> </ul> </li> </ul> <p>Depending on the model format, <code>DataInputConfig</code> requires the following parameters for <code>ml_eia2</code> <a href="https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_OutputConfig.html#sagemaker-Type-OutputConfig-TargetDevice">OutputConfig:TargetDevice</a>.</p> <ul> <li> <p>For TensorFlow models saved in the SavedModel format, specify the input names from <code>signature_def_key</code> and the input model shapes for <code>DataInputConfig</code>. Specify the <code>signature_def_key</code> in <a href="https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_OutputConfig.html#sagemaker-Type-OutputConfig-CompilerOptions"> <code>OutputConfig:CompilerOptions</code> </a> if the model does not use TensorFlow's default signature def key. For example:</p> <ul> <li> <p> <code>"DataInputConfig": \{"inputs": [1, 224, 224, 3]\}</code> </p> </li> <li> <p> <code>"CompilerOptions": \{"signature_def_key": "serving_custom"\}</code> </p> </li> </ul> </li> <li> <p>For TensorFlow models saved as a frozen graph, specify the input tensor names and shapes in <code>DataInputConfig</code> and the output tensor names for <code>output_names</code> in <a href="https://docs.aws.amazon.com/sagemaker/latest/APIReference/API_OutputConfig.html#sagemaker-Type-OutputConfig-CompilerOptions"> <code>OutputConfig:CompilerOptions</code> </a>. For example:</p> <ul> <li> <p> <code>"DataInputConfig": \{"input_tensor:0": [1, 224, 224, 3]\}</code> </p> </li> <li> <p> <code>"CompilerOptions": \{"output_names": ["output_tensor:0"]\}</code> </p> </li> </ul> </li> </ul>
|
|
26
|
+
* @public
|
|
27
|
+
*/
|
|
28
|
+
DataInputConfig?: string | undefined;
|
|
29
|
+
/**
|
|
30
|
+
* <p>Identifies the framework in which the model was trained. For example: TENSORFLOW.</p>
|
|
31
|
+
* @public
|
|
32
|
+
*/
|
|
33
|
+
Framework: Framework | undefined;
|
|
34
|
+
/**
|
|
35
|
+
* <p>Specifies the framework version to use. This API field is only supported for the MXNet, PyTorch, TensorFlow and TensorFlow Lite frameworks.</p> <p>For information about framework versions supported for cloud targets and edge devices, see <a href="https://docs.aws.amazon.com/sagemaker/latest/dg/neo-supported-cloud.html">Cloud Supported Instance Types and Frameworks</a> and <a href="https://docs.aws.amazon.com/sagemaker/latest/dg/neo-supported-devices-edge-frameworks.html">Edge Supported Frameworks</a>.</p>
|
|
36
|
+
* @public
|
|
37
|
+
*/
|
|
38
|
+
FrameworkVersion?: string | undefined;
|
|
39
|
+
}
|
|
40
|
+
/**
|
|
41
|
+
* <p>Contains information about a target platform that you want your model to run on, such as OS, architecture, and accelerators. It is an alternative of <code>TargetDevice</code>.</p>
|
|
42
|
+
* @public
|
|
43
|
+
*/
|
|
44
|
+
export interface TargetPlatform {
|
|
45
|
+
/**
|
|
46
|
+
* <p>Specifies a target platform OS.</p> <ul> <li> <p> <code>LINUX</code>: Linux-based operating systems.</p> </li> <li> <p> <code>ANDROID</code>: Android operating systems. Android API level can be specified using the <code>ANDROID_PLATFORM</code> compiler option. For example, <code>"CompilerOptions": \{'ANDROID_PLATFORM': 28\}</code> </p> </li> </ul>
|
|
47
|
+
* @public
|
|
48
|
+
*/
|
|
49
|
+
Os: TargetPlatformOs | undefined;
|
|
50
|
+
/**
|
|
51
|
+
* <p>Specifies a target platform architecture.</p> <ul> <li> <p> <code>X86_64</code>: 64-bit version of the x86 instruction set.</p> </li> <li> <p> <code>X86</code>: 32-bit version of the x86 instruction set.</p> </li> <li> <p> <code>ARM64</code>: ARMv8 64-bit CPU.</p> </li> <li> <p> <code>ARM_EABIHF</code>: ARMv7 32-bit, Hard Float.</p> </li> <li> <p> <code>ARM_EABI</code>: ARMv7 32-bit, Soft Float. Used by Android 32-bit ARM platform.</p> </li> </ul>
|
|
52
|
+
* @public
|
|
53
|
+
*/
|
|
54
|
+
Arch: TargetPlatformArch | undefined;
|
|
55
|
+
/**
|
|
56
|
+
* <p>Specifies a target platform accelerator (optional).</p> <ul> <li> <p> <code>NVIDIA</code>: Nvidia graphics processing unit. It also requires <code>gpu-code</code>, <code>trt-ver</code>, <code>cuda-ver</code> compiler options</p> </li> <li> <p> <code>MALI</code>: ARM Mali graphics processor</p> </li> <li> <p> <code>INTEL_GRAPHICS</code>: Integrated Intel graphics</p> </li> </ul>
|
|
57
|
+
* @public
|
|
58
|
+
*/
|
|
59
|
+
Accelerator?: TargetPlatformAccelerator | undefined;
|
|
60
|
+
}
|
|
4
61
|
/**
|
|
5
62
|
* <p>Contains information about the output location for the compiled model and the target device that the model runs on. <code>TargetDevice</code> and <code>TargetPlatform</code> are mutually exclusive, so you need to choose one between the two to specify your target device or platform. If you cannot find your device you want to use from the <code>TargetDevice</code> list, use <code>TargetPlatform</code> to describe the platform of your edge device and <code>CompilerOptions</code> if there are specific settings that are required or recommended to use for particular TargetPlatform.</p>
|
|
6
63
|
* @public
|
|
@@ -6934,36 +6991,3 @@ export interface SpaceCodeEditorAppSettings {
|
|
|
6934
6991
|
*/
|
|
6935
6992
|
AppLifecycleManagement?: SpaceAppLifecycleManagement | undefined;
|
|
6936
6993
|
}
|
|
6937
|
-
/**
|
|
6938
|
-
* <p>A file system, created by you in Amazon EFS, that you assign to a user profile or space for an Amazon SageMaker AI Domain. Permitted users can access this file system in Amazon SageMaker AI Studio.</p>
|
|
6939
|
-
* @public
|
|
6940
|
-
*/
|
|
6941
|
-
export interface EFSFileSystem {
|
|
6942
|
-
/**
|
|
6943
|
-
* <p>The ID of your Amazon EFS file system.</p>
|
|
6944
|
-
* @public
|
|
6945
|
-
*/
|
|
6946
|
-
FileSystemId: string | undefined;
|
|
6947
|
-
}
|
|
6948
|
-
/**
|
|
6949
|
-
* <p>A custom file system in Amazon FSx for Lustre.</p>
|
|
6950
|
-
* @public
|
|
6951
|
-
*/
|
|
6952
|
-
export interface FSxLustreFileSystem {
|
|
6953
|
-
/**
|
|
6954
|
-
* <p>Amazon FSx for Lustre file system ID.</p>
|
|
6955
|
-
* @public
|
|
6956
|
-
*/
|
|
6957
|
-
FileSystemId: string | undefined;
|
|
6958
|
-
}
|
|
6959
|
-
/**
|
|
6960
|
-
* <p>A custom file system in Amazon S3. This is only supported in Amazon SageMaker Unified Studio.</p>
|
|
6961
|
-
* @public
|
|
6962
|
-
*/
|
|
6963
|
-
export interface S3FileSystem {
|
|
6964
|
-
/**
|
|
6965
|
-
* <p>The Amazon S3 URI that specifies the location in S3 where files are stored, which is mounted within the Studio environment. For example: <code>s3://<bucket-name>/<prefix>/</code>.</p>
|
|
6966
|
-
* @public
|
|
6967
|
-
*/
|
|
6968
|
-
S3Uri: string | undefined;
|
|
6969
|
-
}
|
|
@@ -1,6 +1,39 @@
|
|
|
1
|
-
import type { _InstanceType, AccountDefaultStatus, ActionStatus, ActivationState, AlgorithmStatus, AppNetworkAccessType, AppSecurityGroupManagement, AppStatus, AppType, AuthMode, AutoMLJobSecondaryStatus, AutoMLJobStatus, AutoMLProblemTypeConfigName, BatchStrategy, CapacityReservationPreference, CaptureStatus, ClusterNodeProvisioningMode, ClusterNodeRecovery, ClusterStatus, CompilationJobStatus, CustomizationTechnique, DetailedModelPackageStatus, DirectInternetAccess, DomainStatus, EdgePackagingJobStatus, EdgePresetDeploymentStatus, EdgePresetDeploymentType, EnabledOrDisabled, EndpointStatus, EvaluationType, ExecutionStatus, FeatureGroupStatus, FeatureStatus, FeatureType, FlowDefinitionStatus, HubContentStatus, HubContentSupportStatus, HubContentType, HubStatus, HumanTaskUiStatus, HyperParameterTuningJobObjectiveType, HyperParameterTuningJobStatus, ImageStatus, ImageVersionStatus, InferenceComponentCapacitySizeType, InferenceComponentStatus, InferenceExperimentStatus, InferenceExperimentType, IPAddressType, IsTrackingServerActive, JobType, JoinSource, LabelingJobStatus, LastUpdateStatusValue, MaintenanceStatus, MlflowAppStatus, ModelApprovalStatus, ModelCardExportJobStatus, ModelCardProcessingStatus, ModelCardStatus, ModelPackageGroupStatus, ModelPackageRegistrationType, ModelPackageStatus, ModelRegistrationMode, ModelVariantStatus, MonitoringType, NotebookInstanceAcceleratorType, NotebookInstanceStatus, ObjectiveStatus, OfflineStoreStatusValue, OptimizationJobDeploymentInstanceType, OptimizationJobStatus, PartnerAppAuthType, PartnerAppStatus, PartnerAppType, Peft,
|
|
2
|
-
import type { ActionSource, AdditionalInferenceSpecificationDefinition, AlgorithmSpecification, AlgorithmStatusDetails, AlgorithmValidationSpecification, ArtifactSource, AsyncInferenceConfig, AutoMLCandidate, AutoMLChannel, AutoMLComputeConfig, AutoMLDataSplitConfig, AutoMLJobArtifacts, AutoMLJobChannel, AutoMLJobCompletionCriteria, AutoMLJobConfig, AutoMLJobObjective, AutoMLOutputDataConfig, AutoMLPartialFailureReason, AutoMLProblemTypeConfig, AutoMLResolvedAttributes, AutoMLSecurityConfig, AutoRollbackConfig, Autotune, AvailableUpgrade, BatchDataCaptureConfig, Channel, CheckpointConfig, ClusterAutoScalingConfigOutput, ClusterEventDetail, ClusterInstanceGroupDetails, ClusterNodeDetails, ClusterOrchestrator, ClusterRestrictedInstanceGroupDetails, ClusterTieredStorageConfig, CodeEditorAppImageConfig, CodeRepository, CognitoConfig, CognitoMemberDefinition, CollectionConfiguration, ComputeQuotaConfig, ComputeQuotaTarget, ContainerDefinition, ContextSource, GitConfig, InferenceSpecification,
|
|
3
|
-
import type { DataCaptureConfig, DataQualityAppSpecification, DataQualityBaselineConfig, DataQualityJobInput, DefaultSpaceSettings, DeploymentConfig, DeviceSelectionConfig, DomainSettings, DriftCheckBaselines, EdgeDeploymentConfig, EdgeDeploymentModelConfig, EdgeOutputConfig,
|
|
1
|
+
import type { _InstanceType, AccountDefaultStatus, ActionStatus, ActivationState, AlgorithmStatus, AppNetworkAccessType, AppSecurityGroupManagement, AppStatus, AppType, AuthMode, AutoMLJobSecondaryStatus, AutoMLJobStatus, AutoMLProblemTypeConfigName, BatchStrategy, CapacityReservationPreference, CaptureStatus, ClusterNodeProvisioningMode, ClusterNodeRecovery, ClusterStatus, CompilationJobStatus, CustomizationTechnique, DeepHealthCheckType, DetailedModelPackageStatus, DirectInternetAccess, DomainStatus, EdgePackagingJobStatus, EdgePresetDeploymentStatus, EdgePresetDeploymentType, EnabledOrDisabled, EndpointStatus, EvaluationType, ExecutionStatus, FeatureGroupStatus, FeatureStatus, FeatureType, FlowDefinitionStatus, HubContentStatus, HubContentSupportStatus, HubContentType, HubStatus, HumanTaskUiStatus, HyperParameterTuningJobObjectiveType, HyperParameterTuningJobStatus, ImageStatus, ImageVersionStatus, InferenceComponentCapacitySizeType, InferenceComponentStatus, InferenceExperimentStatus, InferenceExperimentType, IPAddressType, IsTrackingServerActive, JobType, JoinSource, LabelingJobStatus, LastUpdateStatusValue, MaintenanceStatus, MlflowAppStatus, ModelApprovalStatus, ModelCardExportJobStatus, ModelCardProcessingStatus, ModelCardStatus, ModelPackageGroupStatus, ModelPackageRegistrationType, ModelPackageStatus, ModelRegistrationMode, ModelVariantStatus, MonitoringType, NotebookInstanceAcceleratorType, NotebookInstanceStatus, ObjectiveStatus, OfflineStoreStatusValue, OptimizationJobDeploymentInstanceType, OptimizationJobStatus, PartnerAppAuthType, PartnerAppStatus, PartnerAppType, Peft, ProblemType, ProcessingInstanceType, Processor, ProductionVariantAcceleratorType, ProductionVariantInstanceType, RecommendationJobStatus, RecommendationJobType, RecommendationStatus, RetentionType, RootAccess, RuleEvaluationStatus, SchedulerConfigComponent, SchedulerResourceStatus, ScheduleStatus, ServerlessJobType, SharingType, SkipModelValidation, StageStatus, Statistic, StudioLifecycleConfigAppType, TagPropagation, ThroughputMode, TrackingServerMaintenanceStatus, TrackingServerSize, TrackingServerStatus, TrainingJobStatus, TrialComponentPrimaryStatus, VariantStatus, VendorGuidance, WorkforceIpAddressType } from "./enums";
|
|
2
|
+
import type { ActionSource, AdditionalInferenceSpecificationDefinition, AlgorithmSpecification, AlgorithmStatusDetails, AlgorithmValidationSpecification, ArtifactSource, AsyncInferenceConfig, AutoMLCandidate, AutoMLChannel, AutoMLComputeConfig, AutoMLDataSplitConfig, AutoMLJobArtifacts, AutoMLJobChannel, AutoMLJobCompletionCriteria, AutoMLJobConfig, AutoMLJobObjective, AutoMLOutputDataConfig, AutoMLPartialFailureReason, AutoMLProblemTypeConfig, AutoMLResolvedAttributes, AutoMLSecurityConfig, AutoRollbackConfig, Autotune, AvailableUpgrade, BatchDataCaptureConfig, Channel, CheckpointConfig, ClusterAutoScalingConfigOutput, ClusterEventDetail, ClusterInstanceGroupDetails, ClusterNodeDetails, ClusterOrchestrator, ClusterRestrictedInstanceGroupDetails, ClusterTieredStorageConfig, CodeEditorAppImageConfig, CodeRepository, CognitoConfig, CognitoMemberDefinition, CollectionConfiguration, ComputeQuotaConfig, ComputeQuotaTarget, ContainerDefinition, ContextSource, GitConfig, InferenceSpecification, JupyterLabAppImageConfig, KernelGatewayImageConfig, MetadataProperties, ModelDeployConfig, OutputDataConfig, ResourceConfig, ResourceSpec, SchedulerConfig, StoppingCondition, Tag, TrainingSpecification, TransformInput, TransformOutput, TransformResources, UserContext, VpcConfig } from "./models_0";
|
|
3
|
+
import type { DataCaptureConfig, DataQualityAppSpecification, DataQualityBaselineConfig, DataQualityJobInput, DefaultSpaceSettings, DeploymentConfig, DeviceSelectionConfig, DomainSettings, DriftCheckBaselines, EdgeDeploymentConfig, EdgeDeploymentModelConfig, EdgeOutputConfig, EndpointInfo, ExperimentConfig, ExplainerConfig, FeatureDefinition, FlowDefinitionOutputConfig, HubS3StorageConfig, HumanLoopActivationConfig, HumanLoopConfig, HumanLoopRequestSource, HumanTaskConfig, HyperParameterTrainingJobDefinition, HyperParameterTuningJobConfig, HyperParameterTuningJobWarmStartConfig, InferenceComponentComputeResourceRequirements, InferenceComponentSchedulingConfig, InferenceComponentStartupParameters, InferenceExecutionConfig, InferenceExperimentDataStorageConfig, InferenceExperimentSchedule, InputConfig, InstanceMetadataServiceConfiguration, JupyterServerAppSettings, KernelGatewayAppSettings, LabelingJobAlgorithmsConfig, LabelingJobInputConfig, LabelingJobOutputConfig, LabelingJobStoppingConditions, MetricsConfig, ModelBiasAppSpecification, ModelBiasBaselineConfig, ModelBiasJobInput, ModelCardExportOutputConfig, ModelCardSecurityConfig, ModelExplainabilityAppSpecification, ModelExplainabilityBaselineConfig, ModelExplainabilityJobInput, ModelInfrastructureConfig, ModelLifeCycle, ModelMetrics, ModelPackageModelCard, ModelPackageSecurityConfig, ModelPackageValidationSpecification, ModelQualityAppSpecification, ModelQualityBaselineConfig, ModelQualityJobInput, MonitoringNetworkConfig, MonitoringOutputConfig, MonitoringResources, MonitoringScheduleConfig, MonitoringStoppingCondition, NeoVpcConfig, NotebookInstanceLifecycleHook, OfflineStoreConfig, OnlineStoreConfig, OptimizationConfig, OptimizationJobModelSource, OptimizationJobOutputConfig, OptimizationVpcConfig, OutputConfig, OwnershipSettings, PartnerAppConfig, PartnerAppMaintenanceConfig, ProductionVariant, ProductionVariantManagedInstanceScaling, ProductionVariantRoutingConfig, ProductionVariantServerlessConfig, RecommendationJobInputConfig, RecommendationJobStoppingConditions, RetryStrategy, ShadowModeConfig, SourceAlgorithmSpecification, SpaceAppLifecycleManagement, SpaceCodeEditorAppSettings, UserSettings } from "./models_1";
|
|
4
|
+
/**
|
|
5
|
+
* <p>A file system, created by you in Amazon EFS, that you assign to a user profile or space for an Amazon SageMaker AI Domain. Permitted users can access this file system in Amazon SageMaker AI Studio.</p>
|
|
6
|
+
* @public
|
|
7
|
+
*/
|
|
8
|
+
export interface EFSFileSystem {
|
|
9
|
+
/**
|
|
10
|
+
* <p>The ID of your Amazon EFS file system.</p>
|
|
11
|
+
* @public
|
|
12
|
+
*/
|
|
13
|
+
FileSystemId: string | undefined;
|
|
14
|
+
}
|
|
15
|
+
/**
|
|
16
|
+
* <p>A custom file system in Amazon FSx for Lustre.</p>
|
|
17
|
+
* @public
|
|
18
|
+
*/
|
|
19
|
+
export interface FSxLustreFileSystem {
|
|
20
|
+
/**
|
|
21
|
+
* <p>Amazon FSx for Lustre file system ID.</p>
|
|
22
|
+
* @public
|
|
23
|
+
*/
|
|
24
|
+
FileSystemId: string | undefined;
|
|
25
|
+
}
|
|
26
|
+
/**
|
|
27
|
+
* <p>A custom file system in Amazon S3. This is only supported in Amazon SageMaker Unified Studio.</p>
|
|
28
|
+
* @public
|
|
29
|
+
*/
|
|
30
|
+
export interface S3FileSystem {
|
|
31
|
+
/**
|
|
32
|
+
* <p>The Amazon S3 URI that specifies the location in S3 where files are stored, which is mounted within the Studio environment. For example: <code>s3://<bucket-name>/<prefix>/</code>.</p>
|
|
33
|
+
* @public
|
|
34
|
+
*/
|
|
35
|
+
S3Uri: string | undefined;
|
|
36
|
+
}
|
|
4
37
|
/**
|
|
5
38
|
* <p>A file system, created by you, that you assign to a user profile or space for an Amazon SageMaker AI Domain. Permitted users can access this file system in Amazon SageMaker AI Studio.</p>
|
|
6
39
|
* @public
|
|
@@ -1397,6 +1430,27 @@ export interface DebugRuleEvaluationStatus {
|
|
|
1397
1430
|
*/
|
|
1398
1431
|
LastModifiedTime?: Date | undefined;
|
|
1399
1432
|
}
|
|
1433
|
+
/**
|
|
1434
|
+
* <p>The configuration of deep health checks for an instance group.</p> <note> <p>Overlapping deep health check configurations will be merged into a single operation.</p> </note>
|
|
1435
|
+
* @public
|
|
1436
|
+
*/
|
|
1437
|
+
export interface InstanceGroupHealthCheckConfiguration {
|
|
1438
|
+
/**
|
|
1439
|
+
* <p>The name of the instance group.</p>
|
|
1440
|
+
* @public
|
|
1441
|
+
*/
|
|
1442
|
+
InstanceGroupName: string | undefined;
|
|
1443
|
+
/**
|
|
1444
|
+
* <p>A list of Amazon Elastic Compute Cloud (EC2) instance IDs on which to perform deep health checks.</p> <note> <p>Leave this field blank to perform deep health checks on the entire instance group.</p> </note>
|
|
1445
|
+
* @public
|
|
1446
|
+
*/
|
|
1447
|
+
InstanceIds?: string[] | undefined;
|
|
1448
|
+
/**
|
|
1449
|
+
* <p>A list of deep health checks to be performed.</p>
|
|
1450
|
+
* @public
|
|
1451
|
+
*/
|
|
1452
|
+
DeepHealthChecks: DeepHealthCheckType[] | undefined;
|
|
1453
|
+
}
|
|
1400
1454
|
/**
|
|
1401
1455
|
* @public
|
|
1402
1456
|
*/
|
|
@@ -7965,118 +8019,3 @@ export interface DescribePipelineRequest {
|
|
|
7965
8019
|
*/
|
|
7966
8020
|
PipelineVersionId?: number | undefined;
|
|
7967
8021
|
}
|
|
7968
|
-
/**
|
|
7969
|
-
* @public
|
|
7970
|
-
*/
|
|
7971
|
-
export interface DescribePipelineResponse {
|
|
7972
|
-
/**
|
|
7973
|
-
* <p>The Amazon Resource Name (ARN) of the pipeline.</p>
|
|
7974
|
-
* @public
|
|
7975
|
-
*/
|
|
7976
|
-
PipelineArn?: string | undefined;
|
|
7977
|
-
/**
|
|
7978
|
-
* <p>The name of the pipeline.</p>
|
|
7979
|
-
* @public
|
|
7980
|
-
*/
|
|
7981
|
-
PipelineName?: string | undefined;
|
|
7982
|
-
/**
|
|
7983
|
-
* <p>The display name of the pipeline.</p>
|
|
7984
|
-
* @public
|
|
7985
|
-
*/
|
|
7986
|
-
PipelineDisplayName?: string | undefined;
|
|
7987
|
-
/**
|
|
7988
|
-
* <p>The JSON pipeline definition.</p>
|
|
7989
|
-
* @public
|
|
7990
|
-
*/
|
|
7991
|
-
PipelineDefinition?: string | undefined;
|
|
7992
|
-
/**
|
|
7993
|
-
* <p>The description of the pipeline.</p>
|
|
7994
|
-
* @public
|
|
7995
|
-
*/
|
|
7996
|
-
PipelineDescription?: string | undefined;
|
|
7997
|
-
/**
|
|
7998
|
-
* <p>The Amazon Resource Name (ARN) that the pipeline uses to execute.</p>
|
|
7999
|
-
* @public
|
|
8000
|
-
*/
|
|
8001
|
-
RoleArn?: string | undefined;
|
|
8002
|
-
/**
|
|
8003
|
-
* <p>The status of the pipeline execution.</p>
|
|
8004
|
-
* @public
|
|
8005
|
-
*/
|
|
8006
|
-
PipelineStatus?: PipelineStatus | undefined;
|
|
8007
|
-
/**
|
|
8008
|
-
* <p>The time when the pipeline was created.</p>
|
|
8009
|
-
* @public
|
|
8010
|
-
*/
|
|
8011
|
-
CreationTime?: Date | undefined;
|
|
8012
|
-
/**
|
|
8013
|
-
* <p>The time when the pipeline was last modified.</p>
|
|
8014
|
-
* @public
|
|
8015
|
-
*/
|
|
8016
|
-
LastModifiedTime?: Date | undefined;
|
|
8017
|
-
/**
|
|
8018
|
-
* <p>The time when the pipeline was last run.</p>
|
|
8019
|
-
* @public
|
|
8020
|
-
*/
|
|
8021
|
-
LastRunTime?: Date | undefined;
|
|
8022
|
-
/**
|
|
8023
|
-
* <p>Information about the user who created or modified a SageMaker resource.</p>
|
|
8024
|
-
* @public
|
|
8025
|
-
*/
|
|
8026
|
-
CreatedBy?: UserContext | undefined;
|
|
8027
|
-
/**
|
|
8028
|
-
* <p>Information about the user who created or modified a SageMaker resource.</p>
|
|
8029
|
-
* @public
|
|
8030
|
-
*/
|
|
8031
|
-
LastModifiedBy?: UserContext | undefined;
|
|
8032
|
-
/**
|
|
8033
|
-
* <p>Lists the parallelism configuration applied to the pipeline.</p>
|
|
8034
|
-
* @public
|
|
8035
|
-
*/
|
|
8036
|
-
ParallelismConfiguration?: ParallelismConfiguration | undefined;
|
|
8037
|
-
/**
|
|
8038
|
-
* <p>The display name of the pipeline version.</p>
|
|
8039
|
-
* @public
|
|
8040
|
-
*/
|
|
8041
|
-
PipelineVersionDisplayName?: string | undefined;
|
|
8042
|
-
/**
|
|
8043
|
-
* <p>The description of the pipeline version.</p>
|
|
8044
|
-
* @public
|
|
8045
|
-
*/
|
|
8046
|
-
PipelineVersionDescription?: string | undefined;
|
|
8047
|
-
}
|
|
8048
|
-
/**
|
|
8049
|
-
* @public
|
|
8050
|
-
*/
|
|
8051
|
-
export interface DescribePipelineDefinitionForExecutionRequest {
|
|
8052
|
-
/**
|
|
8053
|
-
* <p>The Amazon Resource Name (ARN) of the pipeline execution.</p>
|
|
8054
|
-
* @public
|
|
8055
|
-
*/
|
|
8056
|
-
PipelineExecutionArn: string | undefined;
|
|
8057
|
-
}
|
|
8058
|
-
/**
|
|
8059
|
-
* @public
|
|
8060
|
-
*/
|
|
8061
|
-
export interface DescribePipelineDefinitionForExecutionResponse {
|
|
8062
|
-
/**
|
|
8063
|
-
* <p>The JSON pipeline definition.</p>
|
|
8064
|
-
* @public
|
|
8065
|
-
*/
|
|
8066
|
-
PipelineDefinition?: string | undefined;
|
|
8067
|
-
/**
|
|
8068
|
-
* <p>The time when the pipeline was created.</p>
|
|
8069
|
-
* @public
|
|
8070
|
-
*/
|
|
8071
|
-
CreationTime?: Date | undefined;
|
|
8072
|
-
}
|
|
8073
|
-
/**
|
|
8074
|
-
* @public
|
|
8075
|
-
*/
|
|
8076
|
-
export interface DescribePipelineExecutionRequest {
|
|
8077
|
-
/**
|
|
8078
|
-
* <p>The Amazon Resource Name (ARN) of the pipeline execution.</p>
|
|
8079
|
-
* @public
|
|
8080
|
-
*/
|
|
8081
|
-
PipelineExecutionArn: string | undefined;
|
|
8082
|
-
}
|