Skip to content

Commit 4ec1484

Browse files
author
AWS
committed
Amazon SageMaker Service Update: Added inference components model data caching feature
1 parent 5d241d9 commit 4ec1484

File tree

2 files changed

+41
-2
lines changed

2 files changed

+41
-2
lines changed
Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,6 @@
1+
{
2+
"type": "feature",
3+
"category": "Amazon SageMaker Service",
4+
"contributor": "",
5+
"description": "Added inference components model data caching feature"
6+
}

services/sagemaker/src/main/resources/codegen-resources/service-2.json

Lines changed: 35 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -12446,7 +12446,7 @@
1244612446
},
1244712447
"PlatformIdentifier":{
1244812448
"shape":"PlatformIdentifier",
12449-
"documentation":"<p>The platform identifier of the notebook instance runtime environment.</p>"
12449+
"documentation":"<p>The platform identifier of the notebook instance runtime environment. The default value is <code>notebook-al2-v2</code>.</p>"
1245012450
},
1245112451
"InstanceMetadataServiceConfiguration":{
1245212452
"shape":"InstanceMetadataServiceConfiguration",
@@ -21172,6 +21172,7 @@
2117221172
},
2117321173
"documentation":"<p>The configuration parameters that specify the IAM roles assumed by the execution role of SageMaker (assumable roles) and the cluster instances or job execution environments (execution roles or runtime roles) to manage and access resources required for running Amazon EMR clusters or Amazon EMR Serverless applications.</p>"
2117421174
},
21175+
"EnableCaching":{"type":"boolean"},
2117521176
"EnableCapture":{"type":"boolean"},
2117621177
"EnableInfraCheck":{
2117721178
"type":"boolean",
@@ -24706,6 +24707,30 @@
2470624707
"box":true,
2470724708
"min":0
2470824709
},
24710+
"InferenceComponentDataCacheConfig":{
24711+
"type":"structure",
24712+
"required":["EnableCaching"],
24713+
"members":{
24714+
"EnableCaching":{
24715+
"shape":"EnableCaching",
24716+
"documentation":"<p>Sets whether the endpoint that hosts the inference component caches the model artifacts and container image.</p> <p>With caching enabled, the endpoint caches this data in each instance that it provisions for the inference component. That way, the inference component deploys faster during the auto scaling process. If caching isn't enabled, the inference component takes longer to deploy because of the time it spends downloading the data.</p>",
24717+
"box":true
24718+
}
24719+
},
24720+
"documentation":"<p>Settings that affect how the inference component caches data.</p>"
24721+
},
24722+
"InferenceComponentDataCacheConfigSummary":{
24723+
"type":"structure",
24724+
"required":["EnableCaching"],
24725+
"members":{
24726+
"EnableCaching":{
24727+
"shape":"EnableCaching",
24728+
"documentation":"<p>Indicates whether the inference component caches model artifacts as part of the auto scaling process.</p>",
24729+
"box":true
24730+
}
24731+
},
24732+
"documentation":"<p>Settings that affect how the inference component caches data.</p>"
24733+
},
2470924734
"InferenceComponentDeploymentConfig":{
2471024735
"type":"structure",
2471124736
"required":["RollingUpdatePolicy"],
@@ -24811,6 +24836,10 @@
2481124836
"BaseInferenceComponentName":{
2481224837
"shape":"InferenceComponentName",
2481324838
"documentation":"<p>The name of an existing inference component that is to contain the inference component that you're creating with your request.</p> <p>Specify this parameter only if your request is meant to create an adapter inference component. An adapter inference component contains the path to an adapter model. The purpose of the adapter model is to tailor the inference output of a base foundation model, which is hosted by the base inference component. The adapter inference component uses the compute resources that you assigned to the base inference component.</p> <p>When you create an adapter inference component, use the <code>Container</code> parameter to specify the location of the adapter artifacts. In the parameter value, use the <code>ArtifactUrl</code> parameter of the <code>InferenceComponentContainerSpecification</code> data type.</p> <p>Before you can create an adapter inference component, you must have an existing inference component that contains the foundation model that you want to adapt.</p>"
24839+
},
24840+
"DataCacheConfig":{
24841+
"shape":"InferenceComponentDataCacheConfig",
24842+
"documentation":"<p>Settings that affect how the inference component caches data.</p>"
2481424843
}
2481524844
},
2481624845
"documentation":"<p>Details about the resources to deploy with this inference component, including the model, container, and compute resources.</p>"
@@ -24837,6 +24866,10 @@
2483724866
"BaseInferenceComponentName":{
2483824867
"shape":"InferenceComponentName",
2483924868
"documentation":"<p>The name of the base inference component that contains this inference component.</p>"
24869+
},
24870+
"DataCacheConfig":{
24871+
"shape":"InferenceComponentDataCacheConfigSummary",
24872+
"documentation":"<p>Settings that affect how the inference component caches data.</p>"
2484024873
}
2484124874
},
2484224875
"documentation":"<p>Details about the resources that are deployed with this inference component.</p>"
@@ -36136,7 +36169,7 @@
3613636169
},
3613736170
"S3DataDistributionType":{
3613836171
"shape":"ProcessingS3DataDistributionType",
36139-
"documentation":"<p>Whether to distribute the data from Amazon S3 to all processing instances with <code>FullyReplicated</code>, or whether the data from Amazon S3 is shared by Amazon S3 key, downloading one shard of data to each processing instance.</p>"
36172+
"documentation":"<p>Whether to distribute the data from Amazon S3 to all processing instances with <code>FullyReplicated</code>, or whether the data from Amazon S3 is sharded by Amazon S3 key, downloading one shard of data to each processing instance.</p>"
3614036173
},
3614136174
"S3CompressionType":{
3614236175
"shape":"ProcessingS3CompressionType",

0 commit comments

Comments
 (0)