diff options
author | arcadia-devtools <arcadia-devtools@yandex-team.ru> | 2022-06-13 01:20:54 +0300 |
---|---|---|
committer | arcadia-devtools <arcadia-devtools@yandex-team.ru> | 2022-06-13 01:20:54 +0300 |
commit | 669e8d5ef59e6f49667af117dfee9a43efbf2441 (patch) | |
tree | 9e4a6bcdb72a727e62d1c29c7c4d143138cc20ed /contrib/python | |
parent | 386da7e65b6beb4b05206eed4955c05526c94b7b (diff) | |
download | ydb-669e8d5ef59e6f49667af117dfee9a43efbf2441.tar.gz |
intermediate changes
ref:5004388c52f14b7b0a9521b0eee797b5081d4f35
Diffstat (limited to 'contrib/python')
9 files changed, 125 insertions, 31 deletions
diff --git a/contrib/python/boto3/py3/.dist-info/METADATA b/contrib/python/boto3/py3/.dist-info/METADATA index 767a92e631..dda810d673 100644 --- a/contrib/python/boto3/py3/.dist-info/METADATA +++ b/contrib/python/boto3/py3/.dist-info/METADATA @@ -1,6 +1,6 @@ Metadata-Version: 2.1 Name: boto3 -Version: 1.23.9 +Version: 1.23.10 Summary: The AWS SDK for Python Home-page: https://github.com/boto/boto3 Author: Amazon Web Services @@ -22,7 +22,7 @@ Classifier: Programming Language :: Python :: 3.10 Requires-Python: >= 3.6 License-File: LICENSE License-File: NOTICE -Requires-Dist: botocore (<1.27.0,>=1.26.9) +Requires-Dist: botocore (<1.27.0,>=1.26.10) Requires-Dist: jmespath (<2.0.0,>=0.7.1) Requires-Dist: s3transfer (<0.6.0,>=0.5.0) Provides-Extra: crt diff --git a/contrib/python/boto3/py3/boto3/__init__.py b/contrib/python/boto3/py3/boto3/__init__.py index 9ff6abe9f0..e67d53bf8a 100644 --- a/contrib/python/boto3/py3/boto3/__init__.py +++ b/contrib/python/boto3/py3/boto3/__init__.py @@ -17,7 +17,7 @@ from boto3.compat import _warn_deprecated_python from boto3.session import Session __author__ = 'Amazon Web Services' -__version__ = '1.23.9' +__version__ = '1.23.10' # The default Boto3 session; autoloaded when needed. diff --git a/contrib/python/botocore/py3/.dist-info/METADATA b/contrib/python/botocore/py3/.dist-info/METADATA index 52a679689d..261902cc34 100644 --- a/contrib/python/botocore/py3/.dist-info/METADATA +++ b/contrib/python/botocore/py3/.dist-info/METADATA @@ -1,6 +1,6 @@ Metadata-Version: 2.1 Name: botocore -Version: 1.26.9 +Version: 1.26.10 Summary: Low-level, data-driven core of boto 3. Home-page: https://github.com/boto/botocore Author: Amazon Web Services diff --git a/contrib/python/botocore/py3/botocore/__init__.py b/contrib/python/botocore/py3/botocore/__init__.py index 00493239cd..6595c3bf0e 100644 --- a/contrib/python/botocore/py3/botocore/__init__.py +++ b/contrib/python/botocore/py3/botocore/__init__.py @@ -16,7 +16,7 @@ import logging import os import re -__version__ = '1.26.9' +__version__ = '1.26.10' class NullHandler(logging.Handler): diff --git a/contrib/python/botocore/py3/botocore/data/appflow/2020-08-23/service-2.json b/contrib/python/botocore/py3/botocore/data/appflow/2020-08-23/service-2.json index ab002df10d..9374c938cf 100644 --- a/contrib/python/botocore/py3/botocore/data/appflow/2020-08-23/service-2.json +++ b/contrib/python/botocore/py3/botocore/data/appflow/2020-08-23/service-2.json @@ -27,7 +27,7 @@ {"shape":"ConnectorAuthenticationException"}, {"shape":"InternalServerException"} ], - "documentation":"<p> Creates a new connector profile associated with your Amazon Web Services account. There is a soft quota of 100 connector profiles per Amazon Web Services account. If you need more connector profiles than this quota allows, you can submit a request to the Amazon AppFlow team through the Amazon AppFlow support channel. </p>" + "documentation":"<p> Creates a new connector profile associated with your Amazon Web Services account. There is a soft quota of 100 connector profiles per Amazon Web Services account. If you need more connector profiles than this quota allows, you can submit a request to the Amazon AppFlow team through the Amazon AppFlow support channel. In each connector profile that you create, you can provide the credentials and properties for only one connector.</p>" }, "CreateFlow":{ "name":"CreateFlow", @@ -108,7 +108,7 @@ {"shape":"ConnectorServerException"}, {"shape":"InternalServerException"} ], - "documentation":"<p> Provides details regarding the entity used with the connector, with a description of the data model for each entity. </p>" + "documentation":"<p> Provides details regarding the entity used with the connector, with a description of the data model for each field in that entity. </p>" }, "DescribeConnectorProfiles":{ "name":"DescribeConnectorProfiles", @@ -385,7 +385,7 @@ }, "AccessToken":{ "type":"string", - "max":2048, + "max":4096, "pattern":"\\S+", "sensitive":true }, @@ -2610,6 +2610,11 @@ "max":2048, "pattern":"[\\w!@#\\-.?,\\s]*" }, + "FlowErrorDeactivationThreshold":{ + "type":"integer", + "max":100, + "min":1 + }, "FlowExecutionList":{ "type":"list", "member":{"shape":"ExecutionRecord"} @@ -2862,6 +2867,7 @@ "error":{"httpStatusCode":500}, "exception":true }, + "JavaBoolean":{"type":"boolean"}, "KMSArn":{ "type":"string", "max":2048, @@ -3598,7 +3604,7 @@ }, "RefreshToken":{ "type":"string", - "max":1024, + "max":2048, "pattern":"\\S+" }, "Region":{ @@ -3734,7 +3740,11 @@ "shape":"PrefixConfig", "documentation":"<p> Determines the prefix that Amazon AppFlow applies to the folder name in the Amazon S3 bucket. You can name folders according to the flow frequency and date. </p>" }, - "aggregationConfig":{"shape":"AggregationConfig"} + "aggregationConfig":{"shape":"AggregationConfig"}, + "preserveSourceDataTyping":{ + "shape":"JavaBoolean", + "documentation":"<p>If your file output format is Parquet, use this parameter to set whether Amazon AppFlow preserves the data types in your source data when it writes the output to Amazon S3. </p> <ul> <li> <p> <code>true</code>: Amazon AppFlow preserves the data types when it writes to Amazon S3. For example, an integer or <code>1</code> in your source data is still an integer in your output.</p> </li> <li> <p> <code>false</code>: Amazon AppFlow converts all of the source data into strings when it writes to Amazon S3. For example, an integer of <code>1</code> in your source data becomes the string <code>\"1\"</code> in the output.</p> </li> </ul>" + } }, "documentation":"<p> The configuration that determines how Amazon AppFlow should format the flow output data when Amazon S3 is used as the destination. </p>" }, @@ -4018,15 +4028,15 @@ }, "scheduleStartTime":{ "shape":"Date", - "documentation":"<p> Specifies the scheduled start time for a schedule-triggered flow. </p>" + "documentation":"<p>The time at which the scheduled flow starts. The time is formatted as a timestamp that follows the ISO 8601 standard, such as <code>2022-04-26T13:00:00-07:00</code>.</p>" }, "scheduleEndTime":{ "shape":"Date", - "documentation":"<p> Specifies the scheduled end time for a schedule-triggered flow. </p>" + "documentation":"<p>The time at which the scheduled flow ends. The time is formatted as a timestamp that follows the ISO 8601 standard, such as <code>2022-04-27T13:00:00-07:00</code>.</p>" }, "timezone":{ "shape":"Timezone", - "documentation":"<p> Specifies the time zone used when referring to the date and time of a scheduled-triggered flow, such as <code>America/New_York</code>. </p>" + "documentation":"<p>Specifies the time zone used when referring to the dates and times of a scheduled flow, such as <code>America/New_York</code>. This time zone is only a descriptive label. It doesn't affect how Amazon AppFlow interprets the timestamps that you specify to schedule the flow.</p> <p>If you want to schedule a flow by using times in a particular time zone, indicate the time zone as a UTC offset in your timestamps. For example, the UTC offsets for the <code>America/New_York</code> timezone are <code>-04:00</code> EDT and <code>-05:00 EST</code>.</p>" }, "scheduleOffset":{ "shape":"ScheduleOffset", @@ -4036,6 +4046,11 @@ "firstExecutionFrom":{ "shape":"Date", "documentation":"<p> Specifies the date range for the records to import from the connector in the first flow run. </p>" + }, + "flowErrorDeactivationThreshold":{ + "shape":"FlowErrorDeactivationThreshold", + "documentation":"<p>Defines how many times a scheduled flow fails consecutively before Amazon AppFlow deactivates it.</p>", + "box":true } }, "documentation":"<p> Specifies the configuration details of a schedule-triggered flow as defined by the user. Currently, these settings only apply to the <code>Scheduled</code> trigger type. </p>" diff --git a/contrib/python/botocore/py3/botocore/data/datasync/2018-11-09/service-2.json b/contrib/python/botocore/py3/botocore/data/datasync/2018-11-09/service-2.json index cd3ab99eab..2c6e18abdb 100644 --- a/contrib/python/botocore/py3/botocore/data/datasync/2018-11-09/service-2.json +++ b/contrib/python/botocore/py3/botocore/data/datasync/2018-11-09/service-2.json @@ -54,7 +54,7 @@ {"shape":"InvalidRequestException"}, {"shape":"InternalException"} ], - "documentation":"<p>Creates an endpoint for an Amazon EFS file system.</p>" + "documentation":"<p>Creates an endpoint for an Amazon EFS file system that DataSync can access for a transfer. For more information, see <a href=\"https://docs.aws.amazon.com/datasync/latest/userguide/create-efs-location.html\">Creating a location for Amazon EFS</a>.</p>" }, "CreateLocationFsxLustre":{ "name":"CreateLocationFsxLustre", @@ -250,7 +250,7 @@ {"shape":"InvalidRequestException"}, {"shape":"InternalException"} ], - "documentation":"<p>Returns metadata, such as the path information about an Amazon EFS location.</p>" + "documentation":"<p>Returns metadata about your DataSync location for an Amazon EFS file system.</p>" }, "DescribeLocationFsxLustre":{ "name":"DescribeLocationFsxLustre", @@ -726,19 +726,31 @@ "members":{ "Subdirectory":{ "shape":"EfsSubdirectory", - "documentation":"<p>A subdirectory in the location’s path. This subdirectory in the EFS file system is used to read data from the EFS source location or write data to the EFS destination. By default, DataSync uses the root directory.</p> <note> <p> <code>Subdirectory</code> must be specified with forward slashes. For example, <code>/path/to/folder</code>.</p> </note>" + "documentation":"<p>Specifies a mount path for your Amazon EFS file system. This is where DataSync reads or writes data (depending on if this is a source or destination location). By default, DataSync uses the root directory, but you can also include subdirectories.</p> <note> <p>You must specify a value with forward slashes (for example, <code>/path/to/folder</code>).</p> </note>" }, "EfsFilesystemArn":{ "shape":"EfsFilesystemArn", - "documentation":"<p>The Amazon Resource Name (ARN) for the Amazon EFS file system.</p>" + "documentation":"<p>Specifies the ARN for the Amazon EFS file system.</p>" }, "Ec2Config":{ "shape":"Ec2Config", - "documentation":"<p>The subnet and security group that the Amazon EFS file system uses. The security group that you provide needs to be able to communicate with the security group on the mount target in the subnet specified.</p> <p>The exact relationship between security group M (of the mount target) and security group S (which you provide for DataSync to use at this stage) is as follows: </p> <ul> <li> <p> Security group M (which you associate with the mount target) must allow inbound access for the Transmission Control Protocol (TCP) on the NFS port (2049) from security group S. You can enable inbound connections either by IP address (CIDR range) or security group. </p> </li> <li> <p>Security group S (provided to DataSync to access EFS) should have a rule that enables outbound connections to the NFS port on one of the file system’s mount targets. You can enable outbound connections either by IP address (CIDR range) or security group.</p> <p>For information about security groups and mount targets, see Security Groups for Amazon EC2 Instances and Mount Targets in the <i>Amazon EFS User Guide.</i> </p> </li> </ul>" + "documentation":"<p>Specifies the subnet and security groups DataSync uses to access your Amazon EFS file system.</p>" }, "Tags":{ "shape":"InputTagList", - "documentation":"<p>The key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location.</p>" + "documentation":"<p>Specifies the key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location.</p>" + }, + "AccessPointArn":{ + "shape":"EfsAccessPointArn", + "documentation":"<p>Specifies the Amazon Resource Name (ARN) of the access point that DataSync uses to access the Amazon EFS file system.</p>" + }, + "FileSystemAccessRoleArn":{ + "shape":"IamRoleArn", + "documentation":"<p>Specifies an Identity and Access Management (IAM) role that DataSync assumes when mounting the Amazon EFS file system.</p>" + }, + "InTransitEncryption":{ + "shape":"EfsInTransitEncryption", + "documentation":"<p>Specifies whether you want DataSync to use TLS encryption when transferring data to or from your Amazon EFS file system.</p> <p>If you specify an access point using <code>AccessPointArn</code> or an IAM role using <code>FileSystemAccessRoleArn</code>, you must set this parameter to <code>TLS1_2</code>.</p>" } }, "documentation":"<p>CreateLocationEfsRequest</p>" @@ -748,7 +760,7 @@ "members":{ "LocationArn":{ "shape":"LocationArn", - "documentation":"<p>The Amazon Resource Name (ARN) of the Amazon EFS file system location that is created.</p>" + "documentation":"<p>The Amazon Resource Name (ARN) of the Amazon EFS file system location that you create.</p>" } }, "documentation":"<p>CreateLocationEfs</p>" @@ -1294,7 +1306,7 @@ "members":{ "LocationArn":{ "shape":"LocationArn", - "documentation":"<p>The Amazon Resource Name (ARN) of the EFS location to describe.</p>" + "documentation":"<p>The Amazon Resource Name (ARN) of the Amazon EFS file system location that you want information about.</p>" } }, "documentation":"<p>DescribeLocationEfsRequest</p>" @@ -1304,16 +1316,28 @@ "members":{ "LocationArn":{ "shape":"LocationArn", - "documentation":"<p>The Amazon Resource Name (ARN) of the EFS location that was described.</p>" + "documentation":"<p>The ARN of the Amazon EFS file system location.</p>" }, "LocationUri":{ "shape":"LocationUri", - "documentation":"<p>The URL of the EFS location that was described.</p>" + "documentation":"<p>The URL of the Amazon EFS file system location.</p>" }, "Ec2Config":{"shape":"Ec2Config"}, "CreationTime":{ "shape":"Time", - "documentation":"<p>The time that the EFS location was created.</p>" + "documentation":"<p>The time that the location was created.</p>" + }, + "AccessPointArn":{ + "shape":"EfsAccessPointArn", + "documentation":"<p>The ARN of the access point that DataSync uses to access the Amazon EFS file system.</p>" + }, + "FileSystemAccessRoleArn":{ + "shape":"IamRoleArn", + "documentation":"<p>The Identity and Access Management (IAM) role that DataSync assumes when mounting the Amazon EFS file system.</p>" + }, + "InTransitEncryption":{ + "shape":"EfsInTransitEncryption", + "documentation":"<p>Whether DataSync uses TLS encryption when transferring data to or from your Amazon EFS file system.</p>" } }, "documentation":"<p>DescribeLocationEfsResponse</p>" @@ -1808,14 +1832,14 @@ "members":{ "SubnetArn":{ "shape":"Ec2SubnetArn", - "documentation":"<p>The ARN of the subnet that DataSync uses to access the target EFS file system.</p>" + "documentation":"<p>Specifies the ARN of a subnet where DataSync creates the <a href=\"https://docs.aws.amazon.com/datasync/latest/userguide/datasync-network.html#required-network-interfaces\">network interfaces</a> for managing traffic during your transfer.</p> <p>The subnet must be located:</p> <ul> <li> <p>In the same virtual private cloud (VPC) as the Amazon EFS file system.</p> </li> <li> <p>In the same Availability Zone as at least one mount target for the Amazon EFS file system.</p> </li> </ul> <note> <p>You don't need to specify a subnet that includes a file system mount target.</p> </note>" }, "SecurityGroupArns":{ "shape":"Ec2SecurityGroupArnList", - "documentation":"<p>The Amazon Resource Names (ARNs) of the security groups that are configured for the Amazon EC2 resource.</p>" + "documentation":"<p>Specifies the Amazon Resource Names (ARNs) of the security groups associated with an Amazon EFS file system's mount target.</p>" } }, - "documentation":"<p>The subnet that DataSync uses to access target EFS file system. The subnet must have at least one mount target for that file system. The security group that you provide needs to be able to communicate with the security group on the mount target in the subnet specified. </p>" + "documentation":"<p>The subnet and security groups that DataSync uses to access your Amazon EFS file system.</p>" }, "Ec2SecurityGroupArn":{ "type":"string", @@ -1833,11 +1857,23 @@ "max":128, "pattern":"^arn:(aws|aws-cn|aws-us-gov|aws-iso|aws-iso-b):ec2:[a-z\\-0-9]*:[0-9]{12}:subnet/.*$" }, + "EfsAccessPointArn":{ + "type":"string", + "max":128, + "pattern":"^arn:(aws|aws-cn|aws-us-gov|aws-iso|aws-iso-b):elasticfilesystem:[a-z\\-0-9]+:[0-9]{12}:access-point/fsap-[0-9a-f]{8,40}$" + }, "EfsFilesystemArn":{ "type":"string", "max":128, "pattern":"^arn:(aws|aws-cn|aws-us-gov|aws-iso|aws-iso-b):elasticfilesystem:[a-z\\-0-9]*:[0-9]{12}:file-system/fs-.*$" }, + "EfsInTransitEncryption":{ + "type":"string", + "enum":[ + "NONE", + "TLS1_2" + ] + }, "EfsSubdirectory":{ "type":"string", "max":4096, diff --git a/contrib/python/botocore/py3/botocore/data/emr-serverless/2021-07-13/service-2.json b/contrib/python/botocore/py3/botocore/data/emr-serverless/2021-07-13/service-2.json index d4e777c980..a9b421f87f 100644 --- a/contrib/python/botocore/py3/botocore/data/emr-serverless/2021-07-13/service-2.json +++ b/contrib/python/botocore/py3/botocore/data/emr-serverless/2021-07-13/service-2.json @@ -5,7 +5,7 @@ "endpointPrefix":"emr-serverless", "jsonVersion":"1.1", "protocol":"rest-json", - "serviceFullName":"EMR Serverless Web Service", + "serviceFullName":"EMR Serverless", "serviceId":"EMR Serverless", "signatureVersion":"v4", "signingName":"emr-serverless", diff --git a/contrib/python/botocore/py3/botocore/data/endpoints.json b/contrib/python/botocore/py3/botocore/data/endpoints.json index 142d128f31..e8166ef7c8 100644 --- a/contrib/python/botocore/py3/botocore/data/endpoints.json +++ b/contrib/python/botocore/py3/botocore/data/endpoints.json @@ -1699,6 +1699,7 @@ "ap-south-1" : { }, "ap-southeast-1" : { }, "ap-southeast-2" : { }, + "ap-southeast-3" : { }, "ca-central-1" : { }, "eu-central-1" : { }, "eu-north-1" : { }, @@ -8737,6 +8738,12 @@ }, "hostname" : "oidc.ap-northeast-2.amazonaws.com" }, + "ap-northeast-3" : { + "credentialScope" : { + "region" : "ap-northeast-3" + }, + "hostname" : "oidc.ap-northeast-3.amazonaws.com" + }, "ap-south-1" : { "credentialScope" : { "region" : "ap-south-1" diff --git a/contrib/python/botocore/py3/botocore/data/sagemaker/2017-07-24/service-2.json b/contrib/python/botocore/py3/botocore/data/sagemaker/2017-07-24/service-2.json index 0cdc3da82c..6fef9891f0 100644 --- a/contrib/python/botocore/py3/botocore/data/sagemaker/2017-07-24/service-2.json +++ b/contrib/python/botocore/py3/botocore/data/sagemaker/2017-07-24/service-2.json @@ -3288,11 +3288,11 @@ "members":{ "TrainingImage":{ "shape":"AlgorithmImage", - "documentation":"<p>The registry path of the Docker image that contains the training algorithm. For information about docker registry paths for built-in algorithms, see <a href=\"https://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-algo-docker-registry-paths.html\">Algorithms Provided by Amazon SageMaker: Common Parameters</a>. SageMaker supports both <code>registry/repository[:tag]</code> and <code>registry/repository[@digest]</code> image path formats. For more information, see <a href=\"https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms.html\">Using Your Own Algorithms with Amazon SageMaker</a>.</p>" + "documentation":"<p>The registry path of the Docker image that contains the training algorithm. For information about docker registry paths for SageMaker built-in algorithms, see <a href=\"https://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-algo-docker-registry-paths.html\">Docker Registry Paths and Example Code</a> in the <i>Amazon SageMaker developer guide</i>. SageMaker supports both <code>registry/repository[:tag]</code> and <code>registry/repository[@digest]</code> image path formats. For more information about using your custom training container, see <a href=\"https://docs.aws.amazon.com/sagemaker/latest/dg/your-algorithms.html\">Using Your Own Algorithms with Amazon SageMaker</a>.</p> <note> <p>You must specify either the algorithm name to the <code>AlgorithmName</code> parameter or the image URI of the algorithm container to the <code>TrainingImage</code> parameter.</p> <p>For more information, see the note in the <code>AlgorithmName</code> parameter description.</p> </note>" }, "AlgorithmName":{ "shape":"ArnOrName", - "documentation":"<p>The name of the algorithm resource to use for the training job. This must be an algorithm resource that you created or subscribe to on Amazon Web Services Marketplace. If you specify a value for this parameter, you can't specify a value for <code>TrainingImage</code>.</p>" + "documentation":"<p>The name of the algorithm resource to use for the training job. This must be an algorithm resource that you created or subscribe to on Amazon Web Services Marketplace.</p> <note> <p>You must specify either the algorithm name to the <code>AlgorithmName</code> parameter or the image URI of the algorithm container to the <code>TrainingImage</code> parameter.</p> <p>Note that the <code>AlgorithmName</code> parameter is mutually exclusive with the <code>TrainingImage</code> parameter. If you specify a value for the <code>AlgorithmName</code> parameter, you can't specify a value for <code>TrainingImage</code>, and vice versa.</p> <p>If you specify values for both parameters, the training job might break; if you don't specify any value for both parameters, the training job might raise a <code>null</code> error.</p> </note>" }, "TrainingInputMode":{"shape":"TrainingInputMode"}, "MetricDefinitions":{ @@ -3583,7 +3583,15 @@ "ml.r5.8xlarge", "ml.r5.12xlarge", "ml.r5.16xlarge", - "ml.r5.24xlarge" + "ml.r5.24xlarge", + "ml.g5.xlarge", + "ml.g5.2xlarge", + "ml.g5.4xlarge", + "ml.g5.8xlarge", + "ml.g5.16xlarge", + "ml.g5.12xlarge", + "ml.g5.24xlarge", + "ml.g5.48xlarge" ] }, "AppList":{ @@ -7081,6 +7089,10 @@ "PlatformIdentifier":{ "shape":"PlatformIdentifier", "documentation":"<p>The platform identifier of the notebook instance runtime environment.</p>" + }, + "InstanceMetadataServiceConfiguration":{ + "shape":"InstanceMetadataServiceConfiguration", + "documentation":"<p>Information on the IMDS configuration of the notebook instance</p>" } } }, @@ -11049,6 +11061,10 @@ "PlatformIdentifier":{ "shape":"PlatformIdentifier", "documentation":"<p>The platform identifier of the notebook instance runtime environment.</p>" + }, + "InstanceMetadataServiceConfiguration":{ + "shape":"InstanceMetadataServiceConfiguration", + "documentation":"<p>Information on the IMDS configuration of the notebook instance</p>" } } }, @@ -14869,6 +14885,17 @@ "member":{"shape":"TrainingInputMode"}, "min":1 }, + "InstanceMetadataServiceConfiguration":{ + "type":"structure", + "required":["MinimumInstanceMetadataServiceVersion"], + "members":{ + "MinimumInstanceMetadataServiceVersion":{ + "shape":"MinimumInstanceMetadataServiceVersion", + "documentation":"<p>Indicates the minimum IMDS version that the notebook instance supports. When passed as part of <code>CreateNotebookInstance</code>, if no value is selected, then it defaults to IMDSv1. This means that both IMDSv1 and IMDSv2 are supported. If passed as part of <code>UpdateNotebookInstance</code>, there is no default.</p>" + } + }, + "documentation":"<p>Information on the IMDS configuration of the notebook instance</p>" + }, "InstanceType":{ "type":"string", "enum":[ @@ -18727,6 +18754,11 @@ }, "documentation":"<p/>" }, + "MinimumInstanceMetadataServiceVersion":{ + "type":"string", + "max":1, + "pattern":"1|2" + }, "ModelApprovalStatus":{ "type":"string", "enum":[ @@ -26390,6 +26422,10 @@ "RootAccess":{ "shape":"RootAccess", "documentation":"<p>Whether root access is enabled or disabled for users of the notebook instance. The default value is <code>Enabled</code>.</p> <note> <p>If you set this to <code>Disabled</code>, users don't have root access on the notebook instance, but lifecycle configuration scripts still run with root permissions.</p> </note>" + }, + "InstanceMetadataServiceConfiguration":{ + "shape":"InstanceMetadataServiceConfiguration", + "documentation":"<p>Information on the IMDS configuration of the notebook instance</p>" } } }, |