h&,E      !"#$%&'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\]^_`abcdefghijklmnopqrstuvwxyz{|}~                                                                                                                                    !!!!!!!!!!!!!""""""""""""""""""""""""""""""#########################$$$$$$$$$$$$$$$$$$$$$$$$$%%%%%%%%%%%%%&&&&&&&&&&&&&&&&&&''''''''''''''((((((((((((((())))))))))))))))))))))))))))*****************************++++++++++++++,,,,,,,,,,,,,,,,,,,,,,,,----------------......................../////////////000000000000000000000000000111111111111111122222222222222222222222233333333333333333333333333444444444444444444444444455555555555555555555555555555555555555555666777 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E E F F F F F F F F F F F F F F F F F F F F F F F F F F F F F F F F F F F G G G G G G G G G G G G G G G G GGGGGGGGGGGGGGGGGGGGGGGHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJKKKKKKKKKKKKKKKKKKKKKKKKKKLLLLLLLLLLLLLLLLLLLLLLLLLLMMMMMMMMMMMMMMMMMMMMMMMMMMNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYZZZZZZZZZZZZZZZZZZZZZZZZZZ[[[[[[[[[[[[[[[[[[[[[[[[[[[[\\\\\\\\\\\\\\\\\\\\\\\\\\\\]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^_______________________________________`````````````````````````````````````aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaabbbbbbbbbbbbbbbbbbbbbbbbbbbb(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?6(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';;amazonka-datasync/Represents a single entry in a list of agents. AgentListEntry< returns an array that contains a list of agents when the  https://docs.aws.amazon.com/datasync/latest/userguide/API_ListAgents.html ListAgents operation is called.See:  smart constructor.amazonka-datasync,The Amazon Resource Name (ARN) of the agent.amazonka-datasyncThe name of the agent.amazonka-datasyncThe status of the agent.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, / - The Amazon Resource Name (ARN) of the agent.,  - The name of the agent.,   - The status of the agent.amazonka-datasync,The Amazon Resource Name (ARN) of the agent.amazonka-datasyncThe name of the agent. amazonka-datasyncThe status of the agent.   (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?<(,+)*(,+)*,+(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';G@amazonka-datasyncThe subnet and security groups that DataSync uses to access your Amazon EFS file system.See: D smart constructor.Bamazonka-datasync:Specifies the ARN of a subnet where DataSync creates the  https://docs.aws.amazon.com/datasync/latest/userguide/datasync-network.html#required-network-interfacesnetwork interfaces, for managing traffic during your transfer.The subnet must be located:In the same virtual private cloud (VPC) as the Amazon EFS file system.In the same Availability Zone as at least one mount target for the Amazon EFS file system.You don't need to specify a subnet that includes a file system mount target.Camazonka-datasyncSpecifies the Amazon Resource Names (ARNs) of the security groups associated with an Amazon EFS file system's mount target.Damazonka-datasyncCreate a value of @" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:B, E= - Specifies the ARN of a subnet where DataSync creates the  https://docs.aws.amazon.com/datasync/latest/userguide/datasync-network.html#required-network-interfacesnetwork interfaces, for managing traffic during your transfer.The subnet must be located:In the same virtual private cloud (VPC) as the Amazon EFS file system.In the same Availability Zone as at least one mount target for the Amazon EFS file system.You don't need to specify a subnet that includes a file system mount target.C, F - Specifies the Amazon Resource Names (ARNs) of the security groups associated with an Amazon EFS file system's mount target.Eamazonka-datasync:Specifies the ARN of a subnet where DataSync creates the  https://docs.aws.amazon.com/datasync/latest/userguide/datasync-network.html#required-network-interfacesnetwork interfaces, for managing traffic during your transfer.The subnet must be located:In the same virtual private cloud (VPC) as the Amazon EFS file system.In the same Availability Zone as at least one mount target for the Amazon EFS file system.You don't need to specify a subnet that includes a file system mount target.Famazonka-datasyncSpecifies the Amazon Resource Names (ARNs) of the security groups associated with an Amazon EFS file system's mount target.Damazonka-datasyncBamazonka-datasyncC@CBADEF@CBADEF(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?HWOSRPQOSRPQSR(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?Iglkjhi glkjhilkj(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?I(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';P$amazonka-datasyncSpecifies which files, folders, and objects to include or exclude when transferring files from source to destination.See:  smart constructor.amazonka-datasyncThe type of filter rule to apply. DataSync only supports the SIMPLE_PATTERN rule type.amazonka-datasyncA single filter string that consists of the patterns to include or exclude. The patterns are delimited by "|" (that is, a pipe), for example: /folder1|/folder2amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The type of filter rule to apply. DataSync only supports the SIMPLE_PATTERN rule type.,  - A single filter string that consists of the patterns to include or exclude. The patterns are delimited by "|" (that is, a pipe), for example: /folder1|/folder2amazonka-datasyncThe type of filter rule to apply. DataSync only supports the SIMPLE_PATTERN rule type.amazonka-datasyncA single filter string that consists of the patterns to include or exclude. The patterns are delimited by "|" (that is, a pipe), for example: /folder1|/folder2 (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?P  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?Q (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?Rh  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';ZMamazonka-datasyncThe NameNode of the Hadoop Distributed File System (HDFS). The NameNode manages the file system's namespace. The NameNode performs operations such as opening, closing, and renaming files and directories. The NameNode contains the information to map blocks of data to the DataNodes.See:  smart constructor.amazonka-datasyncThe hostname of the NameNode in the HDFS cluster. This value is the IP address or Domain Name Service (DNS) name of the NameNode. An agent that's installed on-premises uses this hostname to communicate with the NameNode in the network.amazonka-datasync=The port that the NameNode uses to listen to client requests.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The hostname of the NameNode in the HDFS cluster. This value is the IP address or Domain Name Service (DNS) name of the NameNode. An agent that's installed on-premises uses this hostname to communicate with the NameNode in the network.,  - The port that the NameNode uses to listen to client requests.amazonka-datasyncThe hostname of the NameNode in the HDFS cluster. This value is the IP address or Domain Name Service (DNS) name of the NameNode. An agent that's installed on-premises uses this hostname to communicate with the NameNode in the network.amazonka-datasync=The port that the NameNode uses to listen to client requests.amazonka-datasyncamazonka-datasync (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?[  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?[ (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';mamazonka-datasync2Represents a single entry in a list of locations. LocationListEntry? returns an array that contains a list of locations when the  https://docs.aws.amazon.com/datasync/latest/userguide/API_ListLocations.html ListLocations operation is called.See:  smart constructor.amazonka-datasyncThe Amazon Resource Name (ARN) of the location. For Network File System (NFS) or Amazon EFS, the location is the export path. For Amazon S3, the location is the prefix path that you want to mount and use as the root of the location.amazonka-datasync)Represents a list of URIs of a location.  LocationUri? returns an array that contains a list of locations when the  https://docs.aws.amazon.com/datasync/latest/userguide/API_ListLocations.html ListLocations operation is called.Format: TYPE://GLOBAL_ID/SUBDIR.3TYPE designates the type of location (for example, nfs or s3).GLOBAL_ID is the globally unique identifier of the resource that backs the location. An example for EFS is us-east-2.fs-abcd12348. An example for Amazon S3 is the bucket name, such as myBucket. An example for NFS is a valid IPv4 address or a hostname that is compliant with Domain Name Service (DNS).SUBDIR is a valid file system path, delimited by forward slashes as is the *nix convention. For NFS and Amazon EFS, it's the export path to mount the location. For Amazon S3, it's the prefix path that you mount to and treat as the root of the location.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the location. For Network File System (NFS) or Amazon EFS, the location is the export path. For Amazon S3, the location is the prefix path that you want to mount and use as the root of the location., , - Represents a list of URIs of a location.  LocationUri? returns an array that contains a list of locations when the  https://docs.aws.amazon.com/datasync/latest/userguide/API_ListLocations.html ListLocations operation is called.Format: TYPE://GLOBAL_ID/SUBDIR.3TYPE designates the type of location (for example, nfs or s3).GLOBAL_ID is the globally unique identifier of the resource that backs the location. An example for EFS is us-east-2.fs-abcd12348. An example for Amazon S3 is the bucket name, such as myBucket. An example for NFS is a valid IPv4 address or a hostname that is compliant with Domain Name Service (DNS).SUBDIR is a valid file system path, delimited by forward slashes as is the *nix convention. For NFS and Amazon EFS, it's the export path to mount the location. For Amazon S3, it's the prefix path that you mount to and treat as the root of the location.amazonka-datasyncThe Amazon Resource Name (ARN) of the location. For Network File System (NFS) or Amazon EFS, the location is the export path. For Amazon S3, the location is the prefix path that you want to mount and use as the root of the location.amazonka-datasync)Represents a list of URIs of a location.  LocationUri? returns an array that contains a list of locations when the  https://docs.aws.amazon.com/datasync/latest/userguide/API_ListLocations.html ListLocations operation is called.Format: TYPE://GLOBAL_ID/SUBDIR.3TYPE designates the type of location (for example, nfs or s3).GLOBAL_ID is the globally unique identifier of the resource that backs the location. An example for EFS is us-east-2.fs-abcd12348. An example for Amazon S3 is the bucket name, such as myBucket. An example for NFS is a valid IPv4 address or a hostname that is compliant with Domain Name Service (DNS).SUBDIR is a valid file system path, delimited by forward slashes as is the *nix convention. For NFS and Amazon EFS, it's the export path to mount the location. For Amazon S3, it's the prefix path that you mount to and treat as the root of the location.(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?n (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?oA(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?o (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';|amazonka-datasyncSpecifies how DataSync can access a location using the NFS protocol.See:  smart constructor.amazonka-datasyncSpecifies the NFS version that you want DataSync to use when mounting your NFS share. If the server refuses to use the version specified, the task fails.&You can specify the following options: AUTOMATIC- (default): DataSync chooses NFS version 4.1.NFS3: Stateless protocol version that allows for asynchronous writes on the server.NFSv4_0: Stateful, firewall-friendly protocol version that supports delegations and pseudo file systems.NFSv4_1: Stateful protocol version that supports sessions, directory delegations, and parallel data processing. NFS version 4.1 also includes all features available in version 4.0.DataSync currently only supports NFS version 3 with Amazon FSx for NetApp ONTAP locations.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the NFS version that you want DataSync to use when mounting your NFS share. If the server refuses to use the version specified, the task fails.&You can specify the following options: AUTOMATIC- (default): DataSync chooses NFS version 4.1.NFS3: Stateless protocol version that allows for asynchronous writes on the server.NFSv4_0: Stateful, firewall-friendly protocol version that supports delegations and pseudo file systems.NFSv4_1: Stateful protocol version that supports sessions, directory delegations, and parallel data processing. NFS version 4.1 also includes all features available in version 4.0.DataSync currently only supports NFS version 3 with Amazon FSx for NetApp ONTAP locations.amazonka-datasyncSpecifies the NFS version that you want DataSync to use when mounting your NFS share. If the server refuses to use the version specified, the task fails.&You can specify the following options: AUTOMATIC- (default): DataSync chooses NFS version 4.1.NFS3: Stateless protocol version that allows for asynchronous writes on the server.NFSv4_0: Stateful, firewall-friendly protocol version that supports delegations and pseudo file systems.NFSv4_1: Stateful protocol version that supports sessions, directory delegations, and parallel data processing. NFS version 4.1 also includes all features available in version 4.0.DataSync currently only supports NFS version 3 with Amazon FSx for NetApp ONTAP locations.(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';pamazonka-datasyncSpecifies the Network File System (NFS) protocol configuration that DataSync uses to access your Amazon FSx for OpenZFS or Amazon FSx for NetApp ONTAP file system.See:  smart constructor.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Undocumented member.amazonka-datasyncUndocumented member.(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?((c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-datasyncA list of Amazon Resource Names (ARNs) of agents to use for a Network File System (NFS) location.See:  smart constructor.amazonka-datasync.ARNs of the agents to use for an NFS location.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 1 - ARNs of the agents to use for an NFS location.amazonka-datasync.ARNs of the agents to use for an NFS location.amazonka-datasync(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?] (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-datasync.Narrow down the list of resources returned by  ListLocations. For example, to see all your Amazon S3 locations, create a filter using "Name": "LocationType", "Operator": "Equals", and "Values": "S3".For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/query-resources.htmlfiltering resources.See:  smart constructor.amazonka-datasyncThe name of the filter being used. Each API call supports a list of filters that are available for it (for example,  LocationType for  ListLocations).amazonka-datasyncThe values that you want to filter for. For example, you might want to display only Amazon S3 locations.amazonka-datasyncThe operator that is used to compare filter values (for example, Equals or Contains).amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The name of the filter being used. Each API call supports a list of filters that are available for it (for example,  LocationType for  ListLocations).,  - The values that you want to filter for. For example, you might want to display only Amazon S3 locations.,  - The operator that is used to compare filter values (for example, Equals or Contains).amazonka-datasyncThe name of the filter being used. Each API call supports a list of filters that are available for it (for example,  LocationType for  ListLocations).amazonka-datasyncThe values that you want to filter for. For example, you might want to display only Amazon S3 locations.amazonka-datasyncThe operator that is used to compare filter values (for example, Equals or Contains).amazonka-datasyncamazonka-datasync  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?X (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-datasyncThe VPC endpoint, subnet, and security group that an agent uses to access IP addresses in a VPC (Virtual Private Cloud).See:  smart constructor.amazonka-datasyncThe private endpoint that is configured for an agent that has access to IP addresses in a  https://docs.aws.amazon.com/vpc/latest/userguide/endpoint-service.html PrivateLink. An agent that is configured with this endpoint will not be accessible over the public internet.amazonka-datasyncThe Amazon Resource Names (ARNs) of the security groups that are configured for the EC2 resource that hosts an agent activated in a VPC or an agent that has access to a VPC endpoint.amazonka-datasyncThe Amazon Resource Names (ARNs) of the subnets that are configured for an agent activated in a VPC or an agent that has access to a VPC endpoint.amazonka-datasyncThe ID of the VPC endpoint that is configured for an agent. An agent that is configured with a VPC endpoint will not be accessible over the public internet.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The private endpoint that is configured for an agent that has access to IP addresses in a  https://docs.aws.amazon.com/vpc/latest/userguide/endpoint-service.html PrivateLink. An agent that is configured with this endpoint will not be accessible over the public internet.,  - The Amazon Resource Names (ARNs) of the security groups that are configured for the EC2 resource that hosts an agent activated in a VPC or an agent that has access to a VPC endpoint.,  - The Amazon Resource Names (ARNs) of the subnets that are configured for an agent activated in a VPC or an agent that has access to a VPC endpoint.,  - The ID of the VPC endpoint that is configured for an agent. An agent that is configured with a VPC endpoint will not be accessible over the public internet.amazonka-datasyncThe private endpoint that is configured for an agent that has access to IP addresses in a  https://docs.aws.amazon.com/vpc/latest/userguide/endpoint-service.html PrivateLink. An agent that is configured with this endpoint will not be accessible over the public internet.amazonka-datasyncThe Amazon Resource Names (ARNs) of the security groups that are configured for the EC2 resource that hosts an agent activated in a VPC or an agent that has access to a VPC endpoint.amazonka-datasyncThe Amazon Resource Names (ARNs) of the subnets that are configured for an agent activated in a VPC or an agent that has access to a VPC endpoint.amazonka-datasyncThe ID of the VPC endpoint that is configured for an agent. An agent that is configured with a VPC endpoint will not be accessible over the public internet.   (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';%amazonka-datasyncThe Quality of Protection (QOP) configuration specifies the Remote Procedure Call (RPC) and data transfer privacy settings configured on the Hadoop Distributed File System (HDFS) cluster.See:  smart constructor.amazonka-datasyncThe data transfer protection setting configured on the HDFS cluster. This setting corresponds to your dfs.data.transfer.protection setting in the  hdfs-site.xml file on your Hadoop cluster.amazonka-datasyncThe RPC protection setting configured on the HDFS cluster. This setting corresponds to your hadoop.rpc.protection setting in your  core-site.xml file on your Hadoop cluster.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The data transfer protection setting configured on the HDFS cluster. This setting corresponds to your dfs.data.transfer.protection setting in the  hdfs-site.xml file on your Hadoop cluster.,  - The RPC protection setting configured on the HDFS cluster. This setting corresponds to your hadoop.rpc.protection setting in your  core-site.xml file on your Hadoop cluster.amazonka-datasyncThe data transfer protection setting configured on the HDFS cluster. This setting corresponds to your dfs.data.transfer.protection setting in the  hdfs-site.xml file on your Hadoop cluster.amazonka-datasyncThe RPC protection setting configured on the HDFS cluster. This setting corresponds to your hadoop.rpc.protection setting in your  core-site.xml file on your Hadoop cluster.!(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-datasyncThe Amazon Resource Name (ARN) of the Identity and Access Management (IAM) role used to access an Amazon S3 bucket.For detailed information about using such a role, see Creating a Location for Amazon S3 in the DataSync User Guide.See:  smart constructor.amazonka-datasync4The ARN of the IAM role for accessing the S3 bucket.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 7 - The ARN of the IAM role for accessing the S3 bucket.amazonka-datasync4The ARN of the IAM role for accessing the S3 bucket.amazonka-datasync"(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?F #(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?& $(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";? %(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-datasyncSpecifies how DataSync can access a location using the SMB protocol.See:  smart constructor.amazonka-datasyncSpecifies the SMB version that you want DataSync to use when mounting your SMB share. If you don't specify a version, DataSync defaults to  AUTOMATIC and chooses a version based on negotiation with the SMB server.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the SMB version that you want DataSync to use when mounting your SMB share. If you don't specify a version, DataSync defaults to  AUTOMATIC and chooses a version based on negotiation with the SMB server.amazonka-datasyncSpecifies the SMB version that you want DataSync to use when mounting your SMB share. If you don't specify a version, DataSync defaults to  AUTOMATIC and chooses a version based on negotiation with the SMB server.&(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';* amazonka-datasyncSpecifies the Server Message Block (SMB) protocol configuration that DataSync uses to access your Amazon FSx for NetApp ONTAP file system. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-ontap-location.html#create-ontap-location-access$Accessing FSx for ONTAP file systems.See:  smart constructor.amazonka-datasyncSpecifies the fully qualified domain name (FQDN) of the Microsoft Active Directory that your storage virtual machine (SVM) belongs to.amazonka-datasyncSpecifies the password of a user who has permission to access your SVM.amazonka-datasyncSpecifies a user name that can mount the location and access the files, folders, and metadata that you need in the SVM.If you provide a user in your Active Directory, note the following:If you're using Directory Service for Microsoft Active Directory, the user must be a member of the Amazon Web Services Delegated FSx Administrators group.If you're using a self-managed Active Directory, the user must be a member of either the Domain Admins group or a custom group that you specified for file system administration when you created your file system.Make sure that the user has the permissions it needs to copy the data you want: SE_TCB_NAME: Required to set object ownership and file metadata. With this privilege, you also can copy NTFS discretionary access lists (DACLs).SE_SECURITY_NAME: May be needed to copy NTFS system access control lists (SACLs). This operation specifically requires the Windows privilege, which is granted to members of the Domain Admins group. If you configure your task to copy SACLs, make sure that the user has the required privileges. For information about copying SACLs, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-task.html#configure-ownership-and-permissions)Ownership and permissions-related options.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the fully qualified domain name (FQDN) of the Microsoft Active Directory that your storage virtual machine (SVM) belongs to.,  - Undocumented member.,  - Specifies the password of a user who has permission to access your SVM.,  - Specifies a user name that can mount the location and access the files, folders, and metadata that you need in the SVM.If you provide a user in your Active Directory, note the following:If you're using Directory Service for Microsoft Active Directory, the user must be a member of the Amazon Web Services Delegated FSx Administrators group.If you're using a self-managed Active Directory, the user must be a member of either the Domain Admins group or a custom group that you specified for file system administration when you created your file system.Make sure that the user has the permissions it needs to copy the data you want: SE_TCB_NAME: Required to set object ownership and file metadata. With this privilege, you also can copy NTFS discretionary access lists (DACLs).SE_SECURITY_NAME: May be needed to copy NTFS system access control lists (SACLs). This operation specifically requires the Windows privilege, which is granted to members of the Domain Admins group. If you configure your task to copy SACLs, make sure that the user has the required privileges. For information about copying SACLs, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-task.html#configure-ownership-and-permissions)Ownership and permissions-related options.amazonka-datasyncSpecifies the fully qualified domain name (FQDN) of the Microsoft Active Directory that your storage virtual machine (SVM) belongs to.amazonka-datasyncUndocumented member.amazonka-datasyncSpecifies the password of a user who has permission to access your SVM.amazonka-datasyncSpecifies a user name that can mount the location and access the files, folders, and metadata that you need in the SVM.If you provide a user in your Active Directory, note the following:If you're using Directory Service for Microsoft Active Directory, the user must be a member of the Amazon Web Services Delegated FSx Administrators group.If you're using a self-managed Active Directory, the user must be a member of either the Domain Admins group or a custom group that you specified for file system administration when you created your file system.Make sure that the user has the permissions it needs to copy the data you want: SE_TCB_NAME: Required to set object ownership and file metadata. With this privilege, you also can copy NTFS discretionary access lists (DACLs).SE_SECURITY_NAME: May be needed to copy NTFS system access control lists (SACLs). This operation specifically requires the Windows privilege, which is granted to members of the Domain Admins group. If you configure your task to copy SACLs, make sure that the user has the required privileges. For information about copying SACLs, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-task.html#configure-ownership-and-permissions)Ownership and permissions-related options.amazonka-datasyncamazonka-datasync  '(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';`amazonka-datasyncSpecifies the data transfer protocol that DataSync uses to access your Amazon FSx file system.See:  smart constructor.amazonka-datasyncSpecifies the Network File System (NFS) protocol configuration that DataSync uses to access your FSx for OpenZFS file system or FSx for ONTAP file system's storage virtual machine (SVM).amazonka-datasyncSpecifies the Server Message Block (SMB) protocol configuration that DataSync uses to access your FSx for ONTAP file system's SVM.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the Network File System (NFS) protocol configuration that DataSync uses to access your FSx for OpenZFS file system or FSx for ONTAP file system's storage virtual machine (SVM).,  - Specifies the Server Message Block (SMB) protocol configuration that DataSync uses to access your FSx for ONTAP file system's SVM.amazonka-datasyncSpecifies the Network File System (NFS) protocol configuration that DataSync uses to access your FSx for OpenZFS file system or FSx for ONTAP file system's storage virtual machine (SVM).amazonka-datasyncSpecifies the Server Message Block (SMB) protocol configuration that DataSync uses to access your FSx for ONTAP file system's SVM.((c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-datasyncA key-value pair representing a single tag that's been applied to an Amazon Web Services resource.See:  smart constructor.amazonka-datasync2The value for an Amazon Web Services resource tag.amazonka-datasync0The key for an Amazon Web Services resource tag.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 5 - The value for an Amazon Web Services resource tag., 3 - The key for an Amazon Web Services resource tag.amazonka-datasync2The value for an Amazon Web Services resource tag.amazonka-datasync0The key for an Amazon Web Services resource tag.amazonka-datasync)(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';)amazonka-datasync#Describes the detailed result of a  TaskExecution operation. This result includes the time in milliseconds spent in each phase, the status of the task execution, and the errors encountered.See:  smart constructor.amazonka-datasyncErrors that DataSync encountered during execution of the task. You can use this error code to help troubleshoot issues.amazonka-datasyncDetailed description of an error that was encountered during the task execution. You can use this information to help troubleshoot issues.amazonka-datasyncThe total time in milliseconds that DataSync spent in the PREPARING phase.amazonka-datasync"The status of the PREPARING phase.amazonka-datasyncThe total time in milliseconds that DataSync took to transfer the file from the source to the destination location.amazonka-datasyncThe total time in milliseconds that DataSync spent in the TRANSFERRING phase.amazonka-datasync%The status of the TRANSFERRING phase.amazonka-datasyncThe total time in milliseconds that DataSync spent in the VERIFYING phase.amazonka-datasync"The status of the VERIFYING phase.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Errors that DataSync encountered during execution of the task. You can use this error code to help troubleshoot issues.,  - Detailed description of an error that was encountered during the task execution. You can use this information to help troubleshoot issues.,  - The total time in milliseconds that DataSync spent in the PREPARING phase., % - The status of the PREPARING phase.,  - The total time in milliseconds that DataSync took to transfer the file from the source to the destination location.,  - The total time in milliseconds that DataSync spent in the TRANSFERRING phase., ( - The status of the TRANSFERRING phase.,  - The total time in milliseconds that DataSync spent in the VERIFYING phase., % - The status of the VERIFYING phase.amazonka-datasyncErrors that DataSync encountered during execution of the task. You can use this error code to help troubleshoot issues.amazonka-datasyncDetailed description of an error that was encountered during the task execution. You can use this information to help troubleshoot issues.amazonka-datasyncThe total time in milliseconds that DataSync spent in the PREPARING phase.amazonka-datasync"The status of the PREPARING phase.amazonka-datasyncThe total time in milliseconds that DataSync took to transfer the file from the source to the destination location.amazonka-datasyncThe total time in milliseconds that DataSync spent in the TRANSFERRING phase.amazonka-datasync%The status of the TRANSFERRING phase.amazonka-datasyncThe total time in milliseconds that DataSync spent in the VERIFYING phase.amazonka-datasync"The status of the VERIFYING phase.*(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?! +(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-datasync9Represents a single entry in a list of task executions. TaskExecutionListEntry returns an array that contains a list of specific invocations of a task when the  https://docs.aws.amazon.com/datasync/latest/userguide/API_ListTaskExecutions.htmlListTaskExecutions operation is called.See:  smart constructor.amazonka-datasyncThe status of a task execution.amazonka-datasync=The Amazon Resource Name (ARN) of the task that was executed.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, " - The status of a task execution.,  - The Amazon Resource Name (ARN) of the task that was executed.amazonka-datasyncThe status of a task execution.amazonka-datasync=The Amazon Resource Name (ARN) of the task that was executed.,(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?A-(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-datasyncYou can use API filters to narrow down the list of resources returned by  ListTasks. For example, to retrieve all tasks on a source location, you can use  ListTasks with filter name  LocationId and Operator Equals with the ARN for the location.For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/query-resources.htmlfiltering DataSync resources.See:  smart constructor.amazonka-datasyncThe name of the filter being used. Each API call supports a list of filters that are available for it. For example,  LocationId for  ListTasks.amazonka-datasyncThe values that you want to filter for. For example, you might want to display only tasks for a specific destination location.amazonka-datasyncThe operator that is used to compare filter values (for example, Equals or Contains).amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The name of the filter being used. Each API call supports a list of filters that are available for it. For example,  LocationId for  ListTasks.,  - The values that you want to filter for. For example, you might want to display only tasks for a specific destination location.,  - The operator that is used to compare filter values (for example, Equals or Contains).amazonka-datasyncThe name of the filter being used. Each API call supports a list of filters that are available for it. For example,  LocationId for  ListTasks.amazonka-datasyncThe values that you want to filter for. For example, you might want to display only tasks for a specific destination location.amazonka-datasyncThe operator that is used to compare filter values (for example, Equals or Contains).amazonka-datasyncamazonka-datasync  .(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?/(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-datasyncSpecifies the schedule you want your task to use for repeated executions. For more information, see  https://docs.aws.amazon.com/AmazonCloudWatch/latest/events/ScheduledEvents.htmlSchedule Expressions for Rules.See:  smart constructor.amazonka-datasyncA cron expression that specifies when DataSync initiates a scheduled transfer from a source to a destination location.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A cron expression that specifies when DataSync initiates a scheduled transfer from a source to a destination location.amazonka-datasyncA cron expression that specifies when DataSync initiates a scheduled transfer from a source to a destination location.amazonka-datasync0(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";? 1(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';9amazonka-datasync.Represents a single entry in a list of tasks.  TaskListEntry; returns an array that contains a list of tasks when the  https://docs.aws.amazon.com/datasync/latest/userguide/API_ListTasks.html ListTasks operation is called. A task includes the source and destination file systems to sync and the options to use for the tasks.See:  smart constructor.amazonka-datasyncThe name of the task.amazonka-datasyncThe status of the task.amazonka-datasync+The Amazon Resource Name (ARN) of the task.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The name of the task.,  - The status of the task., . - The Amazon Resource Name (ARN) of the task.amazonka-datasyncThe name of the task.amazonka-datasyncThe status of the task.amazonka-datasync+The Amazon Resource Name (ARN) of the task.  2(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?3(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";? 4(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";? 5(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';z~ amazonka-datasyncConfigures your DataSync task settings. These options include how DataSync handles files, objects, and their associated metadata. You also can specify how DataSync verifies data integrity, set bandwidth limits for your task, among other options.Each task setting has a default value. Unless you need to, you don't have to configure any of these Options before starting your task.See:  smart constructor.amazonka-datasyncSpecifies whether to preserve metadata indicating the last time a file was read or written to. If you set Atime to  BEST_EFFORT., DataSync attempts to preserve the original Atime attribute on all source files (that is, the version before the  PREPARING phase of the task execution).The behavior of Atime isn't fully standard across platforms, so DataSync can only do this on a best-effort basis.Default value:  BEST_EFFORT BEST_EFFORT#: Attempt to preserve the per-file Atime value (recommended).NONE : Ignore Atime.If Atime is set to  BEST_EFFORT, Mtime must be set to PRESERVE.If Atime is set to NONE, Mtime must also be NONE.amazonka-datasyncLimits the bandwidth used by a DataSync task. For example, if you want DataSync to use a maximum of 1 MB, set this value to 1048576 ( =1024*1024).amazonka-datasync8Specifies the POSIX group ID (GID) of the file's owners.For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.html#metadata-copiedMetadata copied by DataSync.Default value:  INT_VALUE-. This preserves the integer value of the ID. INT_VALUE: Preserve the integer value of user ID (UID) and GID (recommended).NONE: Ignore UID and GID.amazonka-datasyncSpecifies the type of logs that DataSync publishes to a Amazon CloudWatch Logs log group. To specify the log group, see  https://docs.aws.amazon.com/datasync/latest/userguide/API_CreateTask.html#DataSync-CreateTask-request-CloudWatchLogGroupArnCloudWatchLogGroupArn. If you set LogLevel to OFF, no logs are published. BASIC= publishes logs on errors for individual files transferred. TRANSFER publishes logs for every file or object that is transferred and integrity checked.amazonka-datasyncSpecifies whether to preserve metadata indicating the last time that a file was written to before the  PREPARING phase of your task execution. This option is required when you need to run the a task more than once.Default Value: PRESERVEPRESERVE: Preserve original Mtime (recommended)NONE : Ignore Mtime.If Mtime is set to PRESERVE, Atime must be set to  BEST_EFFORT.If Mtime is set to NONE, Atime must also be set to NONE.amazonka-datasyncSpecifies whether object tags are preserved when transferring between object storage systems. If you want your DataSync task to ignore object tags, specify the NONE value.Default Value: PRESERVEamazonka-datasyncSpecifies whether data at the destination location should be overwritten or preserved. If set to NEVER, a destination file for example will not be replaced by a source file (even if the destination file differs from the source file). If you modify files in the destination and you sync the files, you can use this value to protect against overwriting those changes.Some storage classes have specific behaviors that can affect your Amazon S3 storage cost. For detailed information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classesConsiderations when working with Amazon S3 storage classes in DataSync .amazonka-datasyncSpecifies which users or groups can access a file for a specific purpose such as reading, writing, or execution of the file.For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.html#metadata-copiedMetadata copied by DataSync.Default value: PRESERVEPRESERVE1: Preserve POSIX-style permissions (recommended).NONE: Ignore permissions.>DataSync can preserve extant permissions of a source location.amazonka-datasyncSpecifies whether files in the destination location that don't exist in the source should be preserved. This option can affect your Amazon S3 storage cost. If your task deletes objects, you might incur minimum storage duration charges for certain storage classes. For detailed information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classesConsiderations when working with Amazon S3 storage classes in DataSync .Default value: PRESERVEPRESERVE.: Ignore such destination files (recommended).REMOVE=: Delete destination files that aren@t present in the source.amazonka-datasyncSpecifies whether DataSync should preserve the metadata of block and character devices in the source location and recreate the files with that device name and metadata on the destination. DataSync copies only the name and metadata of such devices.DataSync can't copy the actual contents of these devices because they're nonterminal and don't return an end-of-file (EOF) marker.Default value: NONENONE': Ignore special devices (recommended).PRESERVE: Preserve character and block device metadata. This option currently isn't supported for Amazon EFS.amazonka-datasyncSpecifies which components of the SMB security descriptor are copied from source to destination objects.This value is only used for transfers between SMB and Amazon FSx for Windows File Server locations or between two FSx for Windows File Server locations. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.htmlhow DataSync handles metadata.Default value:  OWNER_DACL OWNER_DACL: For each copied object, DataSync copies the following metadata:The object owner.NTFS discretionary access control lists (DACLs), which determine whether to grant access to an object.DataSync won't copy NTFS system access control lists (SACLs) with this option.OWNER_DACL_SACL: For each copied object, DataSync copies the following metadata:The object owner.NTFS discretionary access control lists (DACLs), which determine whether to grant access to an object.SACLs, which are used by administrators to log attempts to access a secured object.Copying SACLs requires granting additional permissions to the Windows user that DataSync uses to access your SMB location. For information about choosing a user that ensures sufficient permissions to files, folders, and metadata, see  create-smb-location.html#SMBuseruser.NONE: None of the SMB security descriptor components are copied. Destination objects are owned by the user that was provided for accessing the destination location. DACLs and SACLs are set based on the destination server@s configuration.amazonka-datasyncSpecifies whether tasks should be queued before executing the tasks. The default is ENABLED', which means the tasks will be queued.If you use the same agent to run multiple tasks, you can enable the tasks to run in series. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/run-task.html#queue-task-executionQueueing task executions.amazonka-datasyncDetermines whether DataSync transfers only the data and metadata that differ between the source and the destination location or transfers all the content from the source (without comparing what's in the destination).CHANGED: DataSync copies only data or metadata that is new or different content from the source location to the destination location.ALL: DataSync copies all source location content to the destination (without comparing what's in the destination).amazonka-datasync6Specifies the POSIX user ID (UID) of the file's owner.For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.html#metadata-copiedMetadata copied by DataSync.Default value:  INT_VALUE-. This preserves the integer value of the ID. INT_VALUE: Preserve the integer value of UID and group ID (GID) (recommended).NONE: Ignore UID and GID.amazonka-datasyncSpecifies how and when DataSync checks the integrity of your data during a transfer.Default value: POINT_IN_TIME_CONSISTENTONLY_FILES_TRANSFERRED (recommended): DataSync calculates the checksum of transferred files and metadata at the source location. At the end of the transfer, DataSync then compares this checksum to the checksum calculated on those files at the destination.We recommend this option when transferring to S3 Glacier Flexible Retrieval or S3 Glacier Deep Archive storage classes. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classes5Storage class considerations with Amazon S3 locations.POINT_IN_TIME_CONSISTENT: At the end of the transfer, DataSync scans the entire source and destination to verify that both locations are fully synchronized.You can't use this option when transferring to S3 Glacier Flexible Retrieval or S3 Glacier Deep Archive storage classes. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classes5Storage class considerations with Amazon S3 locations.NONE: DataSync doesn't run additional verification at the end of the transfer. All data transmissions are still integrity-checked with checksum verification during the transfer.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies whether to preserve metadata indicating the last time a file was read or written to. If you set Atime to  BEST_EFFORT., DataSync attempts to preserve the original Atime attribute on all source files (that is, the version before the  PREPARING phase of the task execution).The behavior of Atime isn't fully standard across platforms, so DataSync can only do this on a best-effort basis.Default value:  BEST_EFFORT BEST_EFFORT#: Attempt to preserve the per-file Atime value (recommended).NONE : Ignore Atime.If Atime is set to  BEST_EFFORT, Mtime must be set to PRESERVE.If Atime is set to NONE, Mtime must also be NONE.,  - Limits the bandwidth used by a DataSync task. For example, if you want DataSync to use a maximum of 1 MB, set this value to 1048576 ( =1024*1024)., ; - Specifies the POSIX group ID (GID) of the file's owners.For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.html#metadata-copiedMetadata copied by DataSync.Default value:  INT_VALUE-. This preserves the integer value of the ID. INT_VALUE: Preserve the integer value of user ID (UID) and GID (recommended).NONE: Ignore UID and GID.,  - Specifies the type of logs that DataSync publishes to a Amazon CloudWatch Logs log group. To specify the log group, see  https://docs.aws.amazon.com/datasync/latest/userguide/API_CreateTask.html#DataSync-CreateTask-request-CloudWatchLogGroupArnCloudWatchLogGroupArn. If you set LogLevel to OFF, no logs are published. BASIC= publishes logs on errors for individual files transferred. TRANSFER publishes logs for every file or object that is transferred and integrity checked.,  - Specifies whether to preserve metadata indicating the last time that a file was written to before the  PREPARING phase of your task execution. This option is required when you need to run the a task more than once.Default Value: PRESERVEPRESERVE: Preserve original Mtime (recommended)NONE : Ignore Mtime.If Mtime is set to PRESERVE, Atime must be set to  BEST_EFFORT.If Mtime is set to NONE, Atime must also be set to NONE.,  - Specifies whether object tags are preserved when transferring between object storage systems. If you want your DataSync task to ignore object tags, specify the NONE value.Default Value: PRESERVE,  - Specifies whether data at the destination location should be overwritten or preserved. If set to NEVER, a destination file for example will not be replaced by a source file (even if the destination file differs from the source file). If you modify files in the destination and you sync the files, you can use this value to protect against overwriting those changes.Some storage classes have specific behaviors that can affect your Amazon S3 storage cost. For detailed information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classesConsiderations when working with Amazon S3 storage classes in DataSync .,  - Specifies which users or groups can access a file for a specific purpose such as reading, writing, or execution of the file.For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.html#metadata-copiedMetadata copied by DataSync.Default value: PRESERVEPRESERVE1: Preserve POSIX-style permissions (recommended).NONE: Ignore permissions.>DataSync can preserve extant permissions of a source location.,  - Specifies whether files in the destination location that don't exist in the source should be preserved. This option can affect your Amazon S3 storage cost. If your task deletes objects, you might incur minimum storage duration charges for certain storage classes. For detailed information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classesConsiderations when working with Amazon S3 storage classes in DataSync .Default value: PRESERVEPRESERVE.: Ignore such destination files (recommended).REMOVE=: Delete destination files that aren@t present in the source.,  - Specifies whether DataSync should preserve the metadata of block and character devices in the source location and recreate the files with that device name and metadata on the destination. DataSync copies only the name and metadata of such devices.DataSync can't copy the actual contents of these devices because they're nonterminal and don't return an end-of-file (EOF) marker.Default value: NONENONE': Ignore special devices (recommended).PRESERVE: Preserve character and block device metadata. This option currently isn't supported for Amazon EFS.,  - Specifies which components of the SMB security descriptor are copied from source to destination objects.This value is only used for transfers between SMB and Amazon FSx for Windows File Server locations or between two FSx for Windows File Server locations. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.htmlhow DataSync handles metadata.Default value:  OWNER_DACL OWNER_DACL: For each copied object, DataSync copies the following metadata:The object owner.NTFS discretionary access control lists (DACLs), which determine whether to grant access to an object.DataSync won't copy NTFS system access control lists (SACLs) with this option.OWNER_DACL_SACL: For each copied object, DataSync copies the following metadata:The object owner.NTFS discretionary access control lists (DACLs), which determine whether to grant access to an object.SACLs, which are used by administrators to log attempts to access a secured object.Copying SACLs requires granting additional permissions to the Windows user that DataSync uses to access your SMB location. For information about choosing a user that ensures sufficient permissions to files, folders, and metadata, see  create-smb-location.html#SMBuseruser.NONE: None of the SMB security descriptor components are copied. Destination objects are owned by the user that was provided for accessing the destination location. DACLs and SACLs are set based on the destination server@s configuration.,  - Specifies whether tasks should be queued before executing the tasks. The default is ENABLED', which means the tasks will be queued.If you use the same agent to run multiple tasks, you can enable the tasks to run in series. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/run-task.html#queue-task-executionQueueing task executions.,  - Determines whether DataSync transfers only the data and metadata that differ between the source and the destination location or transfers all the content from the source (without comparing what's in the destination).CHANGED: DataSync copies only data or metadata that is new or different content from the source location to the destination location.ALL: DataSync copies all source location content to the destination (without comparing what's in the destination)., 9 - Specifies the POSIX user ID (UID) of the file's owner.For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.html#metadata-copiedMetadata copied by DataSync.Default value:  INT_VALUE-. This preserves the integer value of the ID. INT_VALUE: Preserve the integer value of UID and group ID (GID) (recommended).NONE: Ignore UID and GID.,  - Specifies how and when DataSync checks the integrity of your data during a transfer.Default value: POINT_IN_TIME_CONSISTENTONLY_FILES_TRANSFERRED (recommended): DataSync calculates the checksum of transferred files and metadata at the source location. At the end of the transfer, DataSync then compares this checksum to the checksum calculated on those files at the destination.We recommend this option when transferring to S3 Glacier Flexible Retrieval or S3 Glacier Deep Archive storage classes. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classes5Storage class considerations with Amazon S3 locations.POINT_IN_TIME_CONSISTENT: At the end of the transfer, DataSync scans the entire source and destination to verify that both locations are fully synchronized.You can't use this option when transferring to S3 Glacier Flexible Retrieval or S3 Glacier Deep Archive storage classes. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classes5Storage class considerations with Amazon S3 locations.NONE: DataSync doesn't run additional verification at the end of the transfer. All data transmissions are still integrity-checked with checksum verification during the transfer.amazonka-datasyncSpecifies whether to preserve metadata indicating the last time a file was read or written to. If you set Atime to  BEST_EFFORT., DataSync attempts to preserve the original Atime attribute on all source files (that is, the version before the  PREPARING phase of the task execution).The behavior of Atime isn't fully standard across platforms, so DataSync can only do this on a best-effort basis.Default value:  BEST_EFFORT BEST_EFFORT#: Attempt to preserve the per-file Atime value (recommended).NONE : Ignore Atime.If Atime is set to  BEST_EFFORT, Mtime must be set to PRESERVE.If Atime is set to NONE, Mtime must also be NONE.amazonka-datasyncLimits the bandwidth used by a DataSync task. For example, if you want DataSync to use a maximum of 1 MB, set this value to 1048576 ( =1024*1024).amazonka-datasync8Specifies the POSIX group ID (GID) of the file's owners.For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.html#metadata-copiedMetadata copied by DataSync.Default value:  INT_VALUE-. This preserves the integer value of the ID. INT_VALUE: Preserve the integer value of user ID (UID) and GID (recommended).NONE: Ignore UID and GID.amazonka-datasyncSpecifies the type of logs that DataSync publishes to a Amazon CloudWatch Logs log group. To specify the log group, see  https://docs.aws.amazon.com/datasync/latest/userguide/API_CreateTask.html#DataSync-CreateTask-request-CloudWatchLogGroupArnCloudWatchLogGroupArn. If you set LogLevel to OFF, no logs are published. BASIC= publishes logs on errors for individual files transferred. TRANSFER publishes logs for every file or object that is transferred and integrity checked.amazonka-datasyncSpecifies whether to preserve metadata indicating the last time that a file was written to before the  PREPARING phase of your task execution. This option is required when you need to run the a task more than once.Default Value: PRESERVEPRESERVE: Preserve original Mtime (recommended)NONE : Ignore Mtime.If Mtime is set to PRESERVE, Atime must be set to  BEST_EFFORT.If Mtime is set to NONE, Atime must also be set to NONE.amazonka-datasyncSpecifies whether object tags are preserved when transferring between object storage systems. If you want your DataSync task to ignore object tags, specify the NONE value.Default Value: PRESERVEamazonka-datasyncSpecifies whether data at the destination location should be overwritten or preserved. If set to NEVER, a destination file for example will not be replaced by a source file (even if the destination file differs from the source file). If you modify files in the destination and you sync the files, you can use this value to protect against overwriting those changes.Some storage classes have specific behaviors that can affect your Amazon S3 storage cost. For detailed information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classesConsiderations when working with Amazon S3 storage classes in DataSync .amazonka-datasyncSpecifies which users or groups can access a file for a specific purpose such as reading, writing, or execution of the file.For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.html#metadata-copiedMetadata copied by DataSync.Default value: PRESERVEPRESERVE1: Preserve POSIX-style permissions (recommended).NONE: Ignore permissions.>DataSync can preserve extant permissions of a source location.amazonka-datasyncSpecifies whether files in the destination location that don't exist in the source should be preserved. This option can affect your Amazon S3 storage cost. If your task deletes objects, you might incur minimum storage duration charges for certain storage classes. For detailed information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classesConsiderations when working with Amazon S3 storage classes in DataSync .Default value: PRESERVEPRESERVE.: Ignore such destination files (recommended).REMOVE=: Delete destination files that aren@t present in the source.amazonka-datasyncSpecifies whether DataSync should preserve the metadata of block and character devices in the source location and recreate the files with that device name and metadata on the destination. DataSync copies only the name and metadata of such devices.DataSync can't copy the actual contents of these devices because they're nonterminal and don't return an end-of-file (EOF) marker.Default value: NONENONE': Ignore special devices (recommended).PRESERVE: Preserve character and block device metadata. This option currently isn't supported for Amazon EFS.amazonka-datasyncSpecifies which components of the SMB security descriptor are copied from source to destination objects.This value is only used for transfers between SMB and Amazon FSx for Windows File Server locations or between two FSx for Windows File Server locations. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.htmlhow DataSync handles metadata.Default value:  OWNER_DACL OWNER_DACL: For each copied object, DataSync copies the following metadata:The object owner.NTFS discretionary access control lists (DACLs), which determine whether to grant access to an object.DataSync won't copy NTFS system access control lists (SACLs) with this option.OWNER_DACL_SACL: For each copied object, DataSync copies the following metadata:The object owner.NTFS discretionary access control lists (DACLs), which determine whether to grant access to an object.SACLs, which are used by administrators to log attempts to access a secured object.Copying SACLs requires granting additional permissions to the Windows user that DataSync uses to access your SMB location. For information about choosing a user that ensures sufficient permissions to files, folders, and metadata, see  create-smb-location.html#SMBuseruser.NONE: None of the SMB security descriptor components are copied. Destination objects are owned by the user that was provided for accessing the destination location. DACLs and SACLs are set based on the destination server@s configuration.amazonka-datasyncSpecifies whether tasks should be queued before executing the tasks. The default is ENABLED', which means the tasks will be queued.If you use the same agent to run multiple tasks, you can enable the tasks to run in series. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/run-task.html#queue-task-executionQueueing task executions.amazonka-datasyncDetermines whether DataSync transfers only the data and metadata that differ between the source and the destination location or transfers all the content from the source (without comparing what's in the destination).CHANGED: DataSync copies only data or metadata that is new or different content from the source location to the destination location.ALL: DataSync copies all source location content to the destination (without comparing what's in the destination).amazonka-datasync6Specifies the POSIX user ID (UID) of the file's owner.For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/special-files.html#metadata-copiedMetadata copied by DataSync.Default value:  INT_VALUE-. This preserves the integer value of the ID. INT_VALUE: Preserve the integer value of UID and group ID (GID) (recommended).NONE: Ignore UID and GID.amazonka-datasyncSpecifies how and when DataSync checks the integrity of your data during a transfer.Default value: POINT_IN_TIME_CONSISTENTONLY_FILES_TRANSFERRED (recommended): DataSync calculates the checksum of transferred files and metadata at the source location. At the end of the transfer, DataSync then compares this checksum to the checksum calculated on those files at the destination.We recommend this option when transferring to S3 Glacier Flexible Retrieval or S3 Glacier Deep Archive storage classes. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classes5Storage class considerations with Amazon S3 locations.POINT_IN_TIME_CONSISTENT: At the end of the transfer, DataSync scans the entire source and destination to verify that both locations are fully synchronized.You can't use this option when transferring to S3 Glacier Flexible Retrieval or S3 Glacier Deep Archive storage classes. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classes5Storage class considerations with Amazon S3 locations.NONE: DataSync doesn't run additional verification at the end of the transfer. All data transmissions are still integrity-checked with checksum verification during the transfer.!!6(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%|amazonka-datasync API version  2018-11-09* of the Amazon DataSync SDK configuration.amazonka-datasyncThis exception is thrown when an error occurs in the DataSync service.amazonka-datasyncThis exception is thrown when the client submits a malformed request. (,+)*@ABCDEFOSRPQglkjhi(,+)*,+OSRPQSRglkjhilkj @ABCDEF7(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-datasyncSee:   smart constructor.amazonka-datasync The response's http status code. amazonka-datasyncTagResourceRequestSee:   smart constructor. amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the resource to apply the tag to. amazonka-datasync:Specifies the tags that you want to apply to the resource. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - Specifies the Amazon Resource Name (ARN) of the resource to apply the tag to. ,  = - Specifies the tags that you want to apply to the resource. amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the resource to apply the tag to. amazonka-datasync:Specifies the tags that you want to apply to the resource. amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  # - The response's http status code. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync   8(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-datasyncStartTaskExecutionResponseSee:   smart constructor. amazonka-datasync&The ARN of the running task execution. amazonka-datasync The response's http status code. amazonka-datasyncStartTaskExecutionRequestSee:   smart constructor. amazonka-datasyncSpecifies a list of filter rules that determines which files to exclude from a task. The list contains a single filter string that consists of the patterns to exclude. The patterns are delimited by "|" (that is, a pipe), for example, "/folder1|/folder2". amazonka-datasyncSpecifies a list of filter rules that determines which files to include when running a task. The pattern should contain a single filter string that consists of the patterns to include. The patterns are delimited by "|" (that is, a pipe), for example, "/folder1|/folder2". amazonka-datasyncSpecifies the tags that you want to apply to the Amazon Resource Name (ARN) representing the task execution.Tags are key-value pairs that help you manage, filter, and search for your DataSync resources. amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the task that you want to start. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - Specifies a list of filter rules that determines which files to exclude from a task. The list contains a single filter string that consists of the patterns to exclude. The patterns are delimited by "|" (that is, a pipe), for example, "/folder1|/folder2". ,   - Specifies a list of filter rules that determines which files to include when running a task. The pattern should contain a single filter string that consists of the patterns to include. The patterns are delimited by "|" (that is, a pipe), for example, "/folder1|/folder2". ,   - Undocumented member. ,   - Specifies the tags that you want to apply to the Amazon Resource Name (ARN) representing the task execution.Tags are key-value pairs that help you manage, filter, and search for your DataSync resources. ,   - Specifies the Amazon Resource Name (ARN) of the task that you want to start. amazonka-datasyncSpecifies a list of filter rules that determines which files to exclude from a task. The list contains a single filter string that consists of the patterns to exclude. The patterns are delimited by "|" (that is, a pipe), for example, "/folder1|/folder2". amazonka-datasyncSpecifies a list of filter rules that determines which files to include when running a task. The pattern should contain a single filter string that consists of the patterns to include. The patterns are delimited by "|" (that is, a pipe), for example, "/folder1|/folder2". amazonka-datasyncUndocumented member. amazonka-datasyncSpecifies the tags that you want to apply to the Amazon Resource Name (ARN) representing the task execution.Tags are key-value pairs that help you manage, filter, and search for your DataSync resources. amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the task that you want to start. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  ) - The ARN of the running task execution. ,  # - The response's http status code. amazonka-datasync&The ARN of the running task execution. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync   9(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-datasyncListTasksResponseSee:   smart constructor. amazonka-datasyncAn opaque string that indicates the position at which to begin returning the next list of tasks. amazonka-datasync*A list of all the tasks that are returned. amazonka-datasync The response's http status code. amazonka-datasyncListTasksRequestSee:   smart constructor. amazonka-datasyncYou can use API filters to narrow down the list of resources returned by  ListTasks. For example, to retrieve all tasks on a specific source location, you can use  ListTasks with filter name  LocationId and Operator Equals with the ARN for the location. amazonka-datasync&The maximum number of tasks to return. amazonka-datasyncAn opaque string that indicates the position at which to begin the next list of tasks. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - You can use API filters to narrow down the list of resources returned by  ListTasks. For example, to retrieve all tasks on a specific source location, you can use  ListTasks with filter name  LocationId and Operator Equals with the ARN for the location. ,  ) - The maximum number of tasks to return. ,   - An opaque string that indicates the position at which to begin the next list of tasks. amazonka-datasyncYou can use API filters to narrow down the list of resources returned by  ListTasks. For example, to retrieve all tasks on a specific source location, you can use  ListTasks with filter name  LocationId and Operator Equals with the ARN for the location. amazonka-datasync&The maximum number of tasks to return. amazonka-datasyncAn opaque string that indicates the position at which to begin the next list of tasks. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - An opaque string that indicates the position at which to begin returning the next list of tasks. ,  - - A list of all the tasks that are returned. ,  # - The response's http status code. amazonka-datasyncAn opaque string that indicates the position at which to begin returning the next list of tasks. amazonka-datasync*A list of all the tasks that are returned. amazonka-datasync The response's http status code. amazonka-datasync   :(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-datasyncListTaskExecutionsResponseSee:   smart constructor. amazonka-datasyncAn opaque string that indicates the position at which to begin returning the next list of executed tasks. amazonka-datasyncA list of executed tasks. amazonka-datasync The response's http status code. amazonka-datasyncListTaskExecutionsSee:   smart constructor. amazonka-datasync-The maximum number of executed tasks to list. amazonka-datasyncAn opaque string that indicates the position at which to begin the next list of the executed tasks. amazonka-datasyncThe Amazon Resource Name (ARN) of the task whose tasks you want to list. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  0 - The maximum number of executed tasks to list. ,   - An opaque string that indicates the position at which to begin the next list of the executed tasks. ,   - The Amazon Resource Name (ARN) of the task whose tasks you want to list. amazonka-datasync-The maximum number of executed tasks to list. amazonka-datasyncAn opaque string that indicates the position at which to begin the next list of the executed tasks. amazonka-datasyncThe Amazon Resource Name (ARN) of the task whose tasks you want to list. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - An opaque string that indicates the position at which to begin returning the next list of executed tasks. ,   - A list of executed tasks. ,  # - The response's http status code. amazonka-datasyncAn opaque string that indicates the position at which to begin returning the next list of executed tasks. amazonka-datasyncA list of executed tasks. amazonka-datasync The response's http status code. amazonka-datasync   ;(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';n amazonka-datasyncListTagsForResourceResponseSee:   smart constructor. amazonka-datasyncThe opaque string that indicates the position to begin the next list of results in the response. amazonka-datasync3An array of tags applied to the specified resource. amazonka-datasync The response's http status code. amazonka-datasyncListTagsForResourceRequestSee:   smart constructor. amazonka-datasync9Specifies how many results that you want in the response. amazonka-datasyncSpecifies an opaque string that indicates the position to begin the next list of results in the response. amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the resource that you want tag information on. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  < - Specifies how many results that you want in the response. ,   - Specifies an opaque string that indicates the position to begin the next list of results in the response. ,   - Specifies the Amazon Resource Name (ARN) of the resource that you want tag information on. amazonka-datasync9Specifies how many results that you want in the response. amazonka-datasyncSpecifies an opaque string that indicates the position to begin the next list of results in the response. amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the resource that you want tag information on. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The opaque string that indicates the position to begin the next list of results in the response. ,  6 - An array of tags applied to the specified resource. ,  # - The response's http status code. amazonka-datasyncThe opaque string that indicates the position to begin the next list of results in the response. amazonka-datasync3An array of tags applied to the specified resource. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync   <(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';́ amazonka-datasyncListLocationsResponseSee:   smart constructor. amazonka-datasync+An array that contains a list of locations. amazonka-datasyncAn opaque string that indicates the position at which to begin returning the next list of locations. amazonka-datasync The response's http status code. amazonka-datasyncListLocationsRequestSee:   smart constructor. amazonka-datasyncYou can use API filters to narrow down the list of resources returned by  ListLocations. For example, to retrieve all tasks on a specific source location, you can use  ListLocations with filter name LocationType S3 and Operator Equals. amazonka-datasync*The maximum number of locations to return. amazonka-datasyncAn opaque string that indicates the position at which to begin the next list of locations. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - You can use API filters to narrow down the list of resources returned by  ListLocations. For example, to retrieve all tasks on a specific source location, you can use  ListLocations with filter name LocationType S3 and Operator Equals. ,  - - The maximum number of locations to return. ,   - An opaque string that indicates the position at which to begin the next list of locations. amazonka-datasyncYou can use API filters to narrow down the list of resources returned by  ListLocations. For example, to retrieve all tasks on a specific source location, you can use  ListLocations with filter name LocationType S3 and Operator Equals. amazonka-datasync*The maximum number of locations to return. amazonka-datasyncAn opaque string that indicates the position at which to begin the next list of locations. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  . - An array that contains a list of locations. ,   - An opaque string that indicates the position at which to begin returning the next list of locations. ,  # - The response's http status code. amazonka-datasync+An array that contains a list of locations. amazonka-datasyncAn opaque string that indicates the position at which to begin returning the next list of locations. amazonka-datasync The response's http status code. amazonka-datasync   =(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-datasyncListAgentsResponseSee:   smart constructor. amazonka-datasync!A list of agents in your account. amazonka-datasyncAn opaque string that indicates the position at which to begin returning the next list of agents. amazonka-datasync The response's http status code. amazonka-datasyncListAgentsRequestSee:   smart constructor. amazonka-datasync%The maximum number of agents to list. amazonka-datasyncAn opaque string that indicates the position at which to begin the next list of agents. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  ( - The maximum number of agents to list. ,   - An opaque string that indicates the position at which to begin the next list of agents. amazonka-datasync%The maximum number of agents to list. amazonka-datasyncAn opaque string that indicates the position at which to begin the next list of agents. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  $ - A list of agents in your account. ,   - An opaque string that indicates the position at which to begin returning the next list of agents. ,  # - The response's http status code. amazonka-datasync!A list of agents in your account. amazonka-datasyncAn opaque string that indicates the position at which to begin returning the next list of agents. amazonka-datasync The response's http status code. amazonka-datasync   >(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';! amazonka-datasyncDescribeTaskExecutionResponseSee:   smart constructor. amazonka-datasyncThe physical number of bytes transferred over the network after compression was applied. In most cases, this number is less than BytesTransferred$ unless the data isn't compressible. amazonka-datasyncThe total number of bytes that are involved in the transfer. For the number of bytes sent over the network, see BytesCompressed. amazonka-datasyncThe number of logical bytes written to the destination Amazon Web Services storage resource. amazonka-datasyncThe estimated physical number of bytes that is to be transferred over the network. amazonka-datasyncThe expected number of files that is to be transferred over the network. This value is calculated during the  PREPARING phase before the  TRANSFERRING phase of the task execution. This value is the expected number of files to be transferred. It's calculated based on comparing the content of the source and destination locations and finding the delta that needs to be transferred. amazonka-datasyncA list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. amazonka-datasyncThe actual number of files that was transferred over the network. This value is calculated and updated on an ongoing basis during the  TRANSFERRING phase of the task execution. It's updated periodically when each file is read from the source and sent over the network.If failures occur during a transfer, this value can be less than EstimatedFilesToTransfer6. In some cases, this value can also be greater than EstimatedFilesToTransfer. This element is implementation-specific for some location types, so don't use it as an indicator for a correct file number or to monitor your task execution. amazonka-datasyncA list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. amazonka-datasync!The result of the task execution. amazonka-datasync-The time that the task execution was started. amazonka-datasync!The status of the task execution.For detailed information about task execution statuses, see Understanding Task Statuses in the DataSync User Guide. amazonka-datasyncThe Amazon Resource Name (ARN) of the task execution that was described. TaskExecutionArn is hierarchical and includes TaskArn! for the task that was executed.For example, a  TaskExecution value with the ARN arn:aws:datasync:us-east-1:111222333444:task/task-0208075f79cedf4a2/execution/exec-08ef1e88ec491019b" executed the task with the ARN arn:aws:datasync:us-east-1:111222333444:task/task-0208075f79cedf4a2. amazonka-datasync The response's http status code. amazonka-datasyncDescribeTaskExecutionRequestSee:   smart constructor. amazonka-datasyncThe Amazon Resource Name (ARN) of the task that is being executed. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) of the task that is being executed. amazonka-datasyncThe Amazon Resource Name (ARN) of the task that is being executed. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The physical number of bytes transferred over the network after compression was applied. In most cases, this number is less than BytesTransferred$ unless the data isn't compressible. ,   - The total number of bytes that are involved in the transfer. For the number of bytes sent over the network, see BytesCompressed. ,   - The number of logical bytes written to the destination Amazon Web Services storage resource. ,   - The estimated physical number of bytes that is to be transferred over the network. ,   - The expected number of files that is to be transferred over the network. This value is calculated during the  PREPARING phase before the  TRANSFERRING phase of the task execution. This value is the expected number of files to be transferred. It's calculated based on comparing the content of the source and destination locations and finding the delta that needs to be transferred. ,   - A list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. ,   - The actual number of files that was transferred over the network. This value is calculated and updated on an ongoing basis during the  TRANSFERRING phase of the task execution. It's updated periodically when each file is read from the source and sent over the network.If failures occur during a transfer, this value can be less than EstimatedFilesToTransfer6. In some cases, this value can also be greater than EstimatedFilesToTransfer. This element is implementation-specific for some location types, so don't use it as an indicator for a correct file number or to monitor your task execution. ,   - A list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. ,   - Undocumented member. ,  $ - The result of the task execution. ,  0 - The time that the task execution was started. ,  $ - The status of the task execution.For detailed information about task execution statuses, see Understanding Task Statuses in the DataSync User Guide. ,   - The Amazon Resource Name (ARN) of the task execution that was described. TaskExecutionArn is hierarchical and includes TaskArn! for the task that was executed.For example, a  TaskExecution value with the ARN arn:aws:datasync:us-east-1:111222333444:task/task-0208075f79cedf4a2/execution/exec-08ef1e88ec491019b" executed the task with the ARN arn:aws:datasync:us-east-1:111222333444:task/task-0208075f79cedf4a2. ,  # - The response's http status code. amazonka-datasyncThe physical number of bytes transferred over the network after compression was applied. In most cases, this number is less than BytesTransferred$ unless the data isn't compressible. amazonka-datasyncThe total number of bytes that are involved in the transfer. For the number of bytes sent over the network, see BytesCompressed. amazonka-datasyncThe number of logical bytes written to the destination Amazon Web Services storage resource. amazonka-datasyncThe estimated physical number of bytes that is to be transferred over the network. amazonka-datasyncThe expected number of files that is to be transferred over the network. This value is calculated during the  PREPARING phase before the  TRANSFERRING phase of the task execution. This value is the expected number of files to be transferred. It's calculated based on comparing the content of the source and destination locations and finding the delta that needs to be transferred. amazonka-datasyncA list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. amazonka-datasyncThe actual number of files that was transferred over the network. This value is calculated and updated on an ongoing basis during the  TRANSFERRING phase of the task execution. It's updated periodically when each file is read from the source and sent over the network.If failures occur during a transfer, this value can be less than EstimatedFilesToTransfer6. In some cases, this value can also be greater than EstimatedFilesToTransfer. This element is implementation-specific for some location types, so don't use it as an indicator for a correct file number or to monitor your task execution. amazonka-datasyncA list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. amazonka-datasyncUndocumented member. amazonka-datasync!The result of the task execution. amazonka-datasync-The time that the task execution was started. amazonka-datasync!The status of the task execution.For detailed information about task execution statuses, see Understanding Task Statuses in the DataSync User Guide. amazonka-datasyncThe Amazon Resource Name (ARN) of the task execution that was described. TaskExecutionArn is hierarchical and includes TaskArn! for the task that was executed.For example, a  TaskExecution value with the ARN arn:aws:datasync:us-east-1:111222333444:task/task-0208075f79cedf4a2/execution/exec-08ef1e88ec491019b" executed the task with the ARN arn:aws:datasync:us-east-1:111222333444:task/task-0208075f79cedf4a2. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync $ $ ?(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';(T( amazonka-datasyncDescribeTaskResponseSee:   smart constructor. amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon CloudWatch log group that was used to monitor and log events in the task.For more information on these groups, see Working with Log Groups and Log Streams in the Amazon CloudWatch User Guide. amazonka-datasync#The time that the task was created. amazonka-datasyncThe Amazon Resource Name (ARN) of the task execution that is transferring files. amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon Web Services storage resource's location. amazonka-datasyncThe Amazon Resource Names (ARNs) of the network interfaces created for your destination location. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/datasync-network.html#required-network-interfacesNetwork interface requirements. amazonka-datasyncErrors that DataSync encountered during execution of the task. You can use this error code to help troubleshoot issues. amazonka-datasyncDetailed description of an error that was encountered during the task execution. You can use this information to help troubleshoot issues. amazonka-datasyncA list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. amazonka-datasyncA list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. amazonka-datasync(The name of the task that was described. amazonka-datasyncThe Amazon Resource Name (ARN) of the task that was described. amazonka-datasync The response's http status code. amazonka-datasyncDescribeTaskRequestSee:   smart constructor. amazonka-datasync7The Amazon Resource Name (ARN) of the task to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  : - The Amazon Resource Name (ARN) of the task to describe. amazonka-datasync7The Amazon Resource Name (ARN) of the task to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) of the Amazon CloudWatch log group that was used to monitor and log events in the task.For more information on these groups, see Working with Log Groups and Log Streams in the Amazon CloudWatch User Guide. ,  & - The time that the task was created. ,   - The Amazon Resource Name (ARN) of the task execution that is transferring files. ,   - The Amazon Resource Name (ARN) of the Amazon Web Services storage resource's location. ,   - The Amazon Resource Names (ARNs) of the network interfaces created for your destination location. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/datasync-network.html#required-network-interfacesNetwork interface requirements. ,   - Errors that DataSync encountered during execution of the task. You can use this error code to help troubleshoot issues. ,   - Detailed description of an error that was encountered during the task execution. You can use this information to help troubleshoot issues. ,   - A list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. ,   - A list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. ,  + - The name of the task that was described. ,  ? - The configuration options that control the behavior of the StartTaskExecution operation. Some options include preserving file or object metadata and verifying data integrity.You can override these options for each task execution. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/API_StartTaskExecution.htmlStartTaskExecution. ,   - The schedule used to periodically transfer files from a source to a destination location. ,   - The Amazon Resource Name (ARN) of the source file system's location. ,   - The Amazon Resource Names (ARNs) of the network interfaces created for your source location. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/datasync-network.html#required-network-interfacesNetwork interface requirements. ,  - - The status of the task that was described.For detailed information about task execution statuses, see Understanding Task Statuses in the DataSync User Guide. ,   - The Amazon Resource Name (ARN) of the task that was described. ,  # - The response's http status code. amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon CloudWatch log group that was used to monitor and log events in the task.For more information on these groups, see Working with Log Groups and Log Streams in the Amazon CloudWatch User Guide. amazonka-datasync#The time that the task was created. amazonka-datasyncThe Amazon Resource Name (ARN) of the task execution that is transferring files. amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon Web Services storage resource's location. amazonka-datasyncThe Amazon Resource Names (ARNs) of the network interfaces created for your destination location. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/datasync-network.html#required-network-interfacesNetwork interface requirements. amazonka-datasyncErrors that DataSync encountered during execution of the task. You can use this error code to help troubleshoot issues. amazonka-datasyncDetailed description of an error that was encountered during the task execution. You can use this information to help troubleshoot issues. amazonka-datasyncA list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. amazonka-datasyncA list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync. amazonka-datasync(The name of the task that was described. amazonka-datasyncThe Amazon Resource Name (ARN) of the task that was described. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync * * @(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';7  amazonka-datasyncDescribeLocationSmbResponseSee:   smart constructor. amazonka-datasyncThe Amazon Resource Name (ARN) of the source SMB file system location that is created. amazonka-datasync+The time that the SMB location was created. amazonka-datasync>The name of the Windows domain that the SMB server belongs to. amazonka-datasyncThe Amazon Resource Name (ARN) of the SMB location that was described. amazonka-datasync6The URL of the source SMB location that was described. amazonka-datasyncThe mount options that are available for DataSync to use to access an SMB location. amazonka-datasyncThe user who can mount the share, has the permissions to access files and folders in the SMB share. amazonka-datasync The response's http status code. amazonka-datasyncDescribeLocationSmbRequestSee:   smart constructor. amazonka-datasync?The Amazon Resource Name (ARN) of the SMB location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) of the SMB location to describe. amazonka-datasync?The Amazon Resource Name (ARN) of the SMB location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) of the source SMB file system location that is created. ,  . - The time that the SMB location was created. ,   - The name of the Windows domain that the SMB server belongs to. ,   - The Amazon Resource Name (ARN) of the SMB location that was described. ,  9 - The URL of the source SMB location that was described. ,   - The mount options that are available for DataSync to use to access an SMB location. ,   - The user who can mount the share, has the permissions to access files and folders in the SMB share. ,  # - The response's http status code. amazonka-datasyncThe Amazon Resource Name (ARN) of the source SMB file system location that is created. amazonka-datasync+The time that the SMB location was created. amazonka-datasync>The name of the Windows domain that the SMB server belongs to. amazonka-datasyncThe Amazon Resource Name (ARN) of the SMB location that was described. amazonka-datasync6The URL of the source SMB location that was described. amazonka-datasyncThe mount options that are available for DataSync to use to access an SMB location. amazonka-datasyncThe user who can mount the share, has the permissions to access files and folders in the SMB share. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync   A(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';L/ amazonka-datasyncDescribeLocationS3ResponseSee:   smart constructor. amazonka-datasyncIf you are using DataSync on an Amazon Web Services Outpost, the Amazon Resource Name (ARNs) of the EC2 agents deployed on your Outpost. For more information about launching a DataSync agent on an Amazon Web Services Outpost, see  https://docs.aws.amazon.com/datasync/latest/userguide/deploy-agents.html#outposts-agent&Deploy your DataSync agent on Outposts. amazonka-datasync8The time that the Amazon S3 bucket location was created. amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon S3 bucket or access point. amazonka-datasync5The URL of the Amazon S3 location that was described. amazonka-datasyncThe Amazon S3 storage class that you chose to store your files in when this location is used as a task destination. For more information about S3 storage classes, see  )http://aws.amazon.com/s3/storage-classes/Amazon S3 Storage Classes. Some storage classes have behaviors that can affect your S3 storage cost. For detailed information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classes?Considerations when working with S3 storage classes in DataSync. amazonka-datasync The response's http status code. amazonka-datasyncDescribeLocationS3RequestSee:   smart constructor. amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon S3 bucket location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) of the Amazon S3 bucket location to describe. amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon S3 bucket location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - If you are using DataSync on an Amazon Web Services Outpost, the Amazon Resource Name (ARNs) of the EC2 agents deployed on your Outpost. For more information about launching a DataSync agent on an Amazon Web Services Outpost, see  https://docs.aws.amazon.com/datasync/latest/userguide/deploy-agents.html#outposts-agent&Deploy your DataSync agent on Outposts. ,  ; - The time that the Amazon S3 bucket location was created. ,   - The Amazon Resource Name (ARN) of the Amazon S3 bucket or access point. ,  8 - The URL of the Amazon S3 location that was described. ,   - Undocumented member. ,   - The Amazon S3 storage class that you chose to store your files in when this location is used as a task destination. For more information about S3 storage classes, see  )http://aws.amazon.com/s3/storage-classes/Amazon S3 Storage Classes. Some storage classes have behaviors that can affect your S3 storage cost. For detailed information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classes?Considerations when working with S3 storage classes in DataSync. ,  # - The response's http status code. amazonka-datasyncIf you are using DataSync on an Amazon Web Services Outpost, the Amazon Resource Name (ARNs) of the EC2 agents deployed on your Outpost. For more information about launching a DataSync agent on an Amazon Web Services Outpost, see  https://docs.aws.amazon.com/datasync/latest/userguide/deploy-agents.html#outposts-agent&Deploy your DataSync agent on Outposts. amazonka-datasync8The time that the Amazon S3 bucket location was created. amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon S3 bucket or access point. amazonka-datasync5The URL of the Amazon S3 location that was described. amazonka-datasyncUndocumented member. amazonka-datasyncThe Amazon S3 storage class that you chose to store your files in when this location is used as a task destination. For more information about S3 storage classes, see  )http://aws.amazon.com/s3/storage-classes/Amazon S3 Storage Classes. Some storage classes have behaviors that can affect your S3 storage cost. For detailed information, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-s3-location.html#using-storage-classes?Considerations when working with S3 storage classes in DataSync. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync   B(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';^ amazonka-datasync%DescribeLocationObjectStorageResponseSee:   smart constructor. amazonka-datasyncThe access key (for example, a user name) required to authenticate with the object storage system. amazonka-datasyncThe ARNs of the DataSync agents that can securely connect with your location. amazonka-datasync'The time that the location was created. amazonka-datasync.The ARN of the object storage system location. amazonka-datasync.The URL of the object storage system location. amazonka-datasyncThe self-signed certificate that DataSync uses to securely authenticate with your object storage system. amazonka-datasyncThe port that your object storage server accepts inbound network traffic on (for example, port 443). amazonka-datasyncThe protocol that your object storage system uses to communicate. amazonka-datasync The response's http status code. amazonka-datasync$DescribeLocationObjectStorageRequestSee:   smart constructor. amazonka-datasyncThe Amazon Resource Name (ARN) of the object storage system location that you want information about. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) of the object storage system location that you want information about. amazonka-datasyncThe Amazon Resource Name (ARN) of the object storage system location that you want information about. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The access key (for example, a user name) required to authenticate with the object storage system. ,   - The ARNs of the DataSync agents that can securely connect with your location. ,  * - The time that the location was created. ,  1 - The ARN of the object storage system location. ,  1 - The URL of the object storage system location. ,   - The self-signed certificate that DataSync uses to securely authenticate with your object storage system.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data. ,   - The port that your object storage server accepts inbound network traffic on (for example, port 443). ,   - The protocol that your object storage system uses to communicate. ,  # - The response's http status code. amazonka-datasyncThe access key (for example, a user name) required to authenticate with the object storage system. amazonka-datasyncThe ARNs of the DataSync agents that can securely connect with your location. amazonka-datasync'The time that the location was created. amazonka-datasync.The ARN of the object storage system location. amazonka-datasync.The URL of the object storage system location. amazonka-datasyncThe self-signed certificate that DataSync uses to securely authenticate with your object storage system.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data. amazonka-datasyncThe port that your object storage server accepts inbound network traffic on (for example, port 443). amazonka-datasyncThe protocol that your object storage system uses to communicate. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync   C(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';i amazonka-datasyncDescribeLocationNfsResponseSee:   smart constructor. amazonka-datasync+The time that the NFS location was created. amazonka-datasyncThe Amazon Resource Name (ARN) of the NFS location that was described. amazonka-datasync6The URL of the source NFS location that was described. amazonka-datasyncThe NFS mount options that DataSync used to mount your NFS share. amazonka-datasync The response's http status code. amazonka-datasyncDescribeLocationNfsRequestSee:   smart constructor. amazonka-datasync?The Amazon Resource Name (ARN) of the NFS location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) of the NFS location to describe. amazonka-datasync?The Amazon Resource Name (ARN) of the NFS location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  . - The time that the NFS location was created. ,   - The Amazon Resource Name (ARN) of the NFS location that was described. ,  9 - The URL of the source NFS location that was described. ,   - The NFS mount options that DataSync used to mount your NFS share. ,   - Undocumented member. ,  # - The response's http status code. amazonka-datasync+The time that the NFS location was created. amazonka-datasyncThe Amazon Resource Name (ARN) of the NFS location that was described. amazonka-datasync6The URL of the source NFS location that was described. amazonka-datasyncThe NFS mount options that DataSync used to mount your NFS share. amazonka-datasyncUndocumented member. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync   D(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';K amazonka-datasyncSee:   smart constructor. amazonka-datasyncThe ARNs of the agents that are used to connect to the HDFS cluster. amazonka-datasyncThe type of authentication used to determine the identity of the user. amazonka-datasync;The size of the data blocks to write into the HDFS cluster. amazonka-datasync,The time that the HDFS location was created. amazonka-datasyncThe Kerberos principal with access to the files and folders on the HDFS cluster. This parameter is used if the AuthenticationType is defined as KERBEROS. amazonka-datasync:The URI of the HDFS cluster's Key Management Server (KMS). amazonka-datasync%The ARN of the HDFS cluster location. amazonka-datasync%The URI of the HDFS cluster location. amazonka-datasync,The NameNode that manage the HDFS namespace. amazonka-datasyncThe Quality of Protection (QOP) configuration specifies the Remote Procedure Call (RPC) and data transfer protection settings configured on the Hadoop Distributed File System (HDFS) cluster. amazonka-datasyncThe number of DataNodes to replicate the data to when writing to the HDFS cluster. amazonka-datasyncThe user name used to identify the client on the host operating system. This parameter is used if the AuthenticationType is defined as SIMPLE. amazonka-datasync The response's http status code. amazonka-datasyncSee:   smart constructor. amazonka-datasyncThe Amazon Resource Name (ARN) of the HDFS cluster location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) of the HDFS cluster location to describe. amazonka-datasyncThe Amazon Resource Name (ARN) of the HDFS cluster location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The ARNs of the agents that are used to connect to the HDFS cluster. ,   - The type of authentication used to determine the identity of the user. ,  > - The size of the data blocks to write into the HDFS cluster. ,  / - The time that the HDFS location was created. ,   - The Kerberos principal with access to the files and folders on the HDFS cluster. This parameter is used if the AuthenticationType is defined as KERBEROS. ,  = - The URI of the HDFS cluster's Key Management Server (KMS). ,  ( - The ARN of the HDFS cluster location. ,  ( - The URI of the HDFS cluster location. ,  / - The NameNode that manage the HDFS namespace. ,   - The Quality of Protection (QOP) configuration specifies the Remote Procedure Call (RPC) and data transfer protection settings configured on the Hadoop Distributed File System (HDFS) cluster. ,   - The number of DataNodes to replicate the data to when writing to the HDFS cluster. ,   - The user name used to identify the client on the host operating system. This parameter is used if the AuthenticationType is defined as SIMPLE. ,  # - The response's http status code. amazonka-datasyncThe ARNs of the agents that are used to connect to the HDFS cluster. amazonka-datasyncThe type of authentication used to determine the identity of the user. amazonka-datasync;The size of the data blocks to write into the HDFS cluster. amazonka-datasync,The time that the HDFS location was created. amazonka-datasyncThe Kerberos principal with access to the files and folders on the HDFS cluster. This parameter is used if the AuthenticationType is defined as KERBEROS. amazonka-datasync:The URI of the HDFS cluster's Key Management Server (KMS). amazonka-datasync%The ARN of the HDFS cluster location. amazonka-datasync%The URI of the HDFS cluster location. amazonka-datasync,The NameNode that manage the HDFS namespace. amazonka-datasyncThe Quality of Protection (QOP) configuration specifies the Remote Procedure Call (RPC) and data transfer protection settings configured on the Hadoop Distributed File System (HDFS) cluster. amazonka-datasyncThe number of DataNodes to replicate the data to when writing to the HDFS cluster. amazonka-datasyncThe user name used to identify the client on the host operating system. This parameter is used if the AuthenticationType is defined as SIMPLE. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync " " E(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';= amazonka-datasyncSee:   smart constructor. amazonka-datasyncThe time that the FSx for Windows File Server location was created. amazonka-datasyncThe name of the Windows domain that the FSx for Windows File Server belongs to. amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for Windows File Server location that was described. amazonka-datasyncThe URL of the FSx for Windows File Server location that was described. amazonka-datasyncThe Amazon Resource Names (ARNs) of the security groups that are configured for the FSx for Windows File Server file system. amazonka-datasyncThe user who has the permissions to access files and folders in the FSx for Windows File Server file system. amazonka-datasync The response's http status code. amazonka-datasyncSee:   smart constructor. amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for Windows File Server location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) of the FSx for Windows File Server location to describe. amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for Windows File Server location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The time that the FSx for Windows File Server location was created. ,   - The name of the Windows domain that the FSx for Windows File Server belongs to. ,   - The Amazon Resource Name (ARN) of the FSx for Windows File Server location that was described. ,   - The URL of the FSx for Windows File Server location that was described. ,   - The Amazon Resource Names (ARNs) of the security groups that are configured for the FSx for Windows File Server file system. ,   - The user who has the permissions to access files and folders in the FSx for Windows File Server file system. ,  # - The response's http status code. amazonka-datasyncThe time that the FSx for Windows File Server location was created. amazonka-datasyncThe name of the Windows domain that the FSx for Windows File Server belongs to. amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for Windows File Server location that was described. amazonka-datasyncThe URL of the FSx for Windows File Server location that was described. amazonka-datasyncThe Amazon Resource Names (ARNs) of the security groups that are configured for the FSx for Windows File Server file system. amazonka-datasyncThe user who has the permissions to access files and folders in the FSx for Windows File Server file system. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync   F(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';: amazonka-datasyncSee:   smart constructor. amazonka-datasync7The time that the FSx for OpenZFS location was created. amazonka-datasync;The ARN of the FSx for OpenZFS location that was described. amazonka-datasyncThe uniform resource identifier (URI) of the FSx for OpenZFS location that was described. Example: 9fsxz://us-west-2.fs-1234567890abcdef02/fsx/folderA/folder amazonka-datasyncThe type of protocol that DataSync uses to access your file system. amazonka-datasyncThe ARNs of the security groups that are configured for the FSx for OpenZFS file system. amazonka-datasync The response's http status code. amazonka-datasyncSee:   smart constructor. amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for OpenZFS location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) of the FSx for OpenZFS location to describe. amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for OpenZFS location to describe. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  : - The time that the FSx for OpenZFS location was created. ,  > - The ARN of the FSx for OpenZFS location that was described. ,   - The uniform resource identifier (URI) of the FSx for OpenZFS location that was described. Example: 9fsxz://us-west-2.fs-1234567890abcdef02/fsx/folderA/folder ,   - The type of protocol that DataSync uses to access your file system. ,   - The ARNs of the security groups that are configured for the FSx for OpenZFS file system. ,  # - The response's http status code. amazonka-datasync7The time that the FSx for OpenZFS location was created. amazonka-datasync;The ARN of the FSx for OpenZFS location that was described. amazonka-datasyncThe uniform resource identifier (URI) of the FSx for OpenZFS location that was described. Example: 9fsxz://us-west-2.fs-1234567890abcdef02/fsx/folderA/folder amazonka-datasyncThe type of protocol that DataSync uses to access your file system. amazonka-datasyncThe ARNs of the security groups that are configured for the FSx for OpenZFS file system. amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync   G(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-datasyncSee:   smart constructor. amazonka-datasync'The time that the location was created. amazonka-datasync)The ARN of the FSx for ONTAP file system. amazonka-datasync2The ARN of the FSx for ONTAP file system location. amazonka-datasyncThe uniform resource identifier (URI) of the FSx for ONTAP file system location. amazonka-datasyncThe security groups that DataSync uses to access your FSx for ONTAP file system. amazonka-datasyncThe ARN of the storage virtual machine (SVM) on your FSx for ONTAP file system where you're copying data to or from. amazonka-datasync The response's http status code. amazonka-datasyncSee:   smart constructor. amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the FSx for ONTAP file system location that you want information about. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - Specifies the Amazon Resource Name (ARN) of the FSx for ONTAP file system location that you want information about. amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the FSx for ONTAP file system location that you want information about. amazonka-datasyncCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  * - The time that the location was created. , , - The ARN of the FSx for ONTAP file system. , 5 - The ARN of the FSx for ONTAP file system location. ,  - The uniform resource identifier (URI) of the FSx for ONTAP file system location. ,  - Undocumented member. ,  - The security groups that DataSync uses to access your FSx for ONTAP file system. ,  - The ARN of the storage virtual machine (SVM) on your FSx for ONTAP file system where you're copying data to or from. , # - The response's http status code. amazonka-datasync'The time that the location was created.amazonka-datasync)The ARN of the FSx for ONTAP file system.amazonka-datasync2The ARN of the FSx for ONTAP file system location.amazonka-datasyncThe uniform resource identifier (URI) of the FSx for ONTAP file system location.amazonka-datasyncUndocumented member.amazonka-datasyncThe security groups that DataSync uses to access your FSx for ONTAP file system.amazonka-datasyncThe ARN of the storage virtual machine (SVM) on your FSx for ONTAP file system where you're copying data to or from.amazonka-datasync The response's http status code. amazonka-datasync amazonka-datasync   H(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';,amazonka-datasyncSee:  smart constructor.amazonka-datasync6The time that the FSx for Lustre location was created.amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for Lustre location that was described.amazonka-datasync:The URI of the FSx for Lustre location that was described.amazonka-datasyncThe Amazon Resource Names (ARNs) of the security groups that are configured for the FSx for Lustre file system.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for Lustre location to describe.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the FSx for Lustre location to describe.amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for Lustre location to describe.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 9 - The time that the FSx for Lustre location was created.,  - The Amazon Resource Name (ARN) of the FSx for Lustre location that was described., = - The URI of the FSx for Lustre location that was described.,  - The Amazon Resource Names (ARNs) of the security groups that are configured for the FSx for Lustre file system., # - The response's http status code.amazonka-datasync6The time that the FSx for Lustre location was created.amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for Lustre location that was described.amazonka-datasync:The URI of the FSx for Lustre location that was described.amazonka-datasyncThe Amazon Resource Names (ARNs) of the security groups that are configured for the FSx for Lustre file system.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncI(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Žamazonka-datasyncDescribeLocationEfsResponseSee:  smart constructor.amazonka-datasyncThe ARN of the access point that DataSync uses to access the Amazon EFS file system.amazonka-datasync'The time that the location was created.amazonka-datasyncThe Identity and Access Management (IAM) role that DataSync assumes when mounting the Amazon EFS file system.amazonka-datasyncDescribes whether DataSync uses Transport Layer Security (TLS) encryption when copying data to or from the Amazon EFS file system.amazonka-datasync/The ARN of the Amazon EFS file system location.amazonka-datasync/The URL of the Amazon EFS file system location.amazonka-datasync The response's http status code.amazonka-datasyncDescribeLocationEfsRequestSee:  smart constructor.amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon EFS file system location that you want information about.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the Amazon EFS file system location that you want information about.amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon EFS file system location that you want information about.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The ARN of the access point that DataSync uses to access the Amazon EFS file system., * - The time that the location was created.,  - Undocumented member.,  - The Identity and Access Management (IAM) role that DataSync assumes when mounting the Amazon EFS file system.,  - Describes whether DataSync uses Transport Layer Security (TLS) encryption when copying data to or from the Amazon EFS file system., 2 - The ARN of the Amazon EFS file system location., 2 - The URL of the Amazon EFS file system location., # - The response's http status code.amazonka-datasyncThe ARN of the access point that DataSync uses to access the Amazon EFS file system.amazonka-datasync'The time that the location was created.amazonka-datasyncUndocumented member.amazonka-datasyncThe Identity and Access Management (IAM) role that DataSync assumes when mounting the Amazon EFS file system.amazonka-datasyncDescribes whether DataSync uses Transport Layer Security (TLS) encryption when copying data to or from the Amazon EFS file system.amazonka-datasync/The ARN of the Amazon EFS file system location.amazonka-datasync/The URL of the Amazon EFS file system location.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncJ(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';"amazonka-datasyncDescribeAgentResponseSee:  smart constructor.amazonka-datasync,The Amazon Resource Name (ARN) of the agent.amazonka-datasyncThe time that the agent was activated (that is, created in your account).amazonka-datasyncThe type of endpoint that your agent is connected to. If the endpoint is a VPC endpoint, the agent is not accessible over the public internet.amazonka-datasync3The time that the agent last connected to DataSync.amazonka-datasyncThe name of the agent.amazonka-datasyncThe subnet and the security group that DataSync used to access a VPC endpoint.amazonka-datasyncThe status of the agent. If the status is ONLINE, then the agent is configured properly and is available to use. The Running status is the normal running status for an agent. If the status is OFFLINE, the agent's VM is turned off or the agent is in an unhealthy state. When the issue that caused the unhealthy state is resolved, the agent returns to ONLINE status.amazonka-datasync The response's http status code.amazonka-datasync DescribeAgentSee:  smart constructor.amazonka-datasync8The Amazon Resource Name (ARN) of the agent to describe.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, ; - The Amazon Resource Name (ARN) of the agent to describe.amazonka-datasync8The Amazon Resource Name (ARN) of the agent to describe.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, / - The Amazon Resource Name (ARN) of the agent.,  - The time that the agent was activated (that is, created in your account).,  - The type of endpoint that your agent is connected to. If the endpoint is a VPC endpoint, the agent is not accessible over the public internet., 6 - The time that the agent last connected to DataSync.,  - The name of the agent.,  - The subnet and the security group that DataSync used to access a VPC endpoint.,  - The status of the agent. If the status is ONLINE, then the agent is configured properly and is available to use. The Running status is the normal running status for an agent. If the status is OFFLINE, the agent's VM is turned off or the agent is in an unhealthy state. When the issue that caused the unhealthy state is resolved, the agent returns to ONLINE status., # - The response's http status code.amazonka-datasync,The Amazon Resource Name (ARN) of the agent.amazonka-datasyncThe time that the agent was activated (that is, created in your account).amazonka-datasyncThe type of endpoint that your agent is connected to. If the endpoint is a VPC endpoint, the agent is not accessible over the public internet.amazonka-datasync3The time that the agent last connected to DataSync.amazonka-datasyncThe name of the agent.amazonka-datasyncThe subnet and the security group that DataSync used to access a VPC endpoint.amazonka-datasyncThe status of the agent. If the status is ONLINE, then the agent is configured properly and is available to use. The Running status is the normal running status for an agent. If the status is OFFLINE, the agent's VM is turned off or the agent is in an unhealthy state. When the issue that caused the unhealthy state is resolved, the agent returns to ONLINE status.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncK(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';ڱamazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasync DeleteTaskSee:  smart constructor.amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the task that you want to delete.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the Amazon Resource Name (ARN) of the task that you want to delete.amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the task that you want to delete.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasync  L(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncDeleteLocationSee:  smart constructor.amazonka-datasync9The Amazon Resource Name (ARN) of the location to delete.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, < - The Amazon Resource Name (ARN) of the location to delete.amazonka-datasync9The Amazon Resource Name (ARN) of the location to delete.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasync  M(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';:amazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncDeleteAgentRequestSee:  smart constructor.amazonka-datasyncThe Amazon Resource Name (ARN) of the agent to delete. Use the  ListAgents operation to return a list of agents for your account and Amazon Web Services Region.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the agent to delete. Use the  ListAgents operation to return a list of agents for your account and Amazon Web Services Region.amazonka-datasyncThe Amazon Resource Name (ARN) of the agent to delete. Use the  ListAgents operation to return a list of agents for your account and Amazon Web Services Region.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasync  N(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Hamazonka-datasyncCreateTaskResponseSee:  smart constructor.amazonka-datasync+The Amazon Resource Name (ARN) of the task.amazonka-datasync The response's http status code.amazonka-datasyncCreateTaskRequestSee:  smart constructor.amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon CloudWatch log group that is used to monitor and log events in the task.amazonka-datasyncSpecifies a list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.amazonka-datasyncSpecifies a list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.amazonka-datasyncThe name of a task. This value is a text reference that is used to identify the task in the console.amazonka-datasyncSpecifies the configuration options for a task. Some options include preserving file or object metadata and verifying data integrity.You can also override these options before starting an individual run of a task (also known as a task execution). For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/API_StartTaskExecution.htmlStartTaskExecution.amazonka-datasyncSpecifies a schedule used to periodically transfer files from a source to a destination location. The schedule should be specified in UTC time. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/task-scheduling.htmlScheduling your task.amazonka-datasyncSpecifies the tags that you want to apply to the Amazon Resource Name (ARN) representing the task.Tags are key-value pairs that help you manage, filter, and search for your DataSync resources.amazonka-datasyncThe Amazon Resource Name (ARN) of the source location for the task.amazonka-datasyncThe Amazon Resource Name (ARN) of an Amazon Web Services storage resource's location.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the Amazon CloudWatch log group that is used to monitor and log events in the task.,  - Specifies a list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.,  - Specifies a list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.,  - The name of a task. This value is a text reference that is used to identify the task in the console.,  - Specifies the configuration options for a task. Some options include preserving file or object metadata and verifying data integrity.You can also override these options before starting an individual run of a task (also known as a task execution). For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/API_StartTaskExecution.htmlStartTaskExecution.,  - Specifies a schedule used to periodically transfer files from a source to a destination location. The schedule should be specified in UTC time. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/task-scheduling.htmlScheduling your task.,  - Specifies the tags that you want to apply to the Amazon Resource Name (ARN) representing the task.Tags are key-value pairs that help you manage, filter, and search for your DataSync resources.,  - The Amazon Resource Name (ARN) of the source location for the task.,  - The Amazon Resource Name (ARN) of an Amazon Web Services storage resource's location.amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon CloudWatch log group that is used to monitor and log events in the task.amazonka-datasyncSpecifies a list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.amazonka-datasyncSpecifies a list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.amazonka-datasyncThe name of a task. This value is a text reference that is used to identify the task in the console.amazonka-datasyncSpecifies the configuration options for a task. Some options include preserving file or object metadata and verifying data integrity.You can also override these options before starting an individual run of a task (also known as a task execution). For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/API_StartTaskExecution.htmlStartTaskExecution.amazonka-datasyncSpecifies a schedule used to periodically transfer files from a source to a destination location. The schedule should be specified in UTC time. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/task-scheduling.htmlScheduling your task.amazonka-datasyncSpecifies the tags that you want to apply to the Amazon Resource Name (ARN) representing the task.Tags are key-value pairs that help you manage, filter, and search for your DataSync resources.amazonka-datasyncThe Amazon Resource Name (ARN) of the source location for the task.amazonka-datasyncThe Amazon Resource Name (ARN) of an Amazon Web Services storage resource's location.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, . - The Amazon Resource Name (ARN) of the task., # - The response's http status code.amazonka-datasync+The Amazon Resource Name (ARN) of the task.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncamazonka-datasyncO(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';'amazonka-datasyncCreateLocationSmbResponseSee:  smart constructor.amazonka-datasyncThe Amazon Resource Name (ARN) of the source SMB file system location that is created.amazonka-datasync The response's http status code.amazonka-datasyncCreateLocationSmbRequestSee:  smart constructor.amazonka-datasync>The name of the Windows domain that the SMB server belongs to.amazonka-datasyncThe name of the Windows domain that the SMB server belongs to.amazonka-datasyncIf you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.For information about NFS export configuration, see 18.7. The /etc/exports Configuration File in the Red Hat Enterprise Linux documentation.amazonka-datasyncThe name of the NFS server. This value is the IP address or Domain Name Service (DNS) name of the NFS server. An agent that is installed on-premises uses this hostname to mount the NFS server in a network.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.This name must either be DNS-compliant or must be an IP version 4 (IPv4) address.amazonka-datasyncContains a list of Amazon Resource Names (ARNs) of agents that are used to connect to an NFS server.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The NFS mount options that DataSync can use to mount your NFS share.,  - The key-value pair that represents the tag that you want to add to the location. The value can be an empty string. We recommend using tags to name your resources.,  - The subdirectory in the NFS file system that is used to read data from the NFS source location or write data to the NFS destination. The NFS path should be a path that's exported by the NFS server, or a subdirectory of that path. The path should be such that it can be mounted by other NFS clients in your network.8To see all the paths exported by your NFS server, run "showmount -e nfs-server-name" from an NFS client that has access to your server. You can specify any directory that appears in the results, and any subdirectory of that directory. Ensure that the NFS export is accessible without Kerberos authentication.To transfer all the data in the folder you specified, DataSync needs to have permissions to read all the data. To ensure this, either configure the NFS export with no_root_squash, or ensure that the permissions for all of the files that you want DataSync allow read access for all users. Doing either enables the agent to read the files. For the agent to access directories, you must additionally enable all execute access.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.For information about NFS export configuration, see 18.7. The /etc/exports Configuration File in the Red Hat Enterprise Linux documentation.,  - The name of the NFS server. This value is the IP address or Domain Name Service (DNS) name of the NFS server. An agent that is installed on-premises uses this hostname to mount the NFS server in a network.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.This name must either be DNS-compliant or must be an IP version 4 (IPv4) address.,  - Contains a list of Amazon Resource Names (ARNs) of agents that are used to connect to an NFS server.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.amazonka-datasyncThe NFS mount options that DataSync can use to mount your NFS share.amazonka-datasyncThe key-value pair that represents the tag that you want to add to the location. The value can be an empty string. We recommend using tags to name your resources.amazonka-datasyncThe subdirectory in the NFS file system that is used to read data from the NFS source location or write data to the NFS destination. The NFS path should be a path that's exported by the NFS server, or a subdirectory of that path. The path should be such that it can be mounted by other NFS clients in your network.8To see all the paths exported by your NFS server, run "showmount -e nfs-server-name" from an NFS client that has access to your server. You can specify any directory that appears in the results, and any subdirectory of that directory. Ensure that the NFS export is accessible without Kerberos authentication.To transfer all the data in the folder you specified, DataSync needs to have permissions to read all the data. To ensure this, either configure the NFS export with no_root_squash, or ensure that the permissions for all of the files that you want DataSync allow read access for all users. Doing either enables the agent to read the files. For the agent to access directories, you must additionally enable all execute access.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.For information about NFS export configuration, see 18.7. The /etc/exports Configuration File in the Red Hat Enterprise Linux documentation.amazonka-datasyncThe name of the NFS server. This value is the IP address or Domain Name Service (DNS) name of the NFS server. An agent that is installed on-premises uses this hostname to mount the NFS server in a network.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.This name must either be DNS-compliant or must be an IP version 4 (IPv4) address.amazonka-datasyncContains a list of Amazon Resource Names (ARNs) of agents that are used to connect to an NFS server.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the source NFS file system location that is created., # - The response's http status code.amazonka-datasyncThe Amazon Resource Name (ARN) of the source NFS file system location that is created.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncamazonka-datasyncamazonka-datasyncS(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';0"amazonka-datasyncSee:  smart constructor.amazonka-datasync;The ARN of the source HDFS cluster location that's created.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasyncThe size of data blocks to write into the HDFS cluster. The block size must be a multiple of 512 bytes. The default block size is 128 mebibytes (MiB).amazonka-datasyncThe Kerberos key table (keytab) that contains mappings between the defined Kerberos principal and the encrypted keys. You can load the keytab from a file by providing the file's address. If you're using the CLI, it performs base64 encoding for you. Otherwise, provide the base64-encoded text.If KERBEROS is specified for AuthenticationType, this parameter is required.amazonka-datasyncThe  krb5.conf file that contains the Kerberos configuration information. You can load the  krb5.conf file by providing the file's address. If you're using the CLI, it performs the base64 encoding for you. Otherwise, provide the base64-encoded text.If KERBEROS is specified for AuthenticationType, this parameter is required.amazonka-datasyncThe Kerberos principal with access to the files and folders on the HDFS cluster.If KERBEROS is specified for AuthenticationType, this parameter is required.amazonka-datasync:The URI of the HDFS cluster's Key Management Server (KMS).amazonka-datasyncThe Quality of Protection (QOP) configuration specifies the Remote Procedure Call (RPC) and data transfer protection settings configured on the Hadoop Distributed File System (HDFS) cluster. If QopConfiguration isn't specified,  RpcProtection and DataTransferProtection default to PRIVACY . If you set  RpcProtection or DataTransferProtection., the other parameter assumes the same value.amazonka-datasyncThe number of DataNodes to replicate the data to when writing to the HDFS cluster. By default, data is replicated to three DataNodes.amazonka-datasyncThe user name used to identify the client on the host operating system.If SIMPLE is specified for AuthenticationType, this parameter is required.amazonka-datasyncA subdirectory in the HDFS cluster. This subdirectory is used to read data from or write data to the HDFS cluster. If the subdirectory isn't specified, it will default to /.amazonka-datasyncThe key-value pair that represents the tag that you want to add to the location. The value can be an empty string. We recommend using tags to name your resources.amazonka-datasyncThe NameNode that manages the HDFS namespace. The NameNode performs operations such as opening, closing, and renaming files and directories. The NameNode contains the information to map blocks of data to the DataNodes. You can use only one NameNode.amazonka-datasyncThe type of authentication used to determine the identity of the user.amazonka-datasyncThe Amazon Resource Names (ARNs) of the agents that are used to connect to the HDFS cluster.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The size of data blocks to write into the HDFS cluster. The block size must be a multiple of 512 bytes. The default block size is 128 mebibytes (MiB).,  - The Kerberos key table (keytab) that contains mappings between the defined Kerberos principal and the encrypted keys. You can load the keytab from a file by providing the file's address. If you're using the CLI, it performs base64 encoding for you. Otherwise, provide the base64-encoded text.If KERBEROS is specified for AuthenticationType%, this parameter is required.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.,  - The  krb5.conf file that contains the Kerberos configuration information. You can load the  krb5.conf file by providing the file's address. If you're using the CLI, it performs the base64 encoding for you. Otherwise, provide the base64-encoded text.If KERBEROS is specified for AuthenticationType%, this parameter is required.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.,  - The Kerberos principal with access to the files and folders on the HDFS cluster.If KERBEROS is specified for AuthenticationType, this parameter is required., = - The URI of the HDFS cluster's Key Management Server (KMS).,  - The Quality of Protection (QOP) configuration specifies the Remote Procedure Call (RPC) and data transfer protection settings configured on the Hadoop Distributed File System (HDFS) cluster. If QopConfiguration isn't specified,  RpcProtection and DataTransferProtection default to PRIVACY . If you set  RpcProtection or DataTransferProtection., the other parameter assumes the same value.,  - The number of DataNodes to replicate the data to when writing to the HDFS cluster. By default, data is replicated to three DataNodes.,  - The user name used to identify the client on the host operating system.If SIMPLE is specified for AuthenticationType, this parameter is required.,  - A subdirectory in the HDFS cluster. This subdirectory is used to read data from or write data to the HDFS cluster. If the subdirectory isn't specified, it will default to /.,  - The key-value pair that represents the tag that you want to add to the location. The value can be an empty string. We recommend using tags to name your resources.,  - The NameNode that manages the HDFS namespace. The NameNode performs operations such as opening, closing, and renaming files and directories. The NameNode contains the information to map blocks of data to the DataNodes. You can use only one NameNode.,  - The type of authentication used to determine the identity of the user.,  - The Amazon Resource Names (ARNs) of the agents that are used to connect to the HDFS cluster.amazonka-datasyncThe size of data blocks to write into the HDFS cluster. The block size must be a multiple of 512 bytes. The default block size is 128 mebibytes (MiB).amazonka-datasyncThe Kerberos key table (keytab) that contains mappings between the defined Kerberos principal and the encrypted keys. You can load the keytab from a file by providing the file's address. If you're using the CLI, it performs base64 encoding for you. Otherwise, provide the base64-encoded text.If KERBEROS is specified for AuthenticationType%, this parameter is required.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.amazonka-datasyncThe  krb5.conf file that contains the Kerberos configuration information. You can load the  krb5.conf file by providing the file's address. If you're using the CLI, it performs the base64 encoding for you. Otherwise, provide the base64-encoded text.If KERBEROS is specified for AuthenticationType%, this parameter is required.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.amazonka-datasyncThe Kerberos principal with access to the files and folders on the HDFS cluster.If KERBEROS is specified for AuthenticationType, this parameter is required.amazonka-datasync:The URI of the HDFS cluster's Key Management Server (KMS).amazonka-datasyncThe Quality of Protection (QOP) configuration specifies the Remote Procedure Call (RPC) and data transfer protection settings configured on the Hadoop Distributed File System (HDFS) cluster. If QopConfiguration isn't specified,  RpcProtection and DataTransferProtection default to PRIVACY . If you set  RpcProtection or DataTransferProtection., the other parameter assumes the same value.amazonka-datasyncThe number of DataNodes to replicate the data to when writing to the HDFS cluster. By default, data is replicated to three DataNodes.amazonka-datasyncThe user name used to identify the client on the host operating system.If SIMPLE is specified for AuthenticationType, this parameter is required.amazonka-datasyncA subdirectory in the HDFS cluster. This subdirectory is used to read data from or write data to the HDFS cluster. If the subdirectory isn't specified, it will default to /.amazonka-datasyncThe key-value pair that represents the tag that you want to add to the location. The value can be an empty string. We recommend using tags to name your resources.amazonka-datasyncThe NameNode that manages the HDFS namespace. The NameNode performs operations such as opening, closing, and renaming files and directories. The NameNode contains the information to map blocks of data to the DataNodes. You can use only one NameNode.amazonka-datasyncThe type of authentication used to determine the identity of the user.amazonka-datasyncThe Amazon Resource Names (ARNs) of the agents that are used to connect to the HDFS cluster.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, > - The ARN of the source HDFS cluster location that's created., # - The response's http status code.amazonka-datasync;The ARN of the source HDFS cluster location that's created.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncamazonka-datasyncamazonka-datasync$$T(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Bamazonka-datasyncSee:  smart constructor.amazonka-datasyncThe ARN of the FSx for Windows File Server file system location you created.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasyncSpecifies the name of the Windows domain that the FSx for Windows File Server belongs to.amazonka-datasyncSpecifies a mount path for your file system using forward slashes. This is where DataSync reads or writes data (depending on if this is a source or destination location).amazonka-datasyncSpecifies labels that help you categorize, filter, and search for your Amazon Web Services resources. We recommend creating at least a name tag for your location.amazonka-datasyncSpecifies the Amazon Resource Name (ARN) for the FSx for Windows File Server file system.amazonka-datasyncSpecifies the ARNs of the security groups that provide access to your file system's preferred subnet.If you choose a security group that doesn't allow connections from within itself, do one of the following:Configure the security group to allow it to communicate within itself.Choose a different security group that can communicate with the mount target's security group.amazonka-datasyncSpecifies the user who has the permissions to access files and folders in the file system.For information about choosing a user name that ensures sufficient permissions to files, folders, and metadata, see  !create-fsx-location.html#FSxWuseruser.amazonka-datasyncSpecifies the password of the user who has the permissions to access files and folders in the file system.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the name of the Windows domain that the FSx for Windows File Server belongs to.,  - Specifies a mount path for your file system using forward slashes. This is where DataSync reads or writes data (depending on if this is a source or destination location).,  - Specifies labels that help you categorize, filter, and search for your Amazon Web Services resources. We recommend creating at least a name tag for your location.,  - Specifies the Amazon Resource Name (ARN) for the FSx for Windows File Server file system.,  - Specifies the ARNs of the security groups that provide access to your file system's preferred subnet.If you choose a security group that doesn't allow connections from within itself, do one of the following:Configure the security group to allow it to communicate within itself.Choose a different security group that can communicate with the mount target's security group.,  - Specifies the user who has the permissions to access files and folders in the file system.For information about choosing a user name that ensures sufficient permissions to files, folders, and metadata, see  !create-fsx-location.html#FSxWuseruser.,  - Specifies the password of the user who has the permissions to access files and folders in the file system.amazonka-datasyncSpecifies the name of the Windows domain that the FSx for Windows File Server belongs to.amazonka-datasyncSpecifies a mount path for your file system using forward slashes. This is where DataSync reads or writes data (depending on if this is a source or destination location).amazonka-datasyncSpecifies labels that help you categorize, filter, and search for your Amazon Web Services resources. We recommend creating at least a name tag for your location.amazonka-datasyncSpecifies the Amazon Resource Name (ARN) for the FSx for Windows File Server file system.amazonka-datasyncSpecifies the ARNs of the security groups that provide access to your file system's preferred subnet.If you choose a security group that doesn't allow connections from within itself, do one of the following:Configure the security group to allow it to communicate within itself.Choose a different security group that can communicate with the mount target's security group.amazonka-datasyncSpecifies the user who has the permissions to access files and folders in the file system.For information about choosing a user name that ensures sufficient permissions to files, folders, and metadata, see  !create-fsx-location.html#FSxWuseruser.amazonka-datasyncSpecifies the password of the user who has the permissions to access files and folders in the file system.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The ARN of the FSx for Windows File Server file system location you created., # - The response's http status code.amazonka-datasyncThe ARN of the FSx for Windows File Server file system location you created.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncamazonka-datasyncamazonka-datasyncamazonka-datasyncU(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';-amazonka-datasyncSee:  smart constructor.amazonka-datasyncThe ARN of the FSx for OpenZFS file system location that you created.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasync;A subdirectory in the location's path that must begin with /fsx. DataSync uses this subdirectory to read or write data (depending on whether the file system is a source or destination location).amazonka-datasyncThe key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location.amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for OpenZFS file system.amazonka-datasyncThe type of protocol that DataSync uses to access your file system.amazonka-datasyncThe ARNs of the security groups that are used to configure the FSx for OpenZFS file system.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, > - A subdirectory in the location's path that must begin with /fsx. DataSync uses this subdirectory to read or write data (depending on whether the file system is a source or destination location).,  - The key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location.,  - The Amazon Resource Name (ARN) of the FSx for OpenZFS file system.,  - The type of protocol that DataSync uses to access your file system.,  - The ARNs of the security groups that are used to configure the FSx for OpenZFS file system.amazonka-datasync;A subdirectory in the location's path that must begin with /fsx. DataSync uses this subdirectory to read or write data (depending on whether the file system is a source or destination location).amazonka-datasyncThe key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location.amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for OpenZFS file system.amazonka-datasyncThe type of protocol that DataSync uses to access your file system.amazonka-datasyncThe ARNs of the security groups that are used to configure the FSx for OpenZFS file system.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The ARN of the FSx for OpenZFS file system location that you created., # - The response's http status code.amazonka-datasyncThe ARN of the FSx for OpenZFS file system location that you created.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncamazonka-datasyncamazonka-datasyncV(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Eamazonka-datasyncSee:  smart constructor.amazonka-datasyncSpecifies the ARN of the FSx for ONTAP file system location that you create.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasyncSpecifies the junction path (also known as a mount point) in the SVM volume where you're copying data to or from (for example, /vol1).Don't specify a junction path in the SVM's root volume. For more information, see  https://docs.aws.amazon.com/fsx/latest/ONTAPGuide/managing-svms.html/Managing FSx for ONTAP storage virtual machines in the &Amazon FSx for NetApp ONTAP User Guide.amazonka-datasyncSpecifies labels that help you categorize, filter, and search for your Amazon Web Services resources. We recommend creating at least a name tag for your location.amazonka-datasyncSpecifies the Amazon EC2 security groups that provide access to your file system's preferred subnet.The security groups must allow outbound traffic on the following ports (depending on the protocol you use):Network File System (NFS): TCP ports 111, 635, and 2049Server Message Block (SMB): TCP port 445Your file system's security groups must also allow inbound traffic on the same ports.amazonka-datasyncSpecifies the ARN of the storage virtual machine (SVM) on your file system where you're copying data to or from.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the junction path (also known as a mount point) in the SVM volume where you're copying data to or from (for example, /vol1).Don't specify a junction path in the SVM's root volume. For more information, see  https://docs.aws.amazon.com/fsx/latest/ONTAPGuide/managing-svms.html/Managing FSx for ONTAP storage virtual machines in the &Amazon FSx for NetApp ONTAP User Guide.,  - Specifies labels that help you categorize, filter, and search for your Amazon Web Services resources. We recommend creating at least a name tag for your location.,  - Undocumented member.,  - Specifies the Amazon EC2 security groups that provide access to your file system's preferred subnet.The security groups must allow outbound traffic on the following ports (depending on the protocol you use):Network File System (NFS): TCP ports 111, 635, and 2049Server Message Block (SMB): TCP port 445Your file system's security groups must also allow inbound traffic on the same ports.,  - Specifies the ARN of the storage virtual machine (SVM) on your file system where you're copying data to or from.amazonka-datasyncSpecifies the junction path (also known as a mount point) in the SVM volume where you're copying data to or from (for example, /vol1).Don't specify a junction path in the SVM's root volume. For more information, see  https://docs.aws.amazon.com/fsx/latest/ONTAPGuide/managing-svms.html/Managing FSx for ONTAP storage virtual machines in the &Amazon FSx for NetApp ONTAP User Guide.amazonka-datasyncSpecifies labels that help you categorize, filter, and search for your Amazon Web Services resources. We recommend creating at least a name tag for your location.amazonka-datasyncUndocumented member.amazonka-datasyncSpecifies the Amazon EC2 security groups that provide access to your file system's preferred subnet.The security groups must allow outbound traffic on the following ports (depending on the protocol you use):Network File System (NFS): TCP ports 111, 635, and 2049Server Message Block (SMB): TCP port 445Your file system's security groups must also allow inbound traffic on the same ports.amazonka-datasyncSpecifies the ARN of the storage virtual machine (SVM) on your file system where you're copying data to or from.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the ARN of the FSx for ONTAP file system location that you create., # - The response's http status code.amazonka-datasyncSpecifies the ARN of the FSx for ONTAP file system location that you create.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncamazonka-datasyncamazonka-datasyncW(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';samazonka-datasyncSee:  smart constructor.amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for Lustre file system location that's created.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasyncA subdirectory in the location's path. This subdirectory in the FSx for Lustre file system is used to read data from the FSx for Lustre source location or write data to the FSx for Lustre destination.amazonka-datasyncThe key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location.amazonka-datasyncThe Amazon Resource Name (ARN) for the FSx for Lustre file system.amazonka-datasyncThe Amazon Resource Names (ARNs) of the security groups that are used to configure the FSx for Lustre file system.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A subdirectory in the location's path. This subdirectory in the FSx for Lustre file system is used to read data from the FSx for Lustre source location or write data to the FSx for Lustre destination.,  - The key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location.,  - The Amazon Resource Name (ARN) for the FSx for Lustre file system.,  - The Amazon Resource Names (ARNs) of the security groups that are used to configure the FSx for Lustre file system.amazonka-datasyncA subdirectory in the location's path. This subdirectory in the FSx for Lustre file system is used to read data from the FSx for Lustre source location or write data to the FSx for Lustre destination.amazonka-datasyncThe key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location.amazonka-datasyncThe Amazon Resource Name (ARN) for the FSx for Lustre file system.amazonka-datasyncThe Amazon Resource Names (ARNs) of the security groups that are used to configure the FSx for Lustre file system.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the FSx for Lustre file system location that's created., # - The response's http status code.amazonka-datasyncThe Amazon Resource Name (ARN) of the FSx for Lustre file system location that's created.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncamazonka-datasyncX(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Ramazonka-datasyncCreateLocationEfsSee:  smart constructor.amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon EFS file system location that you create.amazonka-datasync The response's http status code.amazonka-datasyncCreateLocationEfsRequestSee:  smart constructor.amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the access point that DataSync uses to access the Amazon EFS file system.amazonka-datasyncSpecifies an Identity and Access Management (IAM) role that DataSync assumes when mounting the Amazon EFS file system.amazonka-datasyncSpecifies whether you want DataSync to use Transport Layer Security (TLS) 1.2 encryption when it copies data to or from the Amazon EFS file system.%If you specify an access point using AccessPointArn or an IAM role using FileSystemAccessRoleArn", you must set this parameter to TLS1_2.amazonka-datasyncSpecifies a mount path for your Amazon EFS file system. This is where DataSync reads or writes data (depending on if this is a source or destination location). By default, DataSync uses the root directory, but you can also include subdirectories.=You must specify a value with forward slashes (for example, /path/to/folder).amazonka-datasyncSpecifies the key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location.amazonka-datasync1Specifies the ARN for the Amazon EFS file system.amazonka-datasyncSpecifies the subnet and security groups DataSync uses to access your Amazon EFS file system.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the Amazon Resource Name (ARN) of the access point that DataSync uses to access the Amazon EFS file system.,  - Specifies an Identity and Access Management (IAM) role that DataSync assumes when mounting the Amazon EFS file system.,  - Specifies whether you want DataSync to use Transport Layer Security (TLS) 1.2 encryption when it copies data to or from the Amazon EFS file system.%If you specify an access point using AccessPointArn or an IAM role using FileSystemAccessRoleArn", you must set this parameter to TLS1_2.,  - Specifies a mount path for your Amazon EFS file system. This is where DataSync reads or writes data (depending on if this is a source or destination location). By default, DataSync uses the root directory, but you can also include subdirectories.=You must specify a value with forward slashes (for example, /path/to/folder).,  - Specifies the key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location., 4 - Specifies the ARN for the Amazon EFS file system.,  - Specifies the subnet and security groups DataSync uses to access your Amazon EFS file system.amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the access point that DataSync uses to access the Amazon EFS file system.amazonka-datasyncSpecifies an Identity and Access Management (IAM) role that DataSync assumes when mounting the Amazon EFS file system.amazonka-datasyncSpecifies whether you want DataSync to use Transport Layer Security (TLS) 1.2 encryption when it copies data to or from the Amazon EFS file system.%If you specify an access point using AccessPointArn or an IAM role using FileSystemAccessRoleArn", you must set this parameter to TLS1_2.amazonka-datasyncSpecifies a mount path for your Amazon EFS file system. This is where DataSync reads or writes data (depending on if this is a source or destination location). By default, DataSync uses the root directory, but you can also include subdirectories.=You must specify a value with forward slashes (for example, /path/to/folder).amazonka-datasyncSpecifies the key-value pair that represents a tag that you want to add to the resource. The value can be an empty string. This value helps you manage, filter, and search for your resources. We recommend that you create a name tag for your location.amazonka-datasync1Specifies the ARN for the Amazon EFS file system.amazonka-datasyncSpecifies the subnet and security groups DataSync uses to access your Amazon EFS file system.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the Amazon EFS file system location that you create., # - The response's http status code.amazonka-datasyncThe Amazon Resource Name (ARN) of the Amazon EFS file system location that you create.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncamazonka-datasyncY(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';=amazonka-datasyncCreateAgentResponseSee:  smart constructor.amazonka-datasync5The Amazon Resource Name (ARN) of the agent. Use the  ListAgents operation to return a list of agents for your account and Amazon Web Services Region.amazonka-datasync The response's http status code.amazonka-datasyncCreateAgentRequestSee:  smart constructor.amazonka-datasyncThe name you configured for your agent. This value is a text reference that is used to identify the agent in the console.amazonka-datasyncThe ARNs of the security groups used to protect your data transfer task subnets. See  https://docs.aws.amazon.com/datasync/latest/userguide/API_Ec2Config.html#DataSync-Type-Ec2Config-SecurityGroupArnsSecurityGroupArns.amazonka-datasyncThe Amazon Resource Names (ARNs) of the subnets in which DataSync will create elastic network interfaces for each data transfer task. The agent that runs a task must be private. When you start a task that is associated with an agent created in a VPC, or one that has access to an IP address in a VPC, then the task is also private. In this case, DataSync creates four network interfaces for each task in your subnet. For a data transfer to work, the agent must be able to route to all these four network interfaces.amazonka-datasyncThe key-value pair that represents the tag that you want to associate with the agent. The value can be an empty string. This value helps you manage, filter, and search for your agents.Valid characters for key and value are letters, spaces, and numbers representable in UTF-8 format, and the following special characters: + - = . _ : / @.amazonka-datasyncThe ID of the VPC (virtual private cloud) endpoint that the agent has access to. This is the client-side VPC endpoint, also called a PrivateLink. If you don't have a PrivateLink VPC endpoint, see  https://docs.aws.amazon.com/vpc/latest/userguide/endpoint-service.html#create-endpoint-service-Creating a VPC Endpoint Service Configuration in the Amazon VPC User Guide.!VPC endpoint ID looks like this: vpce-01234d5aff67890e1.amazonka-datasyncYour agent activation key. You can get the activation key either by sending an HTTP GET request with redirects that enable you to get the agent IP address (port 80). Alternatively, you can get it from the DataSync console.The redirect URL returned in the response provides you the activation key for your agent in the query string parameter  activationKey. It might also include other activation-related parameters; however, these are merely defaults. The arguments you pass to this API call determine the actual configuration of your agent.For more information, see Activating an Agent in the /DataSync User Guide./amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The name you configured for your agent. This value is a text reference that is used to identify the agent in the console.,  - The ARNs of the security groups used to protect your data transfer task subnets. See  https://docs.aws.amazon.com/datasync/latest/userguide/API_Ec2Config.html#DataSync-Type-Ec2Config-SecurityGroupArnsSecurityGroupArns.,  - The Amazon Resource Names (ARNs) of the subnets in which DataSync will create elastic network interfaces for each data transfer task. The agent that runs a task must be private. When you start a task that is associated with an agent created in a VPC, or one that has access to an IP address in a VPC, then the task is also private. In this case, DataSync creates four network interfaces for each task in your subnet. For a data transfer to work, the agent must be able to route to all these four network interfaces.,  - The key-value pair that represents the tag that you want to associate with the agent. The value can be an empty string. This value helps you manage, filter, and search for your agents.Valid characters for key and value are letters, spaces, and numbers representable in UTF-8 format, and the following special characters: + - = . _ : / @.,  - The ID of the VPC (virtual private cloud) endpoint that the agent has access to. This is the client-side VPC endpoint, also called a PrivateLink. If you don't have a PrivateLink VPC endpoint, see  https://docs.aws.amazon.com/vpc/latest/userguide/endpoint-service.html#create-endpoint-service-Creating a VPC Endpoint Service Configuration in the Amazon VPC User Guide.!VPC endpoint ID looks like this: vpce-01234d5aff67890e1.,  - Your agent activation key. You can get the activation key either by sending an HTTP GET request with redirects that enable you to get the agent IP address (port 80). Alternatively, you can get it from the DataSync console.The redirect URL returned in the response provides you the activation key for your agent in the query string parameter  activationKey. It might also include other activation-related parameters; however, these are merely defaults. The arguments you pass to this API call determine the actual configuration of your agent.For more information, see Activating an Agent in the /DataSync User Guide./amazonka-datasyncThe name you configured for your agent. This value is a text reference that is used to identify the agent in the console.amazonka-datasyncThe ARNs of the security groups used to protect your data transfer task subnets. See  https://docs.aws.amazon.com/datasync/latest/userguide/API_Ec2Config.html#DataSync-Type-Ec2Config-SecurityGroupArnsSecurityGroupArns.amazonka-datasyncThe Amazon Resource Names (ARNs) of the subnets in which DataSync will create elastic network interfaces for each data transfer task. The agent that runs a task must be private. When you start a task that is associated with an agent created in a VPC, or one that has access to an IP address in a VPC, then the task is also private. In this case, DataSync creates four network interfaces for each task in your subnet. For a data transfer to work, the agent must be able to route to all these four network interfaces.amazonka-datasyncThe key-value pair that represents the tag that you want to associate with the agent. The value can be an empty string. This value helps you manage, filter, and search for your agents.Valid characters for key and value are letters, spaces, and numbers representable in UTF-8 format, and the following special characters: + - = . _ : / @.amazonka-datasyncThe ID of the VPC (virtual private cloud) endpoint that the agent has access to. This is the client-side VPC endpoint, also called a PrivateLink. If you don't have a PrivateLink VPC endpoint, see  https://docs.aws.amazon.com/vpc/latest/userguide/endpoint-service.html#create-endpoint-service-Creating a VPC Endpoint Service Configuration in the Amazon VPC User Guide.!VPC endpoint ID looks like this: vpce-01234d5aff67890e1.amazonka-datasyncYour agent activation key. You can get the activation key either by sending an HTTP GET request with redirects that enable you to get the agent IP address (port 80). Alternatively, you can get it from the DataSync console.The redirect URL returned in the response provides you the activation key for your agent in the query string parameter  activationKey. It might also include other activation-related parameters; however, these are merely defaults. The arguments you pass to this API call determine the actual configuration of your agent.For more information, see Activating an Agent in the /DataSync User Guide./amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 8 - The Amazon Resource Name (ARN) of the agent. Use the  ListAgents operation to return a list of agents for your account and Amazon Web Services Region., # - The response's http status code.amazonka-datasync5The Amazon Resource Name (ARN) of the agent. Use the  ListAgents operation to return a list of agents for your account and Amazon Web Services Region.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncZ(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Damazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncCancelTaskExecutionRequestSee:  smart constructor.amazonka-datasync=The Amazon Resource Name (ARN) of the task execution to stop.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the task execution to stop.amazonka-datasync=The Amazon Resource Name (ARN) of the task execution to stop.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasync  [(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';K amazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncUntagResourceRequestSee:  smart constructor.amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the resource to remove the tags from.amazonka-datasync7Specifies the keys in the tags that you want to remove.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the Amazon Resource Name (ARN) of the resource to remove the tags from., : - Specifies the keys in the tags that you want to remove.amazonka-datasyncSpecifies the Amazon Resource Name (ARN) of the resource to remove the tags from.amazonka-datasync7Specifies the keys in the tags that you want to remove.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncamazonka-datasync  \(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';R amazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncUpdateAgentRequestSee:  smart constructor.amazonka-datasync5The name that you want to use to configure the agent.amazonka-datasync6The Amazon Resource Name (ARN) of the agent to update.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 8 - The name that you want to use to configure the agent., 9 - The Amazon Resource Name (ARN) of the agent to update.amazonka-datasync5The name that you want to use to configure the agent.amazonka-datasync6The Amazon Resource Name (ARN) of the agent to update.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasync  ](c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';s amazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasyncThe ARNs of the agents that are used to connect to the HDFS cluster.amazonka-datasyncThe type of authentication used to determine the identity of the user.amazonka-datasync;The size of the data blocks to write into the HDFS cluster.amazonka-datasyncThe Kerberos key table (keytab) that contains mappings between the defined Kerberos principal and the encrypted keys. You can load the keytab from a file by providing the file's address. If you use the CLI, it performs base64 encoding for you. Otherwise, provide the base64-encoded text.amazonka-datasyncThe  krb5.conf file that contains the Kerberos configuration information. You can load the  krb5.conf file by providing the file's address. If you're using the CLI, it performs the base64 encoding for you. Otherwise, provide the base64-encoded text.amazonka-datasyncThe Kerberos principal with access to the files and folders on the HDFS cluster.amazonka-datasync:The URI of the HDFS cluster's Key Management Server (KMS).amazonka-datasyncThe NameNode that manages the HDFS namespace. The NameNode performs operations such as opening, closing, and renaming files and directories. The NameNode contains the information to map blocks of data to the DataNodes. You can use only one NameNode.amazonka-datasyncThe Quality of Protection (QOP) configuration specifies the Remote Procedure Call (RPC) and data transfer privacy settings configured on the Hadoop Distributed File System (HDFS) cluster.amazonka-datasyncThe number of DataNodes to replicate the data to when writing to the HDFS cluster.amazonka-datasyncThe user name used to identify the client on the host operating system.amazonka-datasyncA subdirectory in the HDFS cluster. This subdirectory is used to read data from or write data to the HDFS cluster.amazonka-datasyncThe Amazon Resource Name (ARN) of the source HDFS cluster location.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The ARNs of the agents that are used to connect to the HDFS cluster.,  - The type of authentication used to determine the identity of the user., > - The size of the data blocks to write into the HDFS cluster.,  - The Kerberos key table (keytab) that contains mappings between the defined Kerberos principal and the encrypted keys. You can load the keytab from a file by providing the file's address. If you use the CLI, it performs base64 encoding for you. Otherwise, provide the base64-encoded text.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.,  - The  krb5.conf file that contains the Kerberos configuration information. You can load the  krb5.conf file by providing the file's address. If you're using the CLI, it performs the base64 encoding for you. Otherwise, provide the base64-encoded text.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.,  - The Kerberos principal with access to the files and folders on the HDFS cluster., = - The URI of the HDFS cluster's Key Management Server (KMS).,  - The NameNode that manages the HDFS namespace. The NameNode performs operations such as opening, closing, and renaming files and directories. The NameNode contains the information to map blocks of data to the DataNodes. You can use only one NameNode.,  - The Quality of Protection (QOP) configuration specifies the Remote Procedure Call (RPC) and data transfer privacy settings configured on the Hadoop Distributed File System (HDFS) cluster.,  - The number of DataNodes to replicate the data to when writing to the HDFS cluster.,  - The user name used to identify the client on the host operating system.,  - A subdirectory in the HDFS cluster. This subdirectory is used to read data from or write data to the HDFS cluster.,  - The Amazon Resource Name (ARN) of the source HDFS cluster location.amazonka-datasyncThe ARNs of the agents that are used to connect to the HDFS cluster.amazonka-datasyncThe type of authentication used to determine the identity of the user.amazonka-datasync;The size of the data blocks to write into the HDFS cluster.amazonka-datasyncThe Kerberos key table (keytab) that contains mappings between the defined Kerberos principal and the encrypted keys. You can load the keytab from a file by providing the file's address. If you use the CLI, it performs base64 encoding for you. Otherwise, provide the base64-encoded text.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.amazonka-datasyncThe  krb5.conf file that contains the Kerberos configuration information. You can load the  krb5.conf file by providing the file's address. If you're using the CLI, it performs the base64 encoding for you. Otherwise, provide the base64-encoded text.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.amazonka-datasyncThe Kerberos principal with access to the files and folders on the HDFS cluster.amazonka-datasync:The URI of the HDFS cluster's Key Management Server (KMS).amazonka-datasyncThe NameNode that manages the HDFS namespace. The NameNode performs operations such as opening, closing, and renaming files and directories. The NameNode contains the information to map blocks of data to the DataNodes. You can use only one NameNode.amazonka-datasyncThe Quality of Protection (QOP) configuration specifies the Remote Procedure Call (RPC) and data transfer privacy settings configured on the Hadoop Distributed File System (HDFS) cluster.amazonka-datasyncThe number of DataNodes to replicate the data to when writing to the HDFS cluster.amazonka-datasyncThe user name used to identify the client on the host operating system.amazonka-datasyncA subdirectory in the HDFS cluster. This subdirectory is used to read data from or write data to the HDFS cluster.amazonka-datasyncThe Amazon Resource Name (ARN) of the source HDFS cluster location.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasync""^(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';D amazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasyncThe subdirectory in the NFS file system that is used to read data from the NFS source location or write data to the NFS destination. The NFS path should be a path that's exported by the NFS server, or a subdirectory of that path. The path should be such that it can be mounted by other NFS clients in your network.8To see all the paths exported by your NFS server, run "showmount -e nfs-server-name" from an NFS client that has access to your server. You can specify any directory that appears in the results, and any subdirectory of that directory. Ensure that the NFS export is accessible without Kerberos authentication.To transfer all the data in the folder that you specified, DataSync must have permissions to read all the data. To ensure this, either configure the NFS export with no_root_squash, or ensure that the files you want DataSync to access have permissions that allow read access for all users. Doing either option enables the agent to read the files. For the agent to access directories, you must additionally enable all execute access.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.For information about NFS export configuration, see 18.7. The /etc/exports Configuration File in the Red Hat Enterprise Linux documentation.amazonka-datasync=The Amazon Resource Name (ARN) of the NFS location to update.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Undocumented member.,  - Undocumented member.,  - The subdirectory in the NFS file system that is used to read data from the NFS source location or write data to the NFS destination. The NFS path should be a path that's exported by the NFS server, or a subdirectory of that path. The path should be such that it can be mounted by other NFS clients in your network.8To see all the paths exported by your NFS server, run "showmount -e nfs-server-name" from an NFS client that has access to your server. You can specify any directory that appears in the results, and any subdirectory of that directory. Ensure that the NFS export is accessible without Kerberos authentication.To transfer all the data in the folder that you specified, DataSync must have permissions to read all the data. To ensure this, either configure the NFS export with no_root_squash, or ensure that the files you want DataSync to access have permissions that allow read access for all users. Doing either option enables the agent to read the files. For the agent to access directories, you must additionally enable all execute access.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.For information about NFS export configuration, see 18.7. The /etc/exports Configuration File in the Red Hat Enterprise Linux documentation.,  - The Amazon Resource Name (ARN) of the NFS location to update.amazonka-datasyncUndocumented member.amazonka-datasyncUndocumented member.amazonka-datasyncThe subdirectory in the NFS file system that is used to read data from the NFS source location or write data to the NFS destination. The NFS path should be a path that's exported by the NFS server, or a subdirectory of that path. The path should be such that it can be mounted by other NFS clients in your network.8To see all the paths exported by your NFS server, run "showmount -e nfs-server-name" from an NFS client that has access to your server. You can specify any directory that appears in the results, and any subdirectory of that directory. Ensure that the NFS export is accessible without Kerberos authentication.To transfer all the data in the folder that you specified, DataSync must have permissions to read all the data. To ensure this, either configure the NFS export with no_root_squash, or ensure that the files you want DataSync to access have permissions that allow read access for all users. Doing either option enables the agent to read the files. For the agent to access directories, you must additionally enable all execute access.>If you are copying data to or from your Snowcone device, see  https://docs.aws.amazon.com/datasync/latest/userguide/create-nfs-location.html#nfs-on-snowconeNFS Server on Snowcone for more information.For information about NFS export configuration, see 18.7. The /etc/exports Configuration File in the Red Hat Enterprise Linux documentation.amazonka-datasync=The Amazon Resource Name (ARN) of the NFS location to update.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasync_(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasyncSpecifies the access key (for example, a user name) if credentials are required to authenticate with the object storage server.amazonka-datasyncSpecifies the Amazon Resource Names (ARNs) of the DataSync agents that can securely connect with your location.amazonka-datasyncSpecifies the secret key (for example, a password) if credentials are required to authenticate with the object storage server.amazonka-datasyncSpecifies a certificate to authenticate with an object storage system that uses a private or self-signed certificate authority (CA). You must specify a Base64-encoded .pem file (for example, 2file:///home/user/.ssh/storage_sys_certificate.pem). The certificate can be up to 32768 bytes (before Base64 encoding).!To use this parameter, configure ServerProtocol to HTTPS.Updating the certificate doesn't interfere with tasks that you have in progress.amazonka-datasyncSpecifies the port that your object storage server accepts inbound network traffic on (for example, port 443).amazonka-datasyncSpecifies the protocol that your object storage server uses to communicate.amazonka-datasyncSpecifies the object prefix for your object storage server. If this is a source location, DataSync only copies objects with this prefix. If this is a destination location, DataSync writes all objects with this prefix.amazonka-datasyncSpecifies the ARN of the object storage system location that you're updating.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the access key (for example, a user name) if credentials are required to authenticate with the object storage server.,  - Specifies the Amazon Resource Names (ARNs) of the DataSync agents that can securely connect with your location.,  - Specifies the secret key (for example, a password) if credentials are required to authenticate with the object storage server.,  - Specifies a certificate to authenticate with an object storage system that uses a private or self-signed certificate authority (CA). You must specify a Base64-encoded .pem file (for example, 2file:///home/user/.ssh/storage_sys_certificate.pem). The certificate can be up to 32768 bytes (before Base64 encoding).!To use this parameter, configure ServerProtocol to HTTPS.Updating the certificate doesn't interfere with tasks that you have in progress.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.,  - Specifies the port that your object storage server accepts inbound network traffic on (for example, port 443).,  - Specifies the protocol that your object storage server uses to communicate.,  - Specifies the object prefix for your object storage server. If this is a source location, DataSync only copies objects with this prefix. If this is a destination location, DataSync writes all objects with this prefix.,  - Specifies the ARN of the object storage system location that you're updating.amazonka-datasyncSpecifies the access key (for example, a user name) if credentials are required to authenticate with the object storage server.amazonka-datasyncSpecifies the Amazon Resource Names (ARNs) of the DataSync agents that can securely connect with your location.amazonka-datasyncSpecifies the secret key (for example, a password) if credentials are required to authenticate with the object storage server.amazonka-datasyncSpecifies a certificate to authenticate with an object storage system that uses a private or self-signed certificate authority (CA). You must specify a Base64-encoded .pem file (for example, 2file:///home/user/.ssh/storage_sys_certificate.pem). The certificate can be up to 32768 bytes (before Base64 encoding).!To use this parameter, configure ServerProtocol to HTTPS.Updating the certificate doesn't interfere with tasks that you have in progress.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.amazonka-datasyncSpecifies the port that your object storage server accepts inbound network traffic on (for example, port 443).amazonka-datasyncSpecifies the protocol that your object storage server uses to communicate.amazonka-datasyncSpecifies the object prefix for your object storage server. If this is a source location, DataSync only copies objects with this prefix. If this is a destination location, DataSync writes all objects with this prefix.amazonka-datasyncSpecifies the ARN of the object storage system location that you're updating.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasync`(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasyncThe Amazon Resource Names (ARNs) of agents to use for a Simple Message Block (SMB) location.amazonka-datasync>The name of the Windows domain that the SMB server belongs to.amazonka-datasyncThe password of the user who can mount the share has the permissions to access files and folders in the SMB share.amazonka-datasyncThe subdirectory in the SMB file system that is used to read data from the SMB source location or write data to the SMB destination. The SMB path should be a path that's exported by the SMB server, or a subdirectory of that path. The path should be such that it can be mounted by other SMB clients in your network. Subdirectory7 must be specified with forward slashes. For example, /path/to/folder.To transfer all the data in the folder that you specified, DataSync must have permissions to mount the SMB share and to access all the data in that share. To ensure this, do either of the following:Ensure that the user/password specified belongs to the user who can mount the share and who has the appropriate permissions for all of the files and directories that you want DataSync to access.Use credentials of a member of the Backup Operators group to mount the share.Doing either of these options enables the agent to access the data. For the agent to access directories, you must also enable all execute access.amazonka-datasyncThe user who can mount the share has the permissions to access files and folders in the SMB share.amazonka-datasync=The Amazon Resource Name (ARN) of the SMB location to update.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Names (ARNs) of agents to use for a Simple Message Block (SMB) location.,  - The name of the Windows domain that the SMB server belongs to.,  - Undocumented member.,  - The password of the user who can mount the share has the permissions to access files and folders in the SMB share.,  - The subdirectory in the SMB file system that is used to read data from the SMB source location or write data to the SMB destination. The SMB path should be a path that's exported by the SMB server, or a subdirectory of that path. The path should be such that it can be mounted by other SMB clients in your network. Subdirectory7 must be specified with forward slashes. For example, /path/to/folder.To transfer all the data in the folder that you specified, DataSync must have permissions to mount the SMB share and to access all the data in that share. To ensure this, do either of the following:Ensure that the user/password specified belongs to the user who can mount the share and who has the appropriate permissions for all of the files and directories that you want DataSync to access.Use credentials of a member of the Backup Operators group to mount the share.Doing either of these options enables the agent to access the data. For the agent to access directories, you must also enable all execute access.,  - The user who can mount the share has the permissions to access files and folders in the SMB share.,  - The Amazon Resource Name (ARN) of the SMB location to update.amazonka-datasyncThe Amazon Resource Names (ARNs) of agents to use for a Simple Message Block (SMB) location.amazonka-datasync>The name of the Windows domain that the SMB server belongs to.amazonka-datasyncUndocumented member.amazonka-datasyncThe password of the user who can mount the share has the permissions to access files and folders in the SMB share.amazonka-datasyncThe subdirectory in the SMB file system that is used to read data from the SMB source location or write data to the SMB destination. The SMB path should be a path that's exported by the SMB server, or a subdirectory of that path. The path should be such that it can be mounted by other SMB clients in your network. Subdirectory7 must be specified with forward slashes. For example, /path/to/folder.To transfer all the data in the folder that you specified, DataSync must have permissions to mount the SMB share and to access all the data in that share. To ensure this, do either of the following:Ensure that the user/password specified belongs to the user who can mount the share and who has the appropriate permissions for all of the files and directories that you want DataSync to access.Use credentials of a member of the Backup Operators group to mount the share.Doing either of these options enables the agent to access the data. For the agent to access directories, you must also enable all execute access.amazonka-datasyncThe user who can mount the share has the permissions to access files and folders in the SMB share.amazonka-datasync=The Amazon Resource Name (ARN) of the SMB location to update.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasynca(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';2amazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncUpdateTaskResponseSee:  smart constructor.amazonka-datasyncThe Amazon Resource Name (ARN) of the resource name of the Amazon CloudWatch log group.amazonka-datasyncSpecifies a list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.amazonka-datasyncSpecifies a list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.amazonka-datasyncThe name of the task to update.amazonka-datasyncSpecifies a schedule used to periodically transfer files from a source to a destination location. You can configure your task to execute hourly, daily, weekly or on specific days of the week. You control when in the day or hour you want the task to execute. The time you specify is UTC time. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/task-scheduling.htmlScheduling your task.amazonka-datasyncThe Amazon Resource Name (ARN) of the resource name of the task to update.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the resource name of the Amazon CloudWatch log group.,  - Specifies a list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.,  - Specifies a list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync., " - The name of the task to update.,  - Undocumented member.,  - Specifies a schedule used to periodically transfer files from a source to a destination location. You can configure your task to execute hourly, daily, weekly or on specific days of the week. You control when in the day or hour you want the task to execute. The time you specify is UTC time. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/task-scheduling.htmlScheduling your task.,  - The Amazon Resource Name (ARN) of the resource name of the task to update.amazonka-datasyncThe Amazon Resource Name (ARN) of the resource name of the Amazon CloudWatch log group.amazonka-datasyncSpecifies a list of filter rules that exclude specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.amazonka-datasyncSpecifies a list of filter rules that include specific data during your transfer. For more information and examples, see  https://docs.aws.amazon.com/datasync/latest/userguide/filtering.html&Filtering data transferred by DataSync.amazonka-datasyncThe name of the task to update.amazonka-datasyncUndocumented member.amazonka-datasyncSpecifies a schedule used to periodically transfer files from a source to a destination location. You can configure your task to execute hourly, daily, weekly or on specific days of the week. You control when in the day or hour you want the task to execute. The time you specify is UTC time. For more information, see  https://docs.aws.amazon.com/datasync/latest/userguide/task-scheduling.htmlScheduling your task.amazonka-datasyncThe Amazon Resource Name (ARN) of the resource name of the task to update.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncb(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';1 amazonka-datasyncSee:  smart constructor.amazonka-datasync The response's http status code.amazonka-datasyncSee:  smart constructor.amazonka-datasyncThe Amazon Resource Name (ARN) of the specific task execution that is being updated.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the specific task execution that is being updated.,  - Undocumented member.amazonka-datasyncThe Amazon Resource Name (ARN) of the specific task execution that is being updated.amazonka-datasyncUndocumented member.amazonka-datasyncCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-datasync The response's http status code.amazonka-datasyncamazonka-datasyncamazonka-datasync  c(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred EF    EFd(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%e(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred~(,+)*@ADOSRPQglkjhi   (,+)*,+OSRPQSRglkjhilkj@ADfghijklmnopqrstuvwxyz{|}~                                                                                                                                    !!!!!!!!!!!!!""""""""""""""""""""""""""""""#########################$$$$$$$$$$$$$$$$$$$$$$$$$%%%%%%%%%%%%%&&&&&&&&&&&&&&&&&&''''''''''''''((((((((((((((())))))))))))))))))))))))))))*****************************++++++++++++++,,,,,,,,,,,,,,,,,,,,,,,,----------------......................../////////////0000000000000000000000000001111111111111111222222222222222222222 2 2 2 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 6 6 6 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = = > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A A B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B B C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C C D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D D DDDDDDDDDDDDEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEEFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFGGGGGGGGGGGGGGGGGGGGGGGGGGGGGGGGGGGGGGGHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHHIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJJKKKKKKKKKKKKKKKKKKKKKKKKKKLLLLLLLLLLLLLLLLLLLLLLLLLLMMMMMMMMMMMMMMMMMMMMMMMMMMNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNNOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQQRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRRSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTTUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUUVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVVWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWWXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYZZZZZZZZZZZZZZZZZZZZZZZZZZ[[[[[[[[[[[[[[[[[[[[[[[[[[[[\\\\\\\\\\\\\\\\\\\\\\\\\\\\]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]]^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^_______________________________________`````````````````````````````````````aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaabbbbbbbbbbbbbbbbbbbbbbbbbbbb,amazonka-datasync-2.0-LCcm4ddRe2z4CES4wun76Z#Amazonka.DataSync.Types.AgentStatus&Amazonka.DataSync.Types.AgentListEntryAmazonka.DataSync.Types.Atime!Amazonka.DataSync.Types.Ec2Config.Amazonka.DataSync.Types.EfsInTransitEncryption$Amazonka.DataSync.Types.EndpointType"Amazonka.DataSync.Types.FilterType"Amazonka.DataSync.Types.FilterRuleAmazonka.DataSync.Types.Gid.Amazonka.DataSync.Types.HdfsAuthenticationType2Amazonka.DataSync.Types.HdfsDataTransferProtection$Amazonka.DataSync.Types.HdfsNameNode)Amazonka.DataSync.Types.HdfsRpcProtection*Amazonka.DataSync.Types.LocationFilterName)Amazonka.DataSync.Types.LocationListEntry Amazonka.DataSync.Types.LogLevelAmazonka.DataSync.Types.Mtime"Amazonka.DataSync.Types.NfsVersion'Amazonka.DataSync.Types.NfsMountOptions&Amazonka.DataSync.Types.FsxProtocolNfs3Amazonka.DataSync.Types.ObjectStorageServerProtocol"Amazonka.DataSync.Types.ObjectTags$Amazonka.DataSync.Types.OnPremConfig Amazonka.DataSync.Types.Operator&Amazonka.DataSync.Types.LocationFilter%Amazonka.DataSync.Types.OverwriteMode#Amazonka.DataSync.Types.PhaseStatus(Amazonka.DataSync.Types.PosixPermissions,Amazonka.DataSync.Types.PreserveDeletedFiles'Amazonka.DataSync.Types.PreserveDevices)Amazonka.DataSync.Types.PrivateLinkConfig(Amazonka.DataSync.Types.QopConfiguration Amazonka.DataSync.Types.S3Config&Amazonka.DataSync.Types.S3StorageClass6Amazonka.DataSync.Types.SmbSecurityDescriptorCopyFlags"Amazonka.DataSync.Types.SmbVersion'Amazonka.DataSync.Types.SmbMountOptions&Amazonka.DataSync.Types.FsxProtocolSmb#Amazonka.DataSync.Types.FsxProtocol$Amazonka.DataSync.Types.TagListEntry1Amazonka.DataSync.Types.TaskExecutionResultDetail+Amazonka.DataSync.Types.TaskExecutionStatus.Amazonka.DataSync.Types.TaskExecutionListEntry&Amazonka.DataSync.Types.TaskFilterName"Amazonka.DataSync.Types.TaskFilter$Amazonka.DataSync.Types.TaskQueueing$Amazonka.DataSync.Types.TaskSchedule"Amazonka.DataSync.Types.TaskStatus%Amazonka.DataSync.Types.TaskListEntry$Amazonka.DataSync.Types.TransferModeAmazonka.DataSync.Types.Uid"Amazonka.DataSync.Types.VerifyModeAmazonka.DataSync.Types.OptionsAmazonka.DataSync.TypesAmazonka.DataSync.TagResource$Amazonka.DataSync.StartTaskExecutionAmazonka.DataSync.ListTasks$Amazonka.DataSync.ListTaskExecutions%Amazonka.DataSync.ListTagsForResourceAmazonka.DataSync.ListLocationsAmazonka.DataSync.ListAgents'Amazonka.DataSync.DescribeTaskExecutionAmazonka.DataSync.DescribeTask%Amazonka.DataSync.DescribeLocationSmb$Amazonka.DataSync.DescribeLocationS3/Amazonka.DataSync.DescribeLocationObjectStorage%Amazonka.DataSync.DescribeLocationNfs&Amazonka.DataSync.DescribeLocationHdfs,Amazonka.DataSync.DescribeLocationFsxWindows,Amazonka.DataSync.DescribeLocationFsxOpenZfs*Amazonka.DataSync.DescribeLocationFsxOntap+Amazonka.DataSync.DescribeLocationFsxLustre%Amazonka.DataSync.DescribeLocationEfsAmazonka.DataSync.DescribeAgentAmazonka.DataSync.DeleteTask Amazonka.DataSync.DeleteLocationAmazonka.DataSync.DeleteAgentAmazonka.DataSync.CreateTask#Amazonka.DataSync.CreateLocationSmb"Amazonka.DataSync.CreateLocationS3-Amazonka.DataSync.CreateLocationObjectStorage#Amazonka.DataSync.CreateLocationNfs$Amazonka.DataSync.CreateLocationHdfs*Amazonka.DataSync.CreateLocationFsxWindows*Amazonka.DataSync.CreateLocationFsxOpenZfs(Amazonka.DataSync.CreateLocationFsxOntap)Amazonka.DataSync.CreateLocationFsxLustre#Amazonka.DataSync.CreateLocationEfsAmazonka.DataSync.CreateAgent%Amazonka.DataSync.CancelTaskExecutionAmazonka.DataSync.UntagResourceAmazonka.DataSync.UpdateAgent$Amazonka.DataSync.UpdateLocationHdfs#Amazonka.DataSync.UpdateLocationNfs-Amazonka.DataSync.UpdateLocationObjectStorage#Amazonka.DataSync.UpdateLocationSmbAmazonka.DataSync.UpdateTask%Amazonka.DataSync.UpdateTaskExecutionAmazonka.DataSync.LensAmazonka.DataSync.WaitersAmazonka.DataSync AgentStatus AgentStatus'fromAgentStatusAgentStatus_ONLINEAgentStatus_OFFLINE$fShowAgentStatus$fReadAgentStatus$fEqAgentStatus$fOrdAgentStatus$fGenericAgentStatus$fHashableAgentStatus$fNFDataAgentStatus$fFromTextAgentStatus$fToTextAgentStatus$fToByteStringAgentStatus$fToLogAgentStatus$fToHeaderAgentStatus$fToQueryAgentStatus$fFromJSONAgentStatus$fFromJSONKeyAgentStatus$fToJSONAgentStatus$fToJSONKeyAgentStatus$fFromXMLAgentStatus$fToXMLAgentStatusAgentListEntryAgentListEntry'$sel:agentArn:AgentListEntry'$sel:name:AgentListEntry'$sel:status:AgentListEntry'newAgentListEntryagentListEntry_agentArnagentListEntry_nameagentListEntry_status$fNFDataAgentListEntry$fHashableAgentListEntry$fFromJSONAgentListEntry$fEqAgentListEntry$fReadAgentListEntry$fShowAgentListEntry$fGenericAgentListEntryAtimeAtime' fromAtime Atime_NONEAtime_BEST_EFFORT $fShowAtime $fReadAtime $fEqAtime $fOrdAtime$fGenericAtime$fHashableAtime $fNFDataAtime$fFromTextAtime $fToTextAtime$fToByteStringAtime $fToLogAtime$fToHeaderAtime$fToQueryAtime$fFromJSONAtime$fFromJSONKeyAtime $fToJSONAtime$fToJSONKeyAtime$fFromXMLAtime $fToXMLAtime Ec2Config Ec2Config'$sel:subnetArn:Ec2Config'!$sel:securityGroupArns:Ec2Config' newEc2Configec2Config_subnetArnec2Config_securityGroupArns$fToJSONEc2Config$fNFDataEc2Config$fHashableEc2Config$fFromJSONEc2Config $fEqEc2Config$fReadEc2Config$fShowEc2Config$fGenericEc2ConfigEfsInTransitEncryptionEfsInTransitEncryption'fromEfsInTransitEncryptionEfsInTransitEncryption_TLS1_2EfsInTransitEncryption_NONE$fShowEfsInTransitEncryption$fReadEfsInTransitEncryption$fEqEfsInTransitEncryption$fOrdEfsInTransitEncryption$fGenericEfsInTransitEncryption $fHashableEfsInTransitEncryption$fNFDataEfsInTransitEncryption $fFromTextEfsInTransitEncryption$fToTextEfsInTransitEncryption$$fToByteStringEfsInTransitEncryption$fToLogEfsInTransitEncryption $fToHeaderEfsInTransitEncryption$fToQueryEfsInTransitEncryption $fFromJSONEfsInTransitEncryption#$fFromJSONKeyEfsInTransitEncryption$fToJSONEfsInTransitEncryption!$fToJSONKeyEfsInTransitEncryption$fFromXMLEfsInTransitEncryption$fToXMLEfsInTransitEncryption EndpointType EndpointType'fromEndpointTypeEndpointType_PUBLICEndpointType_PRIVATE_LINKEndpointType_FIPS$fShowEndpointType$fReadEndpointType$fEqEndpointType$fOrdEndpointType$fGenericEndpointType$fHashableEndpointType$fNFDataEndpointType$fFromTextEndpointType$fToTextEndpointType$fToByteStringEndpointType$fToLogEndpointType$fToHeaderEndpointType$fToQueryEndpointType$fFromJSONEndpointType$fFromJSONKeyEndpointType$fToJSONEndpointType$fToJSONKeyEndpointType$fFromXMLEndpointType$fToXMLEndpointType FilterType FilterType'fromFilterTypeFilterType_SIMPLE_PATTERN$fShowFilterType$fReadFilterType$fEqFilterType$fOrdFilterType$fGenericFilterType$fHashableFilterType$fNFDataFilterType$fFromTextFilterType$fToTextFilterType$fToByteStringFilterType$fToLogFilterType$fToHeaderFilterType$fToQueryFilterType$fFromJSONFilterType$fFromJSONKeyFilterType$fToJSONFilterType$fToJSONKeyFilterType$fFromXMLFilterType$fToXMLFilterType FilterRule FilterRule'$sel:filterType:FilterRule'$sel:value:FilterRule' newFilterRulefilterRule_filterTypefilterRule_value$fToJSONFilterRule$fNFDataFilterRule$fHashableFilterRule$fFromJSONFilterRule$fEqFilterRule$fReadFilterRule$fShowFilterRule$fGenericFilterRuleGidGid'fromGidGid_NONEGid_NAME Gid_INT_VALUEGid_BOTH $fShowGid $fReadGid$fEqGid$fOrdGid $fGenericGid $fHashableGid $fNFDataGid $fFromTextGid $fToTextGid$fToByteStringGid $fToLogGid $fToHeaderGid $fToQueryGid $fFromJSONGid$fFromJSONKeyGid $fToJSONGid$fToJSONKeyGid $fFromXMLGid $fToXMLGidHdfsAuthenticationTypeHdfsAuthenticationType'fromHdfsAuthenticationTypeHdfsAuthenticationType_SIMPLEHdfsAuthenticationType_KERBEROS$fShowHdfsAuthenticationType$fReadHdfsAuthenticationType$fEqHdfsAuthenticationType$fOrdHdfsAuthenticationType$fGenericHdfsAuthenticationType $fHashableHdfsAuthenticationType$fNFDataHdfsAuthenticationType $fFromTextHdfsAuthenticationType$fToTextHdfsAuthenticationType$$fToByteStringHdfsAuthenticationType$fToLogHdfsAuthenticationType $fToHeaderHdfsAuthenticationType$fToQueryHdfsAuthenticationType $fFromJSONHdfsAuthenticationType#$fFromJSONKeyHdfsAuthenticationType$fToJSONHdfsAuthenticationType!$fToJSONKeyHdfsAuthenticationType$fFromXMLHdfsAuthenticationType$fToXMLHdfsAuthenticationTypeHdfsDataTransferProtectionHdfsDataTransferProtection'fromHdfsDataTransferProtection"HdfsDataTransferProtection_PRIVACY$HdfsDataTransferProtection_INTEGRITY#HdfsDataTransferProtection_DISABLED)HdfsDataTransferProtection_AUTHENTICATION $fShowHdfsDataTransferProtection $fReadHdfsDataTransferProtection$fEqHdfsDataTransferProtection$fOrdHdfsDataTransferProtection#$fGenericHdfsDataTransferProtection$$fHashableHdfsDataTransferProtection"$fNFDataHdfsDataTransferProtection$$fFromTextHdfsDataTransferProtection"$fToTextHdfsDataTransferProtection($fToByteStringHdfsDataTransferProtection!$fToLogHdfsDataTransferProtection$$fToHeaderHdfsDataTransferProtection#$fToQueryHdfsDataTransferProtection$$fFromJSONHdfsDataTransferProtection'$fFromJSONKeyHdfsDataTransferProtection"$fToJSONHdfsDataTransferProtection%$fToJSONKeyHdfsDataTransferProtection#$fFromXMLHdfsDataTransferProtection!$fToXMLHdfsDataTransferProtection HdfsNameNode HdfsNameNode'$sel:hostname:HdfsNameNode'$sel:port:HdfsNameNode'newHdfsNameNodehdfsNameNode_hostnamehdfsNameNode_port$fToJSONHdfsNameNode$fNFDataHdfsNameNode$fHashableHdfsNameNode$fFromJSONHdfsNameNode$fEqHdfsNameNode$fReadHdfsNameNode$fShowHdfsNameNode$fGenericHdfsNameNodeHdfsRpcProtectionHdfsRpcProtection'fromHdfsRpcProtectionHdfsRpcProtection_PRIVACYHdfsRpcProtection_INTEGRITYHdfsRpcProtection_DISABLED HdfsRpcProtection_AUTHENTICATION$fShowHdfsRpcProtection$fReadHdfsRpcProtection$fEqHdfsRpcProtection$fOrdHdfsRpcProtection$fGenericHdfsRpcProtection$fHashableHdfsRpcProtection$fNFDataHdfsRpcProtection$fFromTextHdfsRpcProtection$fToTextHdfsRpcProtection$fToByteStringHdfsRpcProtection$fToLogHdfsRpcProtection$fToHeaderHdfsRpcProtection$fToQueryHdfsRpcProtection$fFromJSONHdfsRpcProtection$fFromJSONKeyHdfsRpcProtection$fToJSONHdfsRpcProtection$fToJSONKeyHdfsRpcProtection$fFromXMLHdfsRpcProtection$fToXMLHdfsRpcProtectionLocationFilterNameLocationFilterName'fromLocationFilterNameLocationFilterName_LocationUriLocationFilterName_LocationTypeLocationFilterName_CreationTime$fShowLocationFilterName$fReadLocationFilterName$fEqLocationFilterName$fOrdLocationFilterName$fGenericLocationFilterName$fHashableLocationFilterName$fNFDataLocationFilterName$fFromTextLocationFilterName$fToTextLocationFilterName $fToByteStringLocationFilterName$fToLogLocationFilterName$fToHeaderLocationFilterName$fToQueryLocationFilterName$fFromJSONLocationFilterName$fFromJSONKeyLocationFilterName$fToJSONLocationFilterName$fToJSONKeyLocationFilterName$fFromXMLLocationFilterName$fToXMLLocationFilterNameLocationListEntryLocationListEntry'#$sel:locationArn:LocationListEntry'#$sel:locationUri:LocationListEntry'newLocationListEntrylocationListEntry_locationArnlocationListEntry_locationUri$fNFDataLocationListEntry$fHashableLocationListEntry$fFromJSONLocationListEntry$fEqLocationListEntry$fReadLocationListEntry$fShowLocationListEntry$fGenericLocationListEntryLogLevel LogLevel' fromLogLevelLogLevel_TRANSFER LogLevel_OFFLogLevel_BASIC$fShowLogLevel$fReadLogLevel $fEqLogLevel $fOrdLogLevel$fGenericLogLevel$fHashableLogLevel$fNFDataLogLevel$fFromTextLogLevel$fToTextLogLevel$fToByteStringLogLevel$fToLogLogLevel$fToHeaderLogLevel$fToQueryLogLevel$fFromJSONLogLevel$fFromJSONKeyLogLevel$fToJSONLogLevel$fToJSONKeyLogLevel$fFromXMLLogLevel$fToXMLLogLevelMtimeMtime' fromMtimeMtime_PRESERVE Mtime_NONE $fShowMtime $fReadMtime $fEqMtime $fOrdMtime$fGenericMtime$fHashableMtime $fNFDataMtime$fFromTextMtime $fToTextMtime$fToByteStringMtime $fToLogMtime$fToHeaderMtime$fToQueryMtime$fFromJSONMtime$fFromJSONKeyMtime $fToJSONMtime$fToJSONKeyMtime$fFromXMLMtime $fToXMLMtime NfsVersion NfsVersion'fromNfsVersionNfsVersion_NFS4_1NfsVersion_NFS4_0NfsVersion_NFS3NfsVersion_AUTOMATIC$fShowNfsVersion$fReadNfsVersion$fEqNfsVersion$fOrdNfsVersion$fGenericNfsVersion$fHashableNfsVersion$fNFDataNfsVersion$fFromTextNfsVersion$fToTextNfsVersion$fToByteStringNfsVersion$fToLogNfsVersion$fToHeaderNfsVersion$fToQueryNfsVersion$fFromJSONNfsVersion$fFromJSONKeyNfsVersion$fToJSONNfsVersion$fToJSONKeyNfsVersion$fFromXMLNfsVersion$fToXMLNfsVersionNfsMountOptionsNfsMountOptions'$sel:version:NfsMountOptions'newNfsMountOptionsnfsMountOptions_version$fToJSONNfsMountOptions$fNFDataNfsMountOptions$fHashableNfsMountOptions$fFromJSONNfsMountOptions$fEqNfsMountOptions$fReadNfsMountOptions$fShowNfsMountOptions$fGenericNfsMountOptionsFsxProtocolNfsFsxProtocolNfs'!$sel:mountOptions:FsxProtocolNfs'newFsxProtocolNfsfsxProtocolNfs_mountOptions$fToJSONFsxProtocolNfs$fNFDataFsxProtocolNfs$fHashableFsxProtocolNfs$fFromJSONFsxProtocolNfs$fEqFsxProtocolNfs$fReadFsxProtocolNfs$fShowFsxProtocolNfs$fGenericFsxProtocolNfsObjectStorageServerProtocolObjectStorageServerProtocol'fromObjectStorageServerProtocol!ObjectStorageServerProtocol_HTTPS ObjectStorageServerProtocol_HTTP!$fShowObjectStorageServerProtocol!$fReadObjectStorageServerProtocol$fEqObjectStorageServerProtocol $fOrdObjectStorageServerProtocol$$fGenericObjectStorageServerProtocol%$fHashableObjectStorageServerProtocol#$fNFDataObjectStorageServerProtocol%$fFromTextObjectStorageServerProtocol#$fToTextObjectStorageServerProtocol)$fToByteStringObjectStorageServerProtocol"$fToLogObjectStorageServerProtocol%$fToHeaderObjectStorageServerProtocol$$fToQueryObjectStorageServerProtocol%$fFromJSONObjectStorageServerProtocol($fFromJSONKeyObjectStorageServerProtocol#$fToJSONObjectStorageServerProtocol&$fToJSONKeyObjectStorageServerProtocol$$fFromXMLObjectStorageServerProtocol"$fToXMLObjectStorageServerProtocol ObjectTags ObjectTags'fromObjectTagsObjectTags_PRESERVEObjectTags_NONE$fShowObjectTags$fReadObjectTags$fEqObjectTags$fOrdObjectTags$fGenericObjectTags$fHashableObjectTags$fNFDataObjectTags$fFromTextObjectTags$fToTextObjectTags$fToByteStringObjectTags$fToLogObjectTags$fToHeaderObjectTags$fToQueryObjectTags$fFromJSONObjectTags$fFromJSONKeyObjectTags$fToJSONObjectTags$fToJSONKeyObjectTags$fFromXMLObjectTags$fToXMLObjectTags OnPremConfig OnPremConfig'$sel:agentArns:OnPremConfig'newOnPremConfigonPremConfig_agentArns$fToJSONOnPremConfig$fNFDataOnPremConfig$fHashableOnPremConfig$fFromJSONOnPremConfig$fEqOnPremConfig$fReadOnPremConfig$fShowOnPremConfig$fGenericOnPremConfigOperator Operator' fromOperatorOperator_NotEqualsOperator_NotContainsOperator_LessThanOrEqualOperator_LessThan Operator_InOperator_GreaterThanOrEqualOperator_GreaterThanOperator_EqualsOperator_ContainsOperator_BeginsWith$fShowOperator$fReadOperator $fEqOperator $fOrdOperator$fGenericOperator$fHashableOperator$fNFDataOperator$fFromTextOperator$fToTextOperator$fToByteStringOperator$fToLogOperator$fToHeaderOperator$fToQueryOperator$fFromJSONOperator$fFromJSONKeyOperator$fToJSONOperator$fToJSONKeyOperator$fFromXMLOperator$fToXMLOperatorLocationFilterLocationFilter'$sel:name:LocationFilter'$sel:values:LocationFilter'$sel:operator:LocationFilter'newLocationFilterlocationFilter_namelocationFilter_valueslocationFilter_operator$fToJSONLocationFilter$fNFDataLocationFilter$fHashableLocationFilter$fEqLocationFilter$fReadLocationFilter$fShowLocationFilter$fGenericLocationFilter OverwriteModeOverwriteMode'fromOverwriteModeOverwriteMode_NEVEROverwriteMode_ALWAYS$fShowOverwriteMode$fReadOverwriteMode$fEqOverwriteMode$fOrdOverwriteMode$fGenericOverwriteMode$fHashableOverwriteMode$fNFDataOverwriteMode$fFromTextOverwriteMode$fToTextOverwriteMode$fToByteStringOverwriteMode$fToLogOverwriteMode$fToHeaderOverwriteMode$fToQueryOverwriteMode$fFromJSONOverwriteMode$fFromJSONKeyOverwriteMode$fToJSONOverwriteMode$fToJSONKeyOverwriteMode$fFromXMLOverwriteMode$fToXMLOverwriteMode PhaseStatus PhaseStatus'fromPhaseStatusPhaseStatus_SUCCESSPhaseStatus_PENDINGPhaseStatus_ERROR$fShowPhaseStatus$fReadPhaseStatus$fEqPhaseStatus$fOrdPhaseStatus$fGenericPhaseStatus$fHashablePhaseStatus$fNFDataPhaseStatus$fFromTextPhaseStatus$fToTextPhaseStatus$fToByteStringPhaseStatus$fToLogPhaseStatus$fToHeaderPhaseStatus$fToQueryPhaseStatus$fFromJSONPhaseStatus$fFromJSONKeyPhaseStatus$fToJSONPhaseStatus$fToJSONKeyPhaseStatus$fFromXMLPhaseStatus$fToXMLPhaseStatusPosixPermissionsPosixPermissions'fromPosixPermissionsPosixPermissions_PRESERVEPosixPermissions_NONE$fShowPosixPermissions$fReadPosixPermissions$fEqPosixPermissions$fOrdPosixPermissions$fGenericPosixPermissions$fHashablePosixPermissions$fNFDataPosixPermissions$fFromTextPosixPermissions$fToTextPosixPermissions$fToByteStringPosixPermissions$fToLogPosixPermissions$fToHeaderPosixPermissions$fToQueryPosixPermissions$fFromJSONPosixPermissions$fFromJSONKeyPosixPermissions$fToJSONPosixPermissions$fToJSONKeyPosixPermissions$fFromXMLPosixPermissions$fToXMLPosixPermissionsPreserveDeletedFilesPreserveDeletedFiles'fromPreserveDeletedFilesPreserveDeletedFiles_REMOVEPreserveDeletedFiles_PRESERVE$fShowPreserveDeletedFiles$fReadPreserveDeletedFiles$fEqPreserveDeletedFiles$fOrdPreserveDeletedFiles$fGenericPreserveDeletedFiles$fHashablePreserveDeletedFiles$fNFDataPreserveDeletedFiles$fFromTextPreserveDeletedFiles$fToTextPreserveDeletedFiles"$fToByteStringPreserveDeletedFiles$fToLogPreserveDeletedFiles$fToHeaderPreserveDeletedFiles$fToQueryPreserveDeletedFiles$fFromJSONPreserveDeletedFiles!$fFromJSONKeyPreserveDeletedFiles$fToJSONPreserveDeletedFiles$fToJSONKeyPreserveDeletedFiles$fFromXMLPreserveDeletedFiles$fToXMLPreserveDeletedFilesPreserveDevicesPreserveDevices'fromPreserveDevicesPreserveDevices_PRESERVEPreserveDevices_NONE$fShowPreserveDevices$fReadPreserveDevices$fEqPreserveDevices$fOrdPreserveDevices$fGenericPreserveDevices$fHashablePreserveDevices$fNFDataPreserveDevices$fFromTextPreserveDevices$fToTextPreserveDevices$fToByteStringPreserveDevices$fToLogPreserveDevices$fToHeaderPreserveDevices$fToQueryPreserveDevices$fFromJSONPreserveDevices$fFromJSONKeyPreserveDevices$fToJSONPreserveDevices$fToJSONKeyPreserveDevices$fFromXMLPreserveDevices$fToXMLPreserveDevicesPrivateLinkConfigPrivateLinkConfig'+$sel:privateLinkEndpoint:PrivateLinkConfig')$sel:securityGroupArns:PrivateLinkConfig'"$sel:subnetArns:PrivateLinkConfig'%$sel:vpcEndpointId:PrivateLinkConfig'newPrivateLinkConfig%privateLinkConfig_privateLinkEndpoint#privateLinkConfig_securityGroupArnsprivateLinkConfig_subnetArnsprivateLinkConfig_vpcEndpointId$fNFDataPrivateLinkConfig$fHashablePrivateLinkConfig$fFromJSONPrivateLinkConfig$fEqPrivateLinkConfig$fReadPrivateLinkConfig$fShowPrivateLinkConfig$fGenericPrivateLinkConfigQopConfigurationQopConfiguration'-$sel:dataTransferProtection:QopConfiguration'$$sel:rpcProtection:QopConfiguration'newQopConfiguration'qopConfiguration_dataTransferProtectionqopConfiguration_rpcProtection$fToJSONQopConfiguration$fNFDataQopConfiguration$fHashableQopConfiguration$fFromJSONQopConfiguration$fEqQopConfiguration$fReadQopConfiguration$fShowQopConfiguration$fGenericQopConfigurationS3Config S3Config'"$sel:bucketAccessRoleArn:S3Config' newS3Configs3Config_bucketAccessRoleArn$fToJSONS3Config$fNFDataS3Config$fHashableS3Config$fFromJSONS3Config $fEqS3Config$fReadS3Config$fShowS3Config$fGenericS3ConfigS3StorageClassS3StorageClass'fromS3StorageClassS3StorageClass_STANDARD_IAS3StorageClass_STANDARDS3StorageClass_OUTPOSTSS3StorageClass_ONEZONE_IA"S3StorageClass_INTELLIGENT_TIERING(S3StorageClass_GLACIER_INSTANT_RETRIEVALS3StorageClass_GLACIERS3StorageClass_DEEP_ARCHIVE$fShowS3StorageClass$fReadS3StorageClass$fEqS3StorageClass$fOrdS3StorageClass$fGenericS3StorageClass$fHashableS3StorageClass$fNFDataS3StorageClass$fFromTextS3StorageClass$fToTextS3StorageClass$fToByteStringS3StorageClass$fToLogS3StorageClass$fToHeaderS3StorageClass$fToQueryS3StorageClass$fFromJSONS3StorageClass$fFromJSONKeyS3StorageClass$fToJSONS3StorageClass$fToJSONKeyS3StorageClass$fFromXMLS3StorageClass$fToXMLS3StorageClassSmbSecurityDescriptorCopyFlagsSmbSecurityDescriptorCopyFlags'"fromSmbSecurityDescriptorCopyFlags.SmbSecurityDescriptorCopyFlags_OWNER_DACL_SACL)SmbSecurityDescriptorCopyFlags_OWNER_DACL#SmbSecurityDescriptorCopyFlags_NONE$$fShowSmbSecurityDescriptorCopyFlags$$fReadSmbSecurityDescriptorCopyFlags"$fEqSmbSecurityDescriptorCopyFlags#$fOrdSmbSecurityDescriptorCopyFlags'$fGenericSmbSecurityDescriptorCopyFlags($fHashableSmbSecurityDescriptorCopyFlags&$fNFDataSmbSecurityDescriptorCopyFlags($fFromTextSmbSecurityDescriptorCopyFlags&$fToTextSmbSecurityDescriptorCopyFlags,$fToByteStringSmbSecurityDescriptorCopyFlags%$fToLogSmbSecurityDescriptorCopyFlags($fToHeaderSmbSecurityDescriptorCopyFlags'$fToQuerySmbSecurityDescriptorCopyFlags($fFromJSONSmbSecurityDescriptorCopyFlags+$fFromJSONKeySmbSecurityDescriptorCopyFlags&$fToJSONSmbSecurityDescriptorCopyFlags)$fToJSONKeySmbSecurityDescriptorCopyFlags'$fFromXMLSmbSecurityDescriptorCopyFlags%$fToXMLSmbSecurityDescriptorCopyFlags SmbVersion SmbVersion'fromSmbVersionSmbVersion_SMB3SmbVersion_SMB2SmbVersion_AUTOMATIC$fShowSmbVersion$fReadSmbVersion$fEqSmbVersion$fOrdSmbVersion$fGenericSmbVersion$fHashableSmbVersion$fNFDataSmbVersion$fFromTextSmbVersion$fToTextSmbVersion$fToByteStringSmbVersion$fToLogSmbVersion$fToHeaderSmbVersion$fToQuerySmbVersion$fFromJSONSmbVersion$fFromJSONKeySmbVersion$fToJSONSmbVersion$fToJSONKeySmbVersion$fFromXMLSmbVersion$fToXMLSmbVersionSmbMountOptionsSmbMountOptions'$sel:version:SmbMountOptions'newSmbMountOptionssmbMountOptions_version$fToJSONSmbMountOptions$fNFDataSmbMountOptions$fHashableSmbMountOptions$fFromJSONSmbMountOptions$fEqSmbMountOptions$fReadSmbMountOptions$fShowSmbMountOptions$fGenericSmbMountOptionsFsxProtocolSmbFsxProtocolSmb'$sel:domain:FsxProtocolSmb'!$sel:mountOptions:FsxProtocolSmb'$sel:password:FsxProtocolSmb'$sel:user:FsxProtocolSmb'newFsxProtocolSmbfsxProtocolSmb_domainfsxProtocolSmb_mountOptionsfsxProtocolSmb_passwordfsxProtocolSmb_user$fToJSONFsxProtocolSmb$fNFDataFsxProtocolSmb$fHashableFsxProtocolSmb$fFromJSONFsxProtocolSmb$fEqFsxProtocolSmb$fShowFsxProtocolSmb$fGenericFsxProtocolSmb FsxProtocol FsxProtocol'$sel:nfs:FsxProtocol'$sel:smb:FsxProtocol'newFsxProtocolfsxProtocol_nfsfsxProtocol_smb$fToJSONFsxProtocol$fNFDataFsxProtocol$fHashableFsxProtocol$fFromJSONFsxProtocol$fEqFsxProtocol$fShowFsxProtocol$fGenericFsxProtocol TagListEntry TagListEntry'$sel:value:TagListEntry'$sel:key:TagListEntry'newTagListEntrytagListEntry_valuetagListEntry_key$fToJSONTagListEntry$fNFDataTagListEntry$fHashableTagListEntry$fFromJSONTagListEntry$fEqTagListEntry$fReadTagListEntry$fShowTagListEntry$fGenericTagListEntryTaskExecutionResultDetailTaskExecutionResultDetail')$sel:errorCode:TaskExecutionResultDetail'+$sel:errorDetail:TaskExecutionResultDetail'/$sel:prepareDuration:TaskExecutionResultDetail'-$sel:prepareStatus:TaskExecutionResultDetail'-$sel:totalDuration:TaskExecutionResultDetail'0$sel:transferDuration:TaskExecutionResultDetail'.$sel:transferStatus:TaskExecutionResultDetail'.$sel:verifyDuration:TaskExecutionResultDetail',$sel:verifyStatus:TaskExecutionResultDetail'newTaskExecutionResultDetail#taskExecutionResultDetail_errorCode%taskExecutionResultDetail_errorDetail)taskExecutionResultDetail_prepareDuration'taskExecutionResultDetail_prepareStatus'taskExecutionResultDetail_totalDuration*taskExecutionResultDetail_transferDuration(taskExecutionResultDetail_transferStatus(taskExecutionResultDetail_verifyDuration&taskExecutionResultDetail_verifyStatus!$fNFDataTaskExecutionResultDetail#$fHashableTaskExecutionResultDetail#$fFromJSONTaskExecutionResultDetail$fEqTaskExecutionResultDetail$fReadTaskExecutionResultDetail$fShowTaskExecutionResultDetail"$fGenericTaskExecutionResultDetailTaskExecutionStatusTaskExecutionStatus'fromTaskExecutionStatusTaskExecutionStatus_VERIFYING TaskExecutionStatus_TRANSFERRINGTaskExecutionStatus_SUCCESSTaskExecutionStatus_QUEUEDTaskExecutionStatus_PREPARINGTaskExecutionStatus_LAUNCHINGTaskExecutionStatus_ERROR$fShowTaskExecutionStatus$fReadTaskExecutionStatus$fEqTaskExecutionStatus$fOrdTaskExecutionStatus$fGenericTaskExecutionStatus$fHashableTaskExecutionStatus$fNFDataTaskExecutionStatus$fFromTextTaskExecutionStatus$fToTextTaskExecutionStatus!$fToByteStringTaskExecutionStatus$fToLogTaskExecutionStatus$fToHeaderTaskExecutionStatus$fToQueryTaskExecutionStatus$fFromJSONTaskExecutionStatus $fFromJSONKeyTaskExecutionStatus$fToJSONTaskExecutionStatus$fToJSONKeyTaskExecutionStatus$fFromXMLTaskExecutionStatus$fToXMLTaskExecutionStatusTaskExecutionListEntryTaskExecutionListEntry'#$sel:status:TaskExecutionListEntry'-$sel:taskExecutionArn:TaskExecutionListEntry'newTaskExecutionListEntrytaskExecutionListEntry_status'taskExecutionListEntry_taskExecutionArn$fNFDataTaskExecutionListEntry $fHashableTaskExecutionListEntry $fFromJSONTaskExecutionListEntry$fEqTaskExecutionListEntry$fReadTaskExecutionListEntry$fShowTaskExecutionListEntry$fGenericTaskExecutionListEntryTaskFilterNameTaskFilterName'fromTaskFilterNameTaskFilterName_LocationIdTaskFilterName_CreationTime$fShowTaskFilterName$fReadTaskFilterName$fEqTaskFilterName$fOrdTaskFilterName$fGenericTaskFilterName$fHashableTaskFilterName$fNFDataTaskFilterName$fFromTextTaskFilterName$fToTextTaskFilterName$fToByteStringTaskFilterName$fToLogTaskFilterName$fToHeaderTaskFilterName$fToQueryTaskFilterName$fFromJSONTaskFilterName$fFromJSONKeyTaskFilterName$fToJSONTaskFilterName$fToJSONKeyTaskFilterName$fFromXMLTaskFilterName$fToXMLTaskFilterName TaskFilter TaskFilter'$sel:name:TaskFilter'$sel:values:TaskFilter'$sel:operator:TaskFilter' newTaskFiltertaskFilter_nametaskFilter_valuestaskFilter_operator$fToJSONTaskFilter$fNFDataTaskFilter$fHashableTaskFilter$fEqTaskFilter$fReadTaskFilter$fShowTaskFilter$fGenericTaskFilter TaskQueueing TaskQueueing'fromTaskQueueingTaskQueueing_ENABLEDTaskQueueing_DISABLED$fShowTaskQueueing$fReadTaskQueueing$fEqTaskQueueing$fOrdTaskQueueing$fGenericTaskQueueing$fHashableTaskQueueing$fNFDataTaskQueueing$fFromTextTaskQueueing$fToTextTaskQueueing$fToByteStringTaskQueueing$fToLogTaskQueueing$fToHeaderTaskQueueing$fToQueryTaskQueueing$fFromJSONTaskQueueing$fFromJSONKeyTaskQueueing$fToJSONTaskQueueing$fToJSONKeyTaskQueueing$fFromXMLTaskQueueing$fToXMLTaskQueueing TaskSchedule TaskSchedule'%$sel:scheduleExpression:TaskSchedule'newTaskScheduletaskSchedule_scheduleExpression$fToJSONTaskSchedule$fNFDataTaskSchedule$fHashableTaskSchedule$fFromJSONTaskSchedule$fEqTaskSchedule$fReadTaskSchedule$fShowTaskSchedule$fGenericTaskSchedule TaskStatus TaskStatus'fromTaskStatusTaskStatus_UNAVAILABLETaskStatus_RUNNINGTaskStatus_QUEUEDTaskStatus_CREATINGTaskStatus_AVAILABLE$fShowTaskStatus$fReadTaskStatus$fEqTaskStatus$fOrdTaskStatus$fGenericTaskStatus$fHashableTaskStatus$fNFDataTaskStatus$fFromTextTaskStatus$fToTextTaskStatus$fToByteStringTaskStatus$fToLogTaskStatus$fToHeaderTaskStatus$fToQueryTaskStatus$fFromJSONTaskStatus$fFromJSONKeyTaskStatus$fToJSONTaskStatus$fToJSONKeyTaskStatus$fFromXMLTaskStatus$fToXMLTaskStatus TaskListEntryTaskListEntry'$sel:name:TaskListEntry'$sel:status:TaskListEntry'$sel:taskArn:TaskListEntry'newTaskListEntrytaskListEntry_nametaskListEntry_statustaskListEntry_taskArn$fNFDataTaskListEntry$fHashableTaskListEntry$fFromJSONTaskListEntry$fEqTaskListEntry$fReadTaskListEntry$fShowTaskListEntry$fGenericTaskListEntry TransferMode TransferMode'fromTransferModeTransferMode_CHANGEDTransferMode_ALL$fShowTransferMode$fReadTransferMode$fEqTransferMode$fOrdTransferMode$fGenericTransferMode$fHashableTransferMode$fNFDataTransferMode$fFromTextTransferMode$fToTextTransferMode$fToByteStringTransferMode$fToLogTransferMode$fToHeaderTransferMode$fToQueryTransferMode$fFromJSONTransferMode$fFromJSONKeyTransferMode$fToJSONTransferMode$fToJSONKeyTransferMode$fFromXMLTransferMode$fToXMLTransferModeUidUid'fromUidUid_NONEUid_NAME Uid_INT_VALUEUid_BOTH $fShowUid $fReadUid$fEqUid$fOrdUid $fGenericUid $fHashableUid $fNFDataUid $fFromTextUid $fToTextUid$fToByteStringUid $fToLogUid $fToHeaderUid $fToQueryUid $fFromJSONUid$fFromJSONKeyUid $fToJSONUid$fToJSONKeyUid $fFromXMLUid $fToXMLUid VerifyMode VerifyMode'fromVerifyMode#VerifyMode_POINT_IN_TIME_CONSISTENT!VerifyMode_ONLY_FILES_TRANSFERREDVerifyMode_NONE$fShowVerifyMode$fReadVerifyMode$fEqVerifyMode$fOrdVerifyMode$fGenericVerifyMode$fHashableVerifyMode$fNFDataVerifyMode$fFromTextVerifyMode$fToTextVerifyMode$fToByteStringVerifyMode$fToLogVerifyMode$fToHeaderVerifyMode$fToQueryVerifyMode$fFromJSONVerifyMode$fFromJSONKeyVerifyMode$fToJSONVerifyMode$fToJSONKeyVerifyMode$fFromXMLVerifyMode$fToXMLVerifyModeOptionsOptions'$sel:atime:Options'$sel:bytesPerSecond:Options'$sel:gid:Options'$sel:logLevel:Options'$sel:mtime:Options'$sel:objectTags:Options'$sel:overwriteMode:Options'$sel:posixPermissions:Options'"$sel:preserveDeletedFiles:Options'$sel:preserveDevices:Options')$sel:securityDescriptorCopyFlags:Options'$sel:taskQueueing:Options'$sel:transferMode:Options'$sel:uid:Options'$sel:verifyMode:Options' newOptions options_atimeoptions_bytesPerSecond options_gidoptions_logLevel options_mtimeoptions_objectTagsoptions_overwriteModeoptions_posixPermissionsoptions_preserveDeletedFilesoptions_preserveDevices#options_securityDescriptorCopyFlagsoptions_taskQueueingoptions_transferMode options_uidoptions_verifyMode$fToJSONOptions$fNFDataOptions$fHashableOptions$fFromJSONOptions $fEqOptions $fReadOptions $fShowOptions$fGenericOptionsdefaultService_InternalException_InvalidRequestExceptionTagResourceResponseTagResourceResponse'$$sel:httpStatus:TagResourceResponse' TagResource TagResource'$sel:resourceArn:TagResource'$sel:tags:TagResource'newTagResourcetagResource_resourceArntagResource_tagsnewTagResourceResponsetagResourceResponse_httpStatus$fToQueryTagResource$fToPathTagResource$fToJSONTagResource$fToHeadersTagResource$fNFDataTagResource$fHashableTagResource$fNFDataTagResourceResponse$fAWSRequestTagResource$fEqTagResourceResponse$fReadTagResourceResponse$fShowTagResourceResponse$fGenericTagResourceResponse$fEqTagResource$fReadTagResource$fShowTagResource$fGenericTagResourceStartTaskExecutionResponseStartTaskExecutionResponse'1$sel:taskExecutionArn:StartTaskExecutionResponse'+$sel:httpStatus:StartTaskExecutionResponse'StartTaskExecutionStartTaskExecution'!$sel:excludes:StartTaskExecution'!$sel:includes:StartTaskExecution'($sel:overrideOptions:StartTaskExecution'$sel:tags:StartTaskExecution' $sel:taskArn:StartTaskExecution'newStartTaskExecutionstartTaskExecution_excludesstartTaskExecution_includes"startTaskExecution_overrideOptionsstartTaskExecution_tagsstartTaskExecution_taskArnnewStartTaskExecutionResponse+startTaskExecutionResponse_taskExecutionArn%startTaskExecutionResponse_httpStatus$fToQueryStartTaskExecution$fToPathStartTaskExecution$fToJSONStartTaskExecution$fToHeadersStartTaskExecution$fNFDataStartTaskExecution$fHashableStartTaskExecution"$fNFDataStartTaskExecutionResponse$fAWSRequestStartTaskExecution$fEqStartTaskExecutionResponse $fReadStartTaskExecutionResponse $fShowStartTaskExecutionResponse#$fGenericStartTaskExecutionResponse$fEqStartTaskExecution$fReadStartTaskExecution$fShowStartTaskExecution$fGenericStartTaskExecutionListTasksResponseListTasksResponse'!$sel:nextToken:ListTasksResponse'$sel:tasks:ListTasksResponse'"$sel:httpStatus:ListTasksResponse' ListTasks ListTasks'$sel:filters:ListTasks'$sel:maxResults:ListTasks'$sel:nextToken:ListTasks' newListTaskslistTasks_filterslistTasks_maxResultslistTasks_nextTokennewListTasksResponselistTasksResponse_nextTokenlistTasksResponse_taskslistTasksResponse_httpStatus$fToQueryListTasks$fToPathListTasks$fToJSONListTasks$fToHeadersListTasks$fNFDataListTasks$fHashableListTasks$fAWSPagerListTasks$fNFDataListTasksResponse$fAWSRequestListTasks$fEqListTasksResponse$fReadListTasksResponse$fShowListTasksResponse$fGenericListTasksResponse $fEqListTasks$fReadListTasks$fShowListTasks$fGenericListTasksListTaskExecutionsResponseListTaskExecutionsResponse'*$sel:nextToken:ListTaskExecutionsResponse'/$sel:taskExecutions:ListTaskExecutionsResponse'+$sel:httpStatus:ListTaskExecutionsResponse'ListTaskExecutionsListTaskExecutions'#$sel:maxResults:ListTaskExecutions'"$sel:nextToken:ListTaskExecutions' $sel:taskArn:ListTaskExecutions'newListTaskExecutionslistTaskExecutions_maxResultslistTaskExecutions_nextTokenlistTaskExecutions_taskArnnewListTaskExecutionsResponse$listTaskExecutionsResponse_nextToken)listTaskExecutionsResponse_taskExecutions%listTaskExecutionsResponse_httpStatus$fToQueryListTaskExecutions$fToPathListTaskExecutions$fToJSONListTaskExecutions$fToHeadersListTaskExecutions$fNFDataListTaskExecutions$fHashableListTaskExecutions$fAWSPagerListTaskExecutions"$fNFDataListTaskExecutionsResponse$fAWSRequestListTaskExecutions$fEqListTaskExecutionsResponse $fReadListTaskExecutionsResponse $fShowListTaskExecutionsResponse#$fGenericListTaskExecutionsResponse$fEqListTaskExecutions$fReadListTaskExecutions$fShowListTaskExecutions$fGenericListTaskExecutionsListTagsForResourceResponseListTagsForResourceResponse'+$sel:nextToken:ListTagsForResourceResponse'&$sel:tags:ListTagsForResourceResponse',$sel:httpStatus:ListTagsForResourceResponse'ListTagsForResourceListTagsForResource'$$sel:maxResults:ListTagsForResource'#$sel:nextToken:ListTagsForResource'%$sel:resourceArn:ListTagsForResource'newListTagsForResourcelistTagsForResource_maxResultslistTagsForResource_nextTokenlistTagsForResource_resourceArnnewListTagsForResourceResponse%listTagsForResourceResponse_nextToken listTagsForResourceResponse_tags&listTagsForResourceResponse_httpStatus$fToQueryListTagsForResource$fToPathListTagsForResource$fToJSONListTagsForResource$fToHeadersListTagsForResource$fNFDataListTagsForResource$fHashableListTagsForResource$fAWSPagerListTagsForResource#$fNFDataListTagsForResourceResponse$fAWSRequestListTagsForResource$fEqListTagsForResourceResponse!$fReadListTagsForResourceResponse!$fShowListTagsForResourceResponse$$fGenericListTagsForResourceResponse$fEqListTagsForResource$fReadListTagsForResource$fShowListTagsForResource$fGenericListTagsForResourceListLocationsResponseListLocationsResponse'%$sel:locations:ListLocationsResponse'%$sel:nextToken:ListLocationsResponse'&$sel:httpStatus:ListLocationsResponse' ListLocationsListLocations'$sel:filters:ListLocations'$sel:maxResults:ListLocations'$sel:nextToken:ListLocations'newListLocationslistLocations_filterslistLocations_maxResultslistLocations_nextTokennewListLocationsResponselistLocationsResponse_locationslistLocationsResponse_nextToken listLocationsResponse_httpStatus$fToQueryListLocations$fToPathListLocations$fToJSONListLocations$fToHeadersListLocations$fNFDataListLocations$fHashableListLocations$fAWSPagerListLocations$fNFDataListLocationsResponse$fAWSRequestListLocations$fEqListLocationsResponse$fReadListLocationsResponse$fShowListLocationsResponse$fGenericListLocationsResponse$fEqListLocations$fReadListLocations$fShowListLocations$fGenericListLocationsListAgentsResponseListAgentsResponse'$sel:agents:ListAgentsResponse'"$sel:nextToken:ListAgentsResponse'#$sel:httpStatus:ListAgentsResponse' ListAgents ListAgents'$sel:maxResults:ListAgents'$sel:nextToken:ListAgents' newListAgentslistAgents_maxResultslistAgents_nextTokennewListAgentsResponselistAgentsResponse_agentslistAgentsResponse_nextTokenlistAgentsResponse_httpStatus$fToQueryListAgents$fToPathListAgents$fToJSONListAgents$fToHeadersListAgents$fNFDataListAgents$fHashableListAgents$fAWSPagerListAgents$fNFDataListAgentsResponse$fAWSRequestListAgents$fEqListAgentsResponse$fReadListAgentsResponse$fShowListAgentsResponse$fGenericListAgentsResponse$fEqListAgents$fReadListAgents$fShowListAgents$fGenericListAgentsDescribeTaskExecutionResponseDescribeTaskExecutionResponse'3$sel:bytesCompressed:DescribeTaskExecutionResponse'4$sel:bytesTransferred:DescribeTaskExecutionResponse'0$sel:bytesWritten:DescribeTaskExecutionResponse'<$sel:estimatedBytesToTransfer:DescribeTaskExecutionResponse'<$sel:estimatedFilesToTransfer:DescribeTaskExecutionResponse',$sel:excludes:DescribeTaskExecutionResponse'4$sel:filesTransferred:DescribeTaskExecutionResponse',$sel:includes:DescribeTaskExecutionResponse'+$sel:options:DescribeTaskExecutionResponse'*$sel:result:DescribeTaskExecutionResponse'-$sel:startTime:DescribeTaskExecutionResponse'*$sel:status:DescribeTaskExecutionResponse'4$sel:taskExecutionArn:DescribeTaskExecutionResponse'.$sel:httpStatus:DescribeTaskExecutionResponse'DescribeTaskExecutionDescribeTaskExecution',$sel:taskExecutionArn:DescribeTaskExecution'newDescribeTaskExecution&describeTaskExecution_taskExecutionArn newDescribeTaskExecutionResponse-describeTaskExecutionResponse_bytesCompressed.describeTaskExecutionResponse_bytesTransferred*describeTaskExecutionResponse_bytesWritten6describeTaskExecutionResponse_estimatedBytesToTransfer6describeTaskExecutionResponse_estimatedFilesToTransfer&describeTaskExecutionResponse_excludes.describeTaskExecutionResponse_filesTransferred&describeTaskExecutionResponse_includes%describeTaskExecutionResponse_options$describeTaskExecutionResponse_result'describeTaskExecutionResponse_startTime$describeTaskExecutionResponse_status.describeTaskExecutionResponse_taskExecutionArn(describeTaskExecutionResponse_httpStatus$fToQueryDescribeTaskExecution$fToPathDescribeTaskExecution$fToJSONDescribeTaskExecution $fToHeadersDescribeTaskExecution$fNFDataDescribeTaskExecution$fHashableDescribeTaskExecution%$fNFDataDescribeTaskExecutionResponse!$fAWSRequestDescribeTaskExecution!$fEqDescribeTaskExecutionResponse#$fReadDescribeTaskExecutionResponse#$fShowDescribeTaskExecutionResponse&$fGenericDescribeTaskExecutionResponse$fEqDescribeTaskExecution$fReadDescribeTaskExecution$fShowDescribeTaskExecution$fGenericDescribeTaskExecutionDescribeTaskResponseDescribeTaskResponse'0$sel:cloudWatchLogGroupArn:DescribeTaskResponse''$sel:creationTime:DescribeTaskResponse'2$sel:currentTaskExecutionArn:DescribeTaskResponse'1$sel:destinationLocationArn:DescribeTaskResponse':$sel:destinationNetworkInterfaceArns:DescribeTaskResponse'$$sel:errorCode:DescribeTaskResponse'&$sel:errorDetail:DescribeTaskResponse'#$sel:excludes:DescribeTaskResponse'#$sel:includes:DescribeTaskResponse'$sel:name:DescribeTaskResponse'"$sel:options:DescribeTaskResponse'#$sel:schedule:DescribeTaskResponse',$sel:sourceLocationArn:DescribeTaskResponse'5$sel:sourceNetworkInterfaceArns:DescribeTaskResponse'!$sel:status:DescribeTaskResponse'"$sel:taskArn:DescribeTaskResponse'%$sel:httpStatus:DescribeTaskResponse' DescribeTask DescribeTask'$sel:taskArn:DescribeTask'newDescribeTaskdescribeTask_taskArnnewDescribeTaskResponse*describeTaskResponse_cloudWatchLogGroupArn!describeTaskResponse_creationTime,describeTaskResponse_currentTaskExecutionArn+describeTaskResponse_destinationLocationArn4describeTaskResponse_destinationNetworkInterfaceArnsdescribeTaskResponse_errorCode describeTaskResponse_errorDetaildescribeTaskResponse_excludesdescribeTaskResponse_includesdescribeTaskResponse_namedescribeTaskResponse_optionsdescribeTaskResponse_schedule&describeTaskResponse_sourceLocationArn/describeTaskResponse_sourceNetworkInterfaceArnsdescribeTaskResponse_statusdescribeTaskResponse_taskArndescribeTaskResponse_httpStatus$fToQueryDescribeTask$fToPathDescribeTask$fToJSONDescribeTask$fToHeadersDescribeTask$fNFDataDescribeTask$fHashableDescribeTask$fNFDataDescribeTaskResponse$fAWSRequestDescribeTask$fEqDescribeTaskResponse$fReadDescribeTaskResponse$fShowDescribeTaskResponse$fGenericDescribeTaskResponse$fEqDescribeTask$fReadDescribeTask$fShowDescribeTask$fGenericDescribeTaskDescribeLocationSmbResponseDescribeLocationSmbResponse'+$sel:agentArns:DescribeLocationSmbResponse'.$sel:creationTime:DescribeLocationSmbResponse'($sel:domain:DescribeLocationSmbResponse'-$sel:locationArn:DescribeLocationSmbResponse'-$sel:locationUri:DescribeLocationSmbResponse'.$sel:mountOptions:DescribeLocationSmbResponse'&$sel:user:DescribeLocationSmbResponse',$sel:httpStatus:DescribeLocationSmbResponse'DescribeLocationSmbDescribeLocationSmb'%$sel:locationArn:DescribeLocationSmb'newDescribeLocationSmbdescribeLocationSmb_locationArnnewDescribeLocationSmbResponse%describeLocationSmbResponse_agentArns(describeLocationSmbResponse_creationTime"describeLocationSmbResponse_domain'describeLocationSmbResponse_locationArn'describeLocationSmbResponse_locationUri(describeLocationSmbResponse_mountOptions describeLocationSmbResponse_user&describeLocationSmbResponse_httpStatus$fToQueryDescribeLocationSmb$fToPathDescribeLocationSmb$fToJSONDescribeLocationSmb$fToHeadersDescribeLocationSmb$fNFDataDescribeLocationSmb$fHashableDescribeLocationSmb#$fNFDataDescribeLocationSmbResponse$fAWSRequestDescribeLocationSmb$fEqDescribeLocationSmbResponse!$fReadDescribeLocationSmbResponse!$fShowDescribeLocationSmbResponse$$fGenericDescribeLocationSmbResponse$fEqDescribeLocationSmb$fReadDescribeLocationSmb$fShowDescribeLocationSmb$fGenericDescribeLocationSmbDescribeLocationS3ResponseDescribeLocationS3Response'*$sel:agentArns:DescribeLocationS3Response'-$sel:creationTime:DescribeLocationS3Response',$sel:locationArn:DescribeLocationS3Response',$sel:locationUri:DescribeLocationS3Response')$sel:s3Config:DescribeLocationS3Response'/$sel:s3StorageClass:DescribeLocationS3Response'+$sel:httpStatus:DescribeLocationS3Response'DescribeLocationS3DescribeLocationS3'$$sel:locationArn:DescribeLocationS3'newDescribeLocationS3describeLocationS3_locationArnnewDescribeLocationS3Response$describeLocationS3Response_agentArns'describeLocationS3Response_creationTime&describeLocationS3Response_locationArn&describeLocationS3Response_locationUri#describeLocationS3Response_s3Config)describeLocationS3Response_s3StorageClass%describeLocationS3Response_httpStatus$fToQueryDescribeLocationS3$fToPathDescribeLocationS3$fToJSONDescribeLocationS3$fToHeadersDescribeLocationS3$fNFDataDescribeLocationS3$fHashableDescribeLocationS3"$fNFDataDescribeLocationS3Response$fAWSRequestDescribeLocationS3$fEqDescribeLocationS3Response $fReadDescribeLocationS3Response $fShowDescribeLocationS3Response#$fGenericDescribeLocationS3Response$fEqDescribeLocationS3$fReadDescribeLocationS3$fShowDescribeLocationS3$fGenericDescribeLocationS3%DescribeLocationObjectStorageResponse&DescribeLocationObjectStorageResponse'5$sel:accessKey:DescribeLocationObjectStorageResponse'5$sel:agentArns:DescribeLocationObjectStorageResponse'8$sel:creationTime:DescribeLocationObjectStorageResponse'7$sel:locationArn:DescribeLocationObjectStorageResponse'7$sel:locationUri:DescribeLocationObjectStorageResponse'=$sel:serverCertificate:DescribeLocationObjectStorageResponse'6$sel:serverPort:DescribeLocationObjectStorageResponse':$sel:serverProtocol:DescribeLocationObjectStorageResponse'6$sel:httpStatus:DescribeLocationObjectStorageResponse'DescribeLocationObjectStorageDescribeLocationObjectStorage'/$sel:locationArn:DescribeLocationObjectStorage' newDescribeLocationObjectStorage)describeLocationObjectStorage_locationArn(newDescribeLocationObjectStorageResponse/describeLocationObjectStorageResponse_accessKey/describeLocationObjectStorageResponse_agentArns2describeLocationObjectStorageResponse_creationTime1describeLocationObjectStorageResponse_locationArn1describeLocationObjectStorageResponse_locationUri7describeLocationObjectStorageResponse_serverCertificate0describeLocationObjectStorageResponse_serverPort4describeLocationObjectStorageResponse_serverProtocol0describeLocationObjectStorageResponse_httpStatus&$fToQueryDescribeLocationObjectStorage%$fToPathDescribeLocationObjectStorage%$fToJSONDescribeLocationObjectStorage($fToHeadersDescribeLocationObjectStorage%$fNFDataDescribeLocationObjectStorage'$fHashableDescribeLocationObjectStorage-$fNFDataDescribeLocationObjectStorageResponse)$fAWSRequestDescribeLocationObjectStorage)$fEqDescribeLocationObjectStorageResponse+$fReadDescribeLocationObjectStorageResponse+$fShowDescribeLocationObjectStorageResponse.$fGenericDescribeLocationObjectStorageResponse!$fEqDescribeLocationObjectStorage#$fReadDescribeLocationObjectStorage#$fShowDescribeLocationObjectStorage&$fGenericDescribeLocationObjectStorageDescribeLocationNfsResponseDescribeLocationNfsResponse'.$sel:creationTime:DescribeLocationNfsResponse'-$sel:locationArn:DescribeLocationNfsResponse'-$sel:locationUri:DescribeLocationNfsResponse'.$sel:mountOptions:DescribeLocationNfsResponse'.$sel:onPremConfig:DescribeLocationNfsResponse',$sel:httpStatus:DescribeLocationNfsResponse'DescribeLocationNfsDescribeLocationNfs'%$sel:locationArn:DescribeLocationNfs'newDescribeLocationNfsdescribeLocationNfs_locationArnnewDescribeLocationNfsResponse(describeLocationNfsResponse_creationTime'describeLocationNfsResponse_locationArn'describeLocationNfsResponse_locationUri(describeLocationNfsResponse_mountOptions(describeLocationNfsResponse_onPremConfig&describeLocationNfsResponse_httpStatus$fToQueryDescribeLocationNfs$fToPathDescribeLocationNfs$fToJSONDescribeLocationNfs$fToHeadersDescribeLocationNfs$fNFDataDescribeLocationNfs$fHashableDescribeLocationNfs#$fNFDataDescribeLocationNfsResponse$fAWSRequestDescribeLocationNfs$fEqDescribeLocationNfsResponse!$fReadDescribeLocationNfsResponse!$fShowDescribeLocationNfsResponse$$fGenericDescribeLocationNfsResponse$fEqDescribeLocationNfs$fReadDescribeLocationNfs$fShowDescribeLocationNfs$fGenericDescribeLocationNfsDescribeLocationHdfsResponseDescribeLocationHdfsResponse',$sel:agentArns:DescribeLocationHdfsResponse'5$sel:authenticationType:DescribeLocationHdfsResponse',$sel:blockSize:DescribeLocationHdfsResponse'/$sel:creationTime:DescribeLocationHdfsResponse'4$sel:kerberosPrincipal:DescribeLocationHdfsResponse'4$sel:kmsKeyProviderUri:DescribeLocationHdfsResponse'.$sel:locationArn:DescribeLocationHdfsResponse'.$sel:locationUri:DescribeLocationHdfsResponse',$sel:nameNodes:DescribeLocationHdfsResponse'3$sel:qopConfiguration:DescribeLocationHdfsResponse'4$sel:replicationFactor:DescribeLocationHdfsResponse'-$sel:simpleUser:DescribeLocationHdfsResponse'-$sel:httpStatus:DescribeLocationHdfsResponse'DescribeLocationHdfsDescribeLocationHdfs'&$sel:locationArn:DescribeLocationHdfs'newDescribeLocationHdfs describeLocationHdfs_locationArnnewDescribeLocationHdfsResponse&describeLocationHdfsResponse_agentArns/describeLocationHdfsResponse_authenticationType&describeLocationHdfsResponse_blockSize)describeLocationHdfsResponse_creationTime.describeLocationHdfsResponse_kerberosPrincipal.describeLocationHdfsResponse_kmsKeyProviderUri(describeLocationHdfsResponse_locationArn(describeLocationHdfsResponse_locationUri&describeLocationHdfsResponse_nameNodes-describeLocationHdfsResponse_qopConfiguration.describeLocationHdfsResponse_replicationFactor'describeLocationHdfsResponse_simpleUser'describeLocationHdfsResponse_httpStatus$fToQueryDescribeLocationHdfs$fToPathDescribeLocationHdfs$fToJSONDescribeLocationHdfs$fToHeadersDescribeLocationHdfs$fNFDataDescribeLocationHdfs$fHashableDescribeLocationHdfs$$fNFDataDescribeLocationHdfsResponse $fAWSRequestDescribeLocationHdfs $fEqDescribeLocationHdfsResponse"$fReadDescribeLocationHdfsResponse"$fShowDescribeLocationHdfsResponse%$fGenericDescribeLocationHdfsResponse$fEqDescribeLocationHdfs$fReadDescribeLocationHdfs$fShowDescribeLocationHdfs$fGenericDescribeLocationHdfs"DescribeLocationFsxWindowsResponse#DescribeLocationFsxWindowsResponse'5$sel:creationTime:DescribeLocationFsxWindowsResponse'/$sel:domain:DescribeLocationFsxWindowsResponse'4$sel:locationArn:DescribeLocationFsxWindowsResponse'4$sel:locationUri:DescribeLocationFsxWindowsResponse':$sel:securityGroupArns:DescribeLocationFsxWindowsResponse'-$sel:user:DescribeLocationFsxWindowsResponse'3$sel:httpStatus:DescribeLocationFsxWindowsResponse'DescribeLocationFsxWindowsDescribeLocationFsxWindows',$sel:locationArn:DescribeLocationFsxWindows'newDescribeLocationFsxWindows&describeLocationFsxWindows_locationArn%newDescribeLocationFsxWindowsResponse/describeLocationFsxWindowsResponse_creationTime)describeLocationFsxWindowsResponse_domain.describeLocationFsxWindowsResponse_locationArn.describeLocationFsxWindowsResponse_locationUri4describeLocationFsxWindowsResponse_securityGroupArns'describeLocationFsxWindowsResponse_user-describeLocationFsxWindowsResponse_httpStatus#$fToQueryDescribeLocationFsxWindows"$fToPathDescribeLocationFsxWindows"$fToJSONDescribeLocationFsxWindows%$fToHeadersDescribeLocationFsxWindows"$fNFDataDescribeLocationFsxWindows$$fHashableDescribeLocationFsxWindows*$fNFDataDescribeLocationFsxWindowsResponse&$fAWSRequestDescribeLocationFsxWindows&$fEqDescribeLocationFsxWindowsResponse($fReadDescribeLocationFsxWindowsResponse($fShowDescribeLocationFsxWindowsResponse+$fGenericDescribeLocationFsxWindowsResponse$fEqDescribeLocationFsxWindows $fReadDescribeLocationFsxWindows $fShowDescribeLocationFsxWindows#$fGenericDescribeLocationFsxWindows"DescribeLocationFsxOpenZfsResponse#DescribeLocationFsxOpenZfsResponse'5$sel:creationTime:DescribeLocationFsxOpenZfsResponse'4$sel:locationArn:DescribeLocationFsxOpenZfsResponse'4$sel:locationUri:DescribeLocationFsxOpenZfsResponse'1$sel:protocol:DescribeLocationFsxOpenZfsResponse':$sel:securityGroupArns:DescribeLocationFsxOpenZfsResponse'3$sel:httpStatus:DescribeLocationFsxOpenZfsResponse'DescribeLocationFsxOpenZfsDescribeLocationFsxOpenZfs',$sel:locationArn:DescribeLocationFsxOpenZfs'newDescribeLocationFsxOpenZfs&describeLocationFsxOpenZfs_locationArn%newDescribeLocationFsxOpenZfsResponse/describeLocationFsxOpenZfsResponse_creationTime.describeLocationFsxOpenZfsResponse_locationArn.describeLocationFsxOpenZfsResponse_locationUri+describeLocationFsxOpenZfsResponse_protocol4describeLocationFsxOpenZfsResponse_securityGroupArns-describeLocationFsxOpenZfsResponse_httpStatus#$fToQueryDescribeLocationFsxOpenZfs"$fToPathDescribeLocationFsxOpenZfs"$fToJSONDescribeLocationFsxOpenZfs%$fToHeadersDescribeLocationFsxOpenZfs"$fNFDataDescribeLocationFsxOpenZfs$$fHashableDescribeLocationFsxOpenZfs*$fNFDataDescribeLocationFsxOpenZfsResponse&$fAWSRequestDescribeLocationFsxOpenZfs&$fEqDescribeLocationFsxOpenZfsResponse($fShowDescribeLocationFsxOpenZfsResponse+$fGenericDescribeLocationFsxOpenZfsResponse$fEqDescribeLocationFsxOpenZfs $fReadDescribeLocationFsxOpenZfs $fShowDescribeLocationFsxOpenZfs#$fGenericDescribeLocationFsxOpenZfs DescribeLocationFsxOntapResponse!DescribeLocationFsxOntapResponse'3$sel:creationTime:DescribeLocationFsxOntapResponse'7$sel:fsxFilesystemArn:DescribeLocationFsxOntapResponse'2$sel:locationArn:DescribeLocationFsxOntapResponse'2$sel:locationUri:DescribeLocationFsxOntapResponse'/$sel:protocol:DescribeLocationFsxOntapResponse'8$sel:securityGroupArns:DescribeLocationFsxOntapResponse'?$sel:storageVirtualMachineArn:DescribeLocationFsxOntapResponse'1$sel:httpStatus:DescribeLocationFsxOntapResponse'DescribeLocationFsxOntapDescribeLocationFsxOntap'*$sel:locationArn:DescribeLocationFsxOntap'newDescribeLocationFsxOntap$describeLocationFsxOntap_locationArn#newDescribeLocationFsxOntapResponse-describeLocationFsxOntapResponse_creationTime1describeLocationFsxOntapResponse_fsxFilesystemArn,describeLocationFsxOntapResponse_locationArn,describeLocationFsxOntapResponse_locationUri)describeLocationFsxOntapResponse_protocol2describeLocationFsxOntapResponse_securityGroupArns9describeLocationFsxOntapResponse_storageVirtualMachineArn+describeLocationFsxOntapResponse_httpStatus!$fToQueryDescribeLocationFsxOntap $fToPathDescribeLocationFsxOntap $fToJSONDescribeLocationFsxOntap#$fToHeadersDescribeLocationFsxOntap $fNFDataDescribeLocationFsxOntap"$fHashableDescribeLocationFsxOntap($fNFDataDescribeLocationFsxOntapResponse$$fAWSRequestDescribeLocationFsxOntap$$fEqDescribeLocationFsxOntapResponse&$fShowDescribeLocationFsxOntapResponse)$fGenericDescribeLocationFsxOntapResponse$fEqDescribeLocationFsxOntap$fReadDescribeLocationFsxOntap$fShowDescribeLocationFsxOntap!$fGenericDescribeLocationFsxOntap!DescribeLocationFsxLustreResponse"DescribeLocationFsxLustreResponse'4$sel:creationTime:DescribeLocationFsxLustreResponse'3$sel:locationArn:DescribeLocationFsxLustreResponse'3$sel:locationUri:DescribeLocationFsxLustreResponse'9$sel:securityGroupArns:DescribeLocationFsxLustreResponse'2$sel:httpStatus:DescribeLocationFsxLustreResponse'DescribeLocationFsxLustreDescribeLocationFsxLustre'+$sel:locationArn:DescribeLocationFsxLustre'newDescribeLocationFsxLustre%describeLocationFsxLustre_locationArn$newDescribeLocationFsxLustreResponse.describeLocationFsxLustreResponse_creationTime-describeLocationFsxLustreResponse_locationArn-describeLocationFsxLustreResponse_locationUri3describeLocationFsxLustreResponse_securityGroupArns,describeLocationFsxLustreResponse_httpStatus"$fToQueryDescribeLocationFsxLustre!$fToPathDescribeLocationFsxLustre!$fToJSONDescribeLocationFsxLustre$$fToHeadersDescribeLocationFsxLustre!$fNFDataDescribeLocationFsxLustre#$fHashableDescribeLocationFsxLustre)$fNFDataDescribeLocationFsxLustreResponse%$fAWSRequestDescribeLocationFsxLustre%$fEqDescribeLocationFsxLustreResponse'$fReadDescribeLocationFsxLustreResponse'$fShowDescribeLocationFsxLustreResponse*$fGenericDescribeLocationFsxLustreResponse$fEqDescribeLocationFsxLustre$fReadDescribeLocationFsxLustre$fShowDescribeLocationFsxLustre"$fGenericDescribeLocationFsxLustreDescribeLocationEfsResponseDescribeLocationEfsResponse'0$sel:accessPointArn:DescribeLocationEfsResponse'.$sel:creationTime:DescribeLocationEfsResponse'+$sel:ec2Config:DescribeLocationEfsResponse'9$sel:fileSystemAccessRoleArn:DescribeLocationEfsResponse'5$sel:inTransitEncryption:DescribeLocationEfsResponse'-$sel:locationArn:DescribeLocationEfsResponse'-$sel:locationUri:DescribeLocationEfsResponse',$sel:httpStatus:DescribeLocationEfsResponse'DescribeLocationEfsDescribeLocationEfs'%$sel:locationArn:DescribeLocationEfs'newDescribeLocationEfsdescribeLocationEfs_locationArnnewDescribeLocationEfsResponse*describeLocationEfsResponse_accessPointArn(describeLocationEfsResponse_creationTime%describeLocationEfsResponse_ec2Config3describeLocationEfsResponse_fileSystemAccessRoleArn/describeLocationEfsResponse_inTransitEncryption'describeLocationEfsResponse_locationArn'describeLocationEfsResponse_locationUri&describeLocationEfsResponse_httpStatus$fToQueryDescribeLocationEfs$fToPathDescribeLocationEfs$fToJSONDescribeLocationEfs$fToHeadersDescribeLocationEfs$fNFDataDescribeLocationEfs$fHashableDescribeLocationEfs#$fNFDataDescribeLocationEfsResponse$fAWSRequestDescribeLocationEfs$fEqDescribeLocationEfsResponse!$fReadDescribeLocationEfsResponse!$fShowDescribeLocationEfsResponse$$fGenericDescribeLocationEfsResponse$fEqDescribeLocationEfs$fReadDescribeLocationEfs$fShowDescribeLocationEfs$fGenericDescribeLocationEfsDescribeAgentResponseDescribeAgentResponse'$$sel:agentArn:DescribeAgentResponse'($sel:creationTime:DescribeAgentResponse'($sel:endpointType:DescribeAgentResponse'.$sel:lastConnectionTime:DescribeAgentResponse' $sel:name:DescribeAgentResponse'-$sel:privateLinkConfig:DescribeAgentResponse'"$sel:status:DescribeAgentResponse'&$sel:httpStatus:DescribeAgentResponse' DescribeAgentDescribeAgent'$sel:agentArn:DescribeAgent'newDescribeAgentdescribeAgent_agentArnnewDescribeAgentResponsedescribeAgentResponse_agentArn"describeAgentResponse_creationTime"describeAgentResponse_endpointType(describeAgentResponse_lastConnectionTimedescribeAgentResponse_name'describeAgentResponse_privateLinkConfigdescribeAgentResponse_status describeAgentResponse_httpStatus$fToQueryDescribeAgent$fToPathDescribeAgent$fToJSONDescribeAgent$fToHeadersDescribeAgent$fNFDataDescribeAgent$fHashableDescribeAgent$fNFDataDescribeAgentResponse$fAWSRequestDescribeAgent$fEqDescribeAgentResponse$fReadDescribeAgentResponse$fShowDescribeAgentResponse$fGenericDescribeAgentResponse$fEqDescribeAgent$fReadDescribeAgent$fShowDescribeAgent$fGenericDescribeAgentDeleteTaskResponseDeleteTaskResponse'#$sel:httpStatus:DeleteTaskResponse' DeleteTask DeleteTask'$sel:taskArn:DeleteTask' newDeleteTaskdeleteTask_taskArnnewDeleteTaskResponsedeleteTaskResponse_httpStatus$fToQueryDeleteTask$fToPathDeleteTask$fToJSONDeleteTask$fToHeadersDeleteTask$fNFDataDeleteTask$fHashableDeleteTask$fNFDataDeleteTaskResponse$fAWSRequestDeleteTask$fEqDeleteTaskResponse$fReadDeleteTaskResponse$fShowDeleteTaskResponse$fGenericDeleteTaskResponse$fEqDeleteTask$fReadDeleteTask$fShowDeleteTask$fGenericDeleteTaskDeleteLocationResponseDeleteLocationResponse''$sel:httpStatus:DeleteLocationResponse'DeleteLocationDeleteLocation' $sel:locationArn:DeleteLocation'newDeleteLocationdeleteLocation_locationArnnewDeleteLocationResponse!deleteLocationResponse_httpStatus$fToQueryDeleteLocation$fToPathDeleteLocation$fToJSONDeleteLocation$fToHeadersDeleteLocation$fNFDataDeleteLocation$fHashableDeleteLocation$fNFDataDeleteLocationResponse$fAWSRequestDeleteLocation$fEqDeleteLocationResponse$fReadDeleteLocationResponse$fShowDeleteLocationResponse$fGenericDeleteLocationResponse$fEqDeleteLocation$fReadDeleteLocation$fShowDeleteLocation$fGenericDeleteLocationDeleteAgentResponseDeleteAgentResponse'$$sel:httpStatus:DeleteAgentResponse' DeleteAgent DeleteAgent'$sel:agentArn:DeleteAgent'newDeleteAgentdeleteAgent_agentArnnewDeleteAgentResponsedeleteAgentResponse_httpStatus$fToQueryDeleteAgent$fToPathDeleteAgent$fToJSONDeleteAgent$fToHeadersDeleteAgent$fNFDataDeleteAgent$fHashableDeleteAgent$fNFDataDeleteAgentResponse$fAWSRequestDeleteAgent$fEqDeleteAgentResponse$fReadDeleteAgentResponse$fShowDeleteAgentResponse$fGenericDeleteAgentResponse$fEqDeleteAgent$fReadDeleteAgent$fShowDeleteAgent$fGenericDeleteAgentCreateTaskResponseCreateTaskResponse' $sel:taskArn:CreateTaskResponse'#$sel:httpStatus:CreateTaskResponse' CreateTask CreateTask'&$sel:cloudWatchLogGroupArn:CreateTask'$sel:excludes:CreateTask'$sel:includes:CreateTask'$sel:name:CreateTask'$sel:options:CreateTask'$sel:schedule:CreateTask'$sel:tags:CreateTask'"$sel:sourceLocationArn:CreateTask''$sel:destinationLocationArn:CreateTask' newCreateTask createTask_cloudWatchLogGroupArncreateTask_excludescreateTask_includescreateTask_namecreateTask_optionscreateTask_schedulecreateTask_tagscreateTask_sourceLocationArn!createTask_destinationLocationArnnewCreateTaskResponsecreateTaskResponse_taskArncreateTaskResponse_httpStatus$fToQueryCreateTask$fToPathCreateTask$fToJSONCreateTask$fToHeadersCreateTask$fNFDataCreateTask$fHashableCreateTask$fNFDataCreateTaskResponse$fAWSRequestCreateTask$fEqCreateTaskResponse$fReadCreateTaskResponse$fShowCreateTaskResponse$fGenericCreateTaskResponse$fEqCreateTask$fReadCreateTask$fShowCreateTask$fGenericCreateTaskCreateLocationSmbResponseCreateLocationSmbResponse'+$sel:locationArn:CreateLocationSmbResponse'*$sel:httpStatus:CreateLocationSmbResponse'CreateLocationSmbCreateLocationSmb'$sel:domain:CreateLocationSmb'$$sel:mountOptions:CreateLocationSmb'$sel:tags:CreateLocationSmb'$$sel:subdirectory:CreateLocationSmb'&$sel:serverHostname:CreateLocationSmb'$sel:user:CreateLocationSmb' $sel:password:CreateLocationSmb'!$sel:agentArns:CreateLocationSmb'newCreateLocationSmbcreateLocationSmb_domaincreateLocationSmb_mountOptionscreateLocationSmb_tagscreateLocationSmb_subdirectory createLocationSmb_serverHostnamecreateLocationSmb_usercreateLocationSmb_passwordcreateLocationSmb_agentArnsnewCreateLocationSmbResponse%createLocationSmbResponse_locationArn$createLocationSmbResponse_httpStatus$fToQueryCreateLocationSmb$fToPathCreateLocationSmb$fToJSONCreateLocationSmb$fToHeadersCreateLocationSmb$fNFDataCreateLocationSmb$fHashableCreateLocationSmb!$fNFDataCreateLocationSmbResponse$fAWSRequestCreateLocationSmb$fEqCreateLocationSmbResponse$fReadCreateLocationSmbResponse$fShowCreateLocationSmbResponse"$fGenericCreateLocationSmbResponse$fEqCreateLocationSmb$fShowCreateLocationSmb$fGenericCreateLocationSmbCreateLocationS3ResponseCreateLocationS3Response'*$sel:locationArn:CreateLocationS3Response')$sel:httpStatus:CreateLocationS3Response'CreateLocationS3CreateLocationS3' $sel:agentArns:CreateLocationS3'%$sel:s3StorageClass:CreateLocationS3'#$sel:subdirectory:CreateLocationS3'$sel:tags:CreateLocationS3'"$sel:s3BucketArn:CreateLocationS3'$sel:s3Config:CreateLocationS3'newCreateLocationS3createLocationS3_agentArnscreateLocationS3_s3StorageClasscreateLocationS3_subdirectorycreateLocationS3_tagscreateLocationS3_s3BucketArncreateLocationS3_s3ConfignewCreateLocationS3Response$createLocationS3Response_locationArn#createLocationS3Response_httpStatus$fToQueryCreateLocationS3$fToPathCreateLocationS3$fToJSONCreateLocationS3$fToHeadersCreateLocationS3$fNFDataCreateLocationS3$fHashableCreateLocationS3 $fNFDataCreateLocationS3Response$fAWSRequestCreateLocationS3$fEqCreateLocationS3Response$fReadCreateLocationS3Response$fShowCreateLocationS3Response!$fGenericCreateLocationS3Response$fEqCreateLocationS3$fReadCreateLocationS3$fShowCreateLocationS3$fGenericCreateLocationS3#CreateLocationObjectStorageResponse$CreateLocationObjectStorageResponse'5$sel:locationArn:CreateLocationObjectStorageResponse'4$sel:httpStatus:CreateLocationObjectStorageResponse'CreateLocationObjectStorageCreateLocationObjectStorage'+$sel:accessKey:CreateLocationObjectStorage'+$sel:secretKey:CreateLocationObjectStorage'3$sel:serverCertificate:CreateLocationObjectStorage',$sel:serverPort:CreateLocationObjectStorage'0$sel:serverProtocol:CreateLocationObjectStorage'.$sel:subdirectory:CreateLocationObjectStorage'&$sel:tags:CreateLocationObjectStorage'0$sel:serverHostname:CreateLocationObjectStorage',$sel:bucketName:CreateLocationObjectStorage'+$sel:agentArns:CreateLocationObjectStorage'newCreateLocationObjectStorage%createLocationObjectStorage_accessKey%createLocationObjectStorage_secretKey-createLocationObjectStorage_serverCertificate&createLocationObjectStorage_serverPort*createLocationObjectStorage_serverProtocol(createLocationObjectStorage_subdirectory createLocationObjectStorage_tags*createLocationObjectStorage_serverHostname&createLocationObjectStorage_bucketName%createLocationObjectStorage_agentArns&newCreateLocationObjectStorageResponse/createLocationObjectStorageResponse_locationArn.createLocationObjectStorageResponse_httpStatus$$fToQueryCreateLocationObjectStorage#$fToPathCreateLocationObjectStorage#$fToJSONCreateLocationObjectStorage&$fToHeadersCreateLocationObjectStorage#$fNFDataCreateLocationObjectStorage%$fHashableCreateLocationObjectStorage+$fNFDataCreateLocationObjectStorageResponse'$fAWSRequestCreateLocationObjectStorage'$fEqCreateLocationObjectStorageResponse)$fReadCreateLocationObjectStorageResponse)$fShowCreateLocationObjectStorageResponse,$fGenericCreateLocationObjectStorageResponse$fEqCreateLocationObjectStorage!$fShowCreateLocationObjectStorage$$fGenericCreateLocationObjectStorageCreateLocationNfsResponseCreateLocationNfsResponse'+$sel:locationArn:CreateLocationNfsResponse'*$sel:httpStatus:CreateLocationNfsResponse'CreateLocationNfsCreateLocationNfs'$$sel:mountOptions:CreateLocationNfs'$sel:tags:CreateLocationNfs'$$sel:subdirectory:CreateLocationNfs'&$sel:serverHostname:CreateLocationNfs'$$sel:onPremConfig:CreateLocationNfs'newCreateLocationNfscreateLocationNfs_mountOptionscreateLocationNfs_tagscreateLocationNfs_subdirectory createLocationNfs_serverHostnamecreateLocationNfs_onPremConfignewCreateLocationNfsResponse%createLocationNfsResponse_locationArn$createLocationNfsResponse_httpStatus$fToQueryCreateLocationNfs$fToPathCreateLocationNfs$fToJSONCreateLocationNfs$fToHeadersCreateLocationNfs$fNFDataCreateLocationNfs$fHashableCreateLocationNfs!$fNFDataCreateLocationNfsResponse$fAWSRequestCreateLocationNfs$fEqCreateLocationNfsResponse$fReadCreateLocationNfsResponse$fShowCreateLocationNfsResponse"$fGenericCreateLocationNfsResponse$fEqCreateLocationNfs$fReadCreateLocationNfs$fShowCreateLocationNfs$fGenericCreateLocationNfsCreateLocationHdfsResponseCreateLocationHdfsResponse',$sel:locationArn:CreateLocationHdfsResponse'+$sel:httpStatus:CreateLocationHdfsResponse'CreateLocationHdfsCreateLocationHdfs'"$sel:blockSize:CreateLocationHdfs''$sel:kerberosKeytab:CreateLocationHdfs')$sel:kerberosKrb5Conf:CreateLocationHdfs'*$sel:kerberosPrincipal:CreateLocationHdfs'*$sel:kmsKeyProviderUri:CreateLocationHdfs')$sel:qopConfiguration:CreateLocationHdfs'*$sel:replicationFactor:CreateLocationHdfs'#$sel:simpleUser:CreateLocationHdfs'%$sel:subdirectory:CreateLocationHdfs'$sel:tags:CreateLocationHdfs'"$sel:nameNodes:CreateLocationHdfs'+$sel:authenticationType:CreateLocationHdfs'"$sel:agentArns:CreateLocationHdfs'newCreateLocationHdfscreateLocationHdfs_blockSize!createLocationHdfs_kerberosKeytab#createLocationHdfs_kerberosKrb5Conf$createLocationHdfs_kerberosPrincipal$createLocationHdfs_kmsKeyProviderUri#createLocationHdfs_qopConfiguration$createLocationHdfs_replicationFactorcreateLocationHdfs_simpleUsercreateLocationHdfs_subdirectorycreateLocationHdfs_tagscreateLocationHdfs_nameNodes%createLocationHdfs_authenticationTypecreateLocationHdfs_agentArnsnewCreateLocationHdfsResponse&createLocationHdfsResponse_locationArn%createLocationHdfsResponse_httpStatus$fToQueryCreateLocationHdfs$fToPathCreateLocationHdfs$fToJSONCreateLocationHdfs$fToHeadersCreateLocationHdfs$fNFDataCreateLocationHdfs$fHashableCreateLocationHdfs"$fNFDataCreateLocationHdfsResponse$fAWSRequestCreateLocationHdfs$fEqCreateLocationHdfsResponse $fReadCreateLocationHdfsResponse $fShowCreateLocationHdfsResponse#$fGenericCreateLocationHdfsResponse$fEqCreateLocationHdfs$fReadCreateLocationHdfs$fShowCreateLocationHdfs$fGenericCreateLocationHdfs CreateLocationFsxWindowsResponse!CreateLocationFsxWindowsResponse'2$sel:locationArn:CreateLocationFsxWindowsResponse'1$sel:httpStatus:CreateLocationFsxWindowsResponse'CreateLocationFsxWindowsCreateLocationFsxWindows'%$sel:domain:CreateLocationFsxWindows'+$sel:subdirectory:CreateLocationFsxWindows'#$sel:tags:CreateLocationFsxWindows'/$sel:fsxFilesystemArn:CreateLocationFsxWindows'0$sel:securityGroupArns:CreateLocationFsxWindows'#$sel:user:CreateLocationFsxWindows''$sel:password:CreateLocationFsxWindows'newCreateLocationFsxWindowscreateLocationFsxWindows_domain%createLocationFsxWindows_subdirectorycreateLocationFsxWindows_tags)createLocationFsxWindows_fsxFilesystemArn*createLocationFsxWindows_securityGroupArnscreateLocationFsxWindows_user!createLocationFsxWindows_password#newCreateLocationFsxWindowsResponse,createLocationFsxWindowsResponse_locationArn+createLocationFsxWindowsResponse_httpStatus!$fToQueryCreateLocationFsxWindows $fToPathCreateLocationFsxWindows $fToJSONCreateLocationFsxWindows#$fToHeadersCreateLocationFsxWindows $fNFDataCreateLocationFsxWindows"$fHashableCreateLocationFsxWindows($fNFDataCreateLocationFsxWindowsResponse$$fAWSRequestCreateLocationFsxWindows$$fEqCreateLocationFsxWindowsResponse&$fReadCreateLocationFsxWindowsResponse&$fShowCreateLocationFsxWindowsResponse)$fGenericCreateLocationFsxWindowsResponse$fEqCreateLocationFsxWindows$fShowCreateLocationFsxWindows!$fGenericCreateLocationFsxWindows CreateLocationFsxOpenZfsResponse!CreateLocationFsxOpenZfsResponse'2$sel:locationArn:CreateLocationFsxOpenZfsResponse'1$sel:httpStatus:CreateLocationFsxOpenZfsResponse'CreateLocationFsxOpenZfsCreateLocationFsxOpenZfs'+$sel:subdirectory:CreateLocationFsxOpenZfs'#$sel:tags:CreateLocationFsxOpenZfs'/$sel:fsxFilesystemArn:CreateLocationFsxOpenZfs''$sel:protocol:CreateLocationFsxOpenZfs'0$sel:securityGroupArns:CreateLocationFsxOpenZfs'newCreateLocationFsxOpenZfs%createLocationFsxOpenZfs_subdirectorycreateLocationFsxOpenZfs_tags)createLocationFsxOpenZfs_fsxFilesystemArn!createLocationFsxOpenZfs_protocol*createLocationFsxOpenZfs_securityGroupArns#newCreateLocationFsxOpenZfsResponse,createLocationFsxOpenZfsResponse_locationArn+createLocationFsxOpenZfsResponse_httpStatus!$fToQueryCreateLocationFsxOpenZfs $fToPathCreateLocationFsxOpenZfs $fToJSONCreateLocationFsxOpenZfs#$fToHeadersCreateLocationFsxOpenZfs $fNFDataCreateLocationFsxOpenZfs"$fHashableCreateLocationFsxOpenZfs($fNFDataCreateLocationFsxOpenZfsResponse$$fAWSRequestCreateLocationFsxOpenZfs$$fEqCreateLocationFsxOpenZfsResponse&$fReadCreateLocationFsxOpenZfsResponse&$fShowCreateLocationFsxOpenZfsResponse)$fGenericCreateLocationFsxOpenZfsResponse$fEqCreateLocationFsxOpenZfs$fShowCreateLocationFsxOpenZfs!$fGenericCreateLocationFsxOpenZfsCreateLocationFsxOntapResponseCreateLocationFsxOntapResponse'0$sel:locationArn:CreateLocationFsxOntapResponse'/$sel:httpStatus:CreateLocationFsxOntapResponse'CreateLocationFsxOntapCreateLocationFsxOntap')$sel:subdirectory:CreateLocationFsxOntap'!$sel:tags:CreateLocationFsxOntap'%$sel:protocol:CreateLocationFsxOntap'.$sel:securityGroupArns:CreateLocationFsxOntap'5$sel:storageVirtualMachineArn:CreateLocationFsxOntap'newCreateLocationFsxOntap#createLocationFsxOntap_subdirectorycreateLocationFsxOntap_tagscreateLocationFsxOntap_protocol(createLocationFsxOntap_securityGroupArns/createLocationFsxOntap_storageVirtualMachineArn!newCreateLocationFsxOntapResponse*createLocationFsxOntapResponse_locationArn)createLocationFsxOntapResponse_httpStatus$fToQueryCreateLocationFsxOntap$fToPathCreateLocationFsxOntap$fToJSONCreateLocationFsxOntap!$fToHeadersCreateLocationFsxOntap$fNFDataCreateLocationFsxOntap $fHashableCreateLocationFsxOntap&$fNFDataCreateLocationFsxOntapResponse"$fAWSRequestCreateLocationFsxOntap"$fEqCreateLocationFsxOntapResponse$$fReadCreateLocationFsxOntapResponse$$fShowCreateLocationFsxOntapResponse'$fGenericCreateLocationFsxOntapResponse$fEqCreateLocationFsxOntap$fShowCreateLocationFsxOntap$fGenericCreateLocationFsxOntapCreateLocationFsxLustreResponse CreateLocationFsxLustreResponse'1$sel:locationArn:CreateLocationFsxLustreResponse'0$sel:httpStatus:CreateLocationFsxLustreResponse'CreateLocationFsxLustreCreateLocationFsxLustre'*$sel:subdirectory:CreateLocationFsxLustre'"$sel:tags:CreateLocationFsxLustre'.$sel:fsxFilesystemArn:CreateLocationFsxLustre'/$sel:securityGroupArns:CreateLocationFsxLustre'newCreateLocationFsxLustre$createLocationFsxLustre_subdirectorycreateLocationFsxLustre_tags(createLocationFsxLustre_fsxFilesystemArn)createLocationFsxLustre_securityGroupArns"newCreateLocationFsxLustreResponse+createLocationFsxLustreResponse_locationArn*createLocationFsxLustreResponse_httpStatus $fToQueryCreateLocationFsxLustre$fToPathCreateLocationFsxLustre$fToJSONCreateLocationFsxLustre"$fToHeadersCreateLocationFsxLustre$fNFDataCreateLocationFsxLustre!$fHashableCreateLocationFsxLustre'$fNFDataCreateLocationFsxLustreResponse#$fAWSRequestCreateLocationFsxLustre#$fEqCreateLocationFsxLustreResponse%$fReadCreateLocationFsxLustreResponse%$fShowCreateLocationFsxLustreResponse($fGenericCreateLocationFsxLustreResponse$fEqCreateLocationFsxLustre$fReadCreateLocationFsxLustre$fShowCreateLocationFsxLustre $fGenericCreateLocationFsxLustreCreateLocationEfsResponseCreateLocationEfsResponse'+$sel:locationArn:CreateLocationEfsResponse'*$sel:httpStatus:CreateLocationEfsResponse'CreateLocationEfsCreateLocationEfs'&$sel:accessPointArn:CreateLocationEfs'/$sel:fileSystemAccessRoleArn:CreateLocationEfs'+$sel:inTransitEncryption:CreateLocationEfs'$$sel:subdirectory:CreateLocationEfs'$sel:tags:CreateLocationEfs'($sel:efsFilesystemArn:CreateLocationEfs'!$sel:ec2Config:CreateLocationEfs'newCreateLocationEfs createLocationEfs_accessPointArn)createLocationEfs_fileSystemAccessRoleArn%createLocationEfs_inTransitEncryptioncreateLocationEfs_subdirectorycreateLocationEfs_tags"createLocationEfs_efsFilesystemArncreateLocationEfs_ec2ConfignewCreateLocationEfsResponse%createLocationEfsResponse_locationArn$createLocationEfsResponse_httpStatus$fToQueryCreateLocationEfs$fToPathCreateLocationEfs$fToJSONCreateLocationEfs$fToHeadersCreateLocationEfs$fNFDataCreateLocationEfs$fHashableCreateLocationEfs!$fNFDataCreateLocationEfsResponse$fAWSRequestCreateLocationEfs$fEqCreateLocationEfsResponse$fReadCreateLocationEfsResponse$fShowCreateLocationEfsResponse"$fGenericCreateLocationEfsResponse$fEqCreateLocationEfs$fReadCreateLocationEfs$fShowCreateLocationEfs$fGenericCreateLocationEfsCreateAgentResponseCreateAgentResponse'"$sel:agentArn:CreateAgentResponse'$$sel:httpStatus:CreateAgentResponse' CreateAgent CreateAgent'$sel:agentName:CreateAgent'#$sel:securityGroupArns:CreateAgent'$sel:subnetArns:CreateAgent'$sel:tags:CreateAgent'$sel:vpcEndpointId:CreateAgent'$sel:activationKey:CreateAgent'newCreateAgentcreateAgent_agentNamecreateAgent_securityGroupArnscreateAgent_subnetArnscreateAgent_tagscreateAgent_vpcEndpointIdcreateAgent_activationKeynewCreateAgentResponsecreateAgentResponse_agentArncreateAgentResponse_httpStatus$fToQueryCreateAgent$fToPathCreateAgent$fToJSONCreateAgent$fToHeadersCreateAgent$fNFDataCreateAgent$fHashableCreateAgent$fNFDataCreateAgentResponse$fAWSRequestCreateAgent$fEqCreateAgentResponse$fReadCreateAgentResponse$fShowCreateAgentResponse$fGenericCreateAgentResponse$fEqCreateAgent$fReadCreateAgent$fShowCreateAgent$fGenericCreateAgentCancelTaskExecutionResponseCancelTaskExecutionResponse',$sel:httpStatus:CancelTaskExecutionResponse'CancelTaskExecutionCancelTaskExecution'*$sel:taskExecutionArn:CancelTaskExecution'newCancelTaskExecution$cancelTaskExecution_taskExecutionArnnewCancelTaskExecutionResponse&cancelTaskExecutionResponse_httpStatus$fToQueryCancelTaskExecution$fToPathCancelTaskExecution$fToJSONCancelTaskExecution$fToHeadersCancelTaskExecution$fNFDataCancelTaskExecution$fHashableCancelTaskExecution#$fNFDataCancelTaskExecutionResponse$fAWSRequestCancelTaskExecution$fEqCancelTaskExecutionResponse!$fReadCancelTaskExecutionResponse!$fShowCancelTaskExecutionResponse$$fGenericCancelTaskExecutionResponse$fEqCancelTaskExecution$fReadCancelTaskExecution$fShowCancelTaskExecution$fGenericCancelTaskExecutionUntagResourceResponseUntagResourceResponse'&$sel:httpStatus:UntagResourceResponse' UntagResourceUntagResource'$sel:resourceArn:UntagResource'$sel:keys:UntagResource'newUntagResourceuntagResource_resourceArnuntagResource_keysnewUntagResourceResponse untagResourceResponse_httpStatus$fToQueryUntagResource$fToPathUntagResource$fToJSONUntagResource$fToHeadersUntagResource$fNFDataUntagResource$fHashableUntagResource$fNFDataUntagResourceResponse$fAWSRequestUntagResource$fEqUntagResourceResponse$fReadUntagResourceResponse$fShowUntagResourceResponse$fGenericUntagResourceResponse$fEqUntagResource$fReadUntagResource$fShowUntagResource$fGenericUntagResourceUpdateAgentResponseUpdateAgentResponse'$$sel:httpStatus:UpdateAgentResponse' UpdateAgent UpdateAgent'$sel:name:UpdateAgent'$sel:agentArn:UpdateAgent'newUpdateAgentupdateAgent_nameupdateAgent_agentArnnewUpdateAgentResponseupdateAgentResponse_httpStatus$fToQueryUpdateAgent$fToPathUpdateAgent$fToJSONUpdateAgent$fToHeadersUpdateAgent$fNFDataUpdateAgent$fHashableUpdateAgent$fNFDataUpdateAgentResponse$fAWSRequestUpdateAgent$fEqUpdateAgentResponse$fReadUpdateAgentResponse$fShowUpdateAgentResponse$fGenericUpdateAgentResponse$fEqUpdateAgent$fReadUpdateAgent$fShowUpdateAgent$fGenericUpdateAgentUpdateLocationHdfsResponseUpdateLocationHdfsResponse'+$sel:httpStatus:UpdateLocationHdfsResponse'UpdateLocationHdfsUpdateLocationHdfs'"$sel:agentArns:UpdateLocationHdfs'+$sel:authenticationType:UpdateLocationHdfs'"$sel:blockSize:UpdateLocationHdfs''$sel:kerberosKeytab:UpdateLocationHdfs')$sel:kerberosKrb5Conf:UpdateLocationHdfs'*$sel:kerberosPrincipal:UpdateLocationHdfs'*$sel:kmsKeyProviderUri:UpdateLocationHdfs'"$sel:nameNodes:UpdateLocationHdfs')$sel:qopConfiguration:UpdateLocationHdfs'*$sel:replicationFactor:UpdateLocationHdfs'#$sel:simpleUser:UpdateLocationHdfs'%$sel:subdirectory:UpdateLocationHdfs'$$sel:locationArn:UpdateLocationHdfs'newUpdateLocationHdfsupdateLocationHdfs_agentArns%updateLocationHdfs_authenticationTypeupdateLocationHdfs_blockSize!updateLocationHdfs_kerberosKeytab#updateLocationHdfs_kerberosKrb5Conf$updateLocationHdfs_kerberosPrincipal$updateLocationHdfs_kmsKeyProviderUriupdateLocationHdfs_nameNodes#updateLocationHdfs_qopConfiguration$updateLocationHdfs_replicationFactorupdateLocationHdfs_simpleUserupdateLocationHdfs_subdirectoryupdateLocationHdfs_locationArnnewUpdateLocationHdfsResponse%updateLocationHdfsResponse_httpStatus$fToQueryUpdateLocationHdfs$fToPathUpdateLocationHdfs$fToJSONUpdateLocationHdfs$fToHeadersUpdateLocationHdfs$fNFDataUpdateLocationHdfs$fHashableUpdateLocationHdfs"$fNFDataUpdateLocationHdfsResponse$fAWSRequestUpdateLocationHdfs$fEqUpdateLocationHdfsResponse $fReadUpdateLocationHdfsResponse $fShowUpdateLocationHdfsResponse#$fGenericUpdateLocationHdfsResponse$fEqUpdateLocationHdfs$fReadUpdateLocationHdfs$fShowUpdateLocationHdfs$fGenericUpdateLocationHdfsUpdateLocationNfsResponseUpdateLocationNfsResponse'*$sel:httpStatus:UpdateLocationNfsResponse'UpdateLocationNfsUpdateLocationNfs'$$sel:mountOptions:UpdateLocationNfs'$$sel:onPremConfig:UpdateLocationNfs'$$sel:subdirectory:UpdateLocationNfs'#$sel:locationArn:UpdateLocationNfs'newUpdateLocationNfsupdateLocationNfs_mountOptionsupdateLocationNfs_onPremConfigupdateLocationNfs_subdirectoryupdateLocationNfs_locationArnnewUpdateLocationNfsResponse$updateLocationNfsResponse_httpStatus$fToQueryUpdateLocationNfs$fToPathUpdateLocationNfs$fToJSONUpdateLocationNfs$fToHeadersUpdateLocationNfs$fNFDataUpdateLocationNfs$fHashableUpdateLocationNfs!$fNFDataUpdateLocationNfsResponse$fAWSRequestUpdateLocationNfs$fEqUpdateLocationNfsResponse$fReadUpdateLocationNfsResponse$fShowUpdateLocationNfsResponse"$fGenericUpdateLocationNfsResponse$fEqUpdateLocationNfs$fReadUpdateLocationNfs$fShowUpdateLocationNfs$fGenericUpdateLocationNfs#UpdateLocationObjectStorageResponse$UpdateLocationObjectStorageResponse'4$sel:httpStatus:UpdateLocationObjectStorageResponse'UpdateLocationObjectStorageUpdateLocationObjectStorage'+$sel:accessKey:UpdateLocationObjectStorage'+$sel:agentArns:UpdateLocationObjectStorage'+$sel:secretKey:UpdateLocationObjectStorage'3$sel:serverCertificate:UpdateLocationObjectStorage',$sel:serverPort:UpdateLocationObjectStorage'0$sel:serverProtocol:UpdateLocationObjectStorage'.$sel:subdirectory:UpdateLocationObjectStorage'-$sel:locationArn:UpdateLocationObjectStorage'newUpdateLocationObjectStorage%updateLocationObjectStorage_accessKey%updateLocationObjectStorage_agentArns%updateLocationObjectStorage_secretKey-updateLocationObjectStorage_serverCertificate&updateLocationObjectStorage_serverPort*updateLocationObjectStorage_serverProtocol(updateLocationObjectStorage_subdirectory'updateLocationObjectStorage_locationArn&newUpdateLocationObjectStorageResponse.updateLocationObjectStorageResponse_httpStatus$$fToQueryUpdateLocationObjectStorage#$fToPathUpdateLocationObjectStorage#$fToJSONUpdateLocationObjectStorage&$fToHeadersUpdateLocationObjectStorage#$fNFDataUpdateLocationObjectStorage%$fHashableUpdateLocationObjectStorage+$fNFDataUpdateLocationObjectStorageResponse'$fAWSRequestUpdateLocationObjectStorage'$fEqUpdateLocationObjectStorageResponse)$fReadUpdateLocationObjectStorageResponse)$fShowUpdateLocationObjectStorageResponse,$fGenericUpdateLocationObjectStorageResponse$fEqUpdateLocationObjectStorage!$fShowUpdateLocationObjectStorage$$fGenericUpdateLocationObjectStorageUpdateLocationSmbResponseUpdateLocationSmbResponse'*$sel:httpStatus:UpdateLocationSmbResponse'UpdateLocationSmbUpdateLocationSmb'!$sel:agentArns:UpdateLocationSmb'$sel:domain:UpdateLocationSmb'$$sel:mountOptions:UpdateLocationSmb' $sel:password:UpdateLocationSmb'$$sel:subdirectory:UpdateLocationSmb'$sel:user:UpdateLocationSmb'#$sel:locationArn:UpdateLocationSmb'newUpdateLocationSmbupdateLocationSmb_agentArnsupdateLocationSmb_domainupdateLocationSmb_mountOptionsupdateLocationSmb_passwordupdateLocationSmb_subdirectoryupdateLocationSmb_userupdateLocationSmb_locationArnnewUpdateLocationSmbResponse$updateLocationSmbResponse_httpStatus$fToQueryUpdateLocationSmb$fToPathUpdateLocationSmb$fToJSONUpdateLocationSmb$fToHeadersUpdateLocationSmb$fNFDataUpdateLocationSmb$fHashableUpdateLocationSmb!$fNFDataUpdateLocationSmbResponse$fAWSRequestUpdateLocationSmb$fEqUpdateLocationSmbResponse$fReadUpdateLocationSmbResponse$fShowUpdateLocationSmbResponse"$fGenericUpdateLocationSmbResponse$fEqUpdateLocationSmb$fShowUpdateLocationSmb$fGenericUpdateLocationSmbUpdateTaskResponseUpdateTaskResponse'#$sel:httpStatus:UpdateTaskResponse' UpdateTask UpdateTask'&$sel:cloudWatchLogGroupArn:UpdateTask'$sel:excludes:UpdateTask'$sel:includes:UpdateTask'$sel:name:UpdateTask'$sel:options:UpdateTask'$sel:schedule:UpdateTask'$sel:taskArn:UpdateTask' newUpdateTask updateTask_cloudWatchLogGroupArnupdateTask_excludesupdateTask_includesupdateTask_nameupdateTask_optionsupdateTask_scheduleupdateTask_taskArnnewUpdateTaskResponseupdateTaskResponse_httpStatus$fToQueryUpdateTask$fToPathUpdateTask$fToJSONUpdateTask$fToHeadersUpdateTask$fNFDataUpdateTask$fHashableUpdateTask$fNFDataUpdateTaskResponse$fAWSRequestUpdateTask$fEqUpdateTaskResponse$fReadUpdateTaskResponse$fShowUpdateTaskResponse$fGenericUpdateTaskResponse$fEqUpdateTask$fReadUpdateTask$fShowUpdateTask$fGenericUpdateTaskUpdateTaskExecutionResponseUpdateTaskExecutionResponse',$sel:httpStatus:UpdateTaskExecutionResponse'UpdateTaskExecutionUpdateTaskExecution'*$sel:taskExecutionArn:UpdateTaskExecution'!$sel:options:UpdateTaskExecution'newUpdateTaskExecution$updateTaskExecution_taskExecutionArnupdateTaskExecution_optionsnewUpdateTaskExecutionResponse&updateTaskExecutionResponse_httpStatus$fToQueryUpdateTaskExecution$fToPathUpdateTaskExecution$fToJSONUpdateTaskExecution$fToHeadersUpdateTaskExecution$fNFDataUpdateTaskExecution$fHashableUpdateTaskExecution#$fNFDataUpdateTaskExecutionResponse$fAWSRequestUpdateTaskExecution$fEqUpdateTaskExecutionResponse!$fReadUpdateTaskExecutionResponse!$fShowUpdateTaskExecutionResponse$$fGenericUpdateTaskExecutionResponse$fEqUpdateTaskExecution$fReadUpdateTaskExecution$fShowUpdateTaskExecution$fGenericUpdateTaskExecution