The Amazon Resource Names (ARNs) of the agents that are used to connect to the HDFS cluster.
Required? | True |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
Aliases | AgentArns |
The type of authentication used to determine the identity of the user.
Required? | True |
Position? | 1 |
Accept pipeline input? | True (ByValue, ByPropertyName) |
The size of data blocks to write into the HDFS cluster. The block size must be a multiple of 512 bytes. The default block size is 128 mebibytes (MiB).
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
Amazon.PowerShell.Cmdlets.DSYN.AmazonDataSyncClientCmdlet.ClientConfig
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
This parameter overrides confirmation prompts to force the cmdlet to continue its operation. This parameter should always be used with caution.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
The Kerberos key table (keytab) that contains mappings between the defined Kerberos principal and the encrypted keys. You can load the keytab from a file by providing the file's address. If you're using the CLI, it performs base64 encoding for you. Otherwise, provide the base64-encoded text. If KERBEROS is specified for AuthenticationType, this parameter is required.The cmdlet will automatically convert the supplied parameter of type string, string[], System.IO.FileInfo or System.IO.Stream to byte[] before supplying it to the service.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
The krb5.conf file that contains the Kerberos configuration information. You can load the krb5.conf file by providing the file's address. If you're using the CLI, it performs the base64 encoding for you. Otherwise, provide the base64-encoded text. If KERBEROS is specified for AuthenticationType, this parameter is required.The cmdlet will automatically convert the supplied parameter of type string, string[], System.IO.FileInfo or System.IO.Stream to byte[] before supplying it to the service.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
The Kerberos principal with access to the files and folders on the HDFS cluster. If KERBEROS is specified for AuthenticationType, this parameter is required.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
The URI of the HDFS cluster's Key Management Server (KMS).
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
The NameNode that manages the HDFS namespace. The NameNode performs operations such as opening, closing, and renaming files and directories. The NameNode contains the information to map blocks of data to the DataNodes. You can use only one NameNode.
Required? | True |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
Aliases | NameNodes |
Changes the cmdlet behavior to return the value passed to the AuthenticationType parameter. The -PassThru parameter is deprecated, use -Select '^AuthenticationType' instead. This parameter will be removed in a future version.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
The data transfer protection setting configured on the HDFS cluster. This setting corresponds to your dfs.data.transfer.protection setting in the hdfs-site.xml file on your Hadoop cluster.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
The RPC protection setting configured on the HDFS cluster. This setting corresponds to your hadoop.rpc.protection setting in your core-site.xml file on your Hadoop cluster.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
-ReplicationFactor <
Int32>
The number of DataNodes to replicate the data to when writing to the HDFS cluster. By default, data is replicated to three DataNodes.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
Use the -Select parameter to control the cmdlet output. The default value is 'LocationArn'. Specifying -Select '*' will result in the cmdlet returning the whole service response (Amazon.DataSync.Model.CreateLocationHdfsResponse). Specifying the name of a property of type Amazon.DataSync.Model.CreateLocationHdfsResponse will result in that property being returned. Specifying -Select '^ParameterName' will result in the cmdlet returning the selected cmdlet parameter value.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
The user name used to identify the client on the host operating system. If SIMPLE is specified for AuthenticationType, this parameter is required.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
A subdirectory in the HDFS cluster. This subdirectory is used to read data from or write data to the HDFS cluster. If the subdirectory isn't specified, it will default to /.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
The key-value pair that represents the tag that you want to add to the location. The value can be an empty string. We recommend using tags to name your resources.
Required? | False |
Position? | Named |
Accept pipeline input? | True (ByPropertyName) |
Aliases | Tags |