Before you use Hadoop Files V2 Connector in tasks, an administrator must verify the following prerequisites:
•Use Hadoop Files V2 Connector on the Linux 64-bit operating system.
•Before you connect to the Hadoop cluster, enable all the required ports for the data node, name node, KMS, and Zookeeper services.
Access Kerberos-enabled Hadoop cluster
Configure the /etc/hosts file and copy the Kerberos configuration file for HDFS instances that use Kerberos authentication.
1Open the /etc/hosts file located in the /etc directory on the Secure Agent machine on Linux.
2To configure the Secure Agent to work with the Kerberos Key Distribution Center (KDC), make an entry of the KDC hosts in the /etc/hosts file.
3Copy the krb5.conf configuration file from the /etc directory in the Hadoop cluster node to either of the following directories available within your Secure Agent installation:
4If there is a jdk directory within the <Secure agent installation directory>\apps folder, navigate to the following directory and copy the krb5.conf configuration file from the /etc directory in the Hadoop cluster node to the following directory within your Secure Agent installation:
5If the cluster is SSL enabled, import the certificate alias file to either of the following directories available within your Secure Agent installation:
6If there is a jdk directory within the <Secure agent installation directory>\apps folder, navigate to the following directory and import the certificate alias file to the following directory within your Secure Agent installation:
If you have a Cloudera, Amazon EMR, Hortonworks, or Microsoft HDInsight instance that does not use Kerberos authentication and runs in a Hadoop cluster environment, perform the following steps:
1Copy the conf configuration files from the /etc/hadoop/conf directory in the Hadoop cluster node and place it in the Secure Agent location with full permission.