Installation and Upgrade Guide > Part VII: Install and Configure for MapR > Manual Configuration Tasks for MapR > Generate MapR Tickets
  

Generate MapR Tickets

To run mappings on a MapR cluster that uses Kerberos or MapR Ticket authentication with information in Hive tables, generate a MapR ticket for the following users:
Data Integration Service User
The Data Integration Service user requires an account on the MapR cluster and a MapR ticket on the Data Integration Service machine.
When the MapR cluster uses both Kerberos and Ticket authentication, you generate a ticket for the Data Integration Service user for each authentication system.
Developer Tool User
The Developer tool user requires an account on the MapR cluster and a MapR ticket on the machine where the Developer tool is installed.
After you generate and save MapR tickets, you perform additional steps to configure application services to communicate with the MapR cluster.

Create the Developer Tool User

To enable the Developer tool to communicate with the MapR cluster, create and configure an account for the Developer tool user on every node in the cluster.

Generate Tickets

After you create a MapR user account for the Data Integration Service user and the Developer tool user, generate a MapR ticket for each user and save it to a local directory, depending on the user requirements that are listed below.
To generate a MapR ticket, refer to MapR documentation.

Data Integration Service User Ticket

Generate a MapR ticket for the Data Integration Service user. Name the ticket file using the following naming convention:
maprticket_<user name>
Save the ticket file in the /tmp directory of the machine that runs the Data Integration Service.
When the MapR cluster is configured to enable a user to use Kerberos authentication and MapR Ticket authentication, you generate a MapR ticketfile for the user for each authentication mode. Save one ticketfile in /tmp. Save the other ticketfile in any directory on the Data Integration Service machine, and provide the location as the value for the MAPR_TICKETFILE_LOCATION property in the Data Integration Service Process properties.

Developer Tool User Ticket

Generate a MapR ticket for the Developer tool user. Name the ticket file using the following naming convention:
maprticket_<user name>
Save the ticket file in the %TEMP% directory of the machine the runs the Developer tool.

Configure Informatica Application Services

Configure properties on the following Informatica application services:
Data Integration Service
When the MapR cluster is secured with MapR Kerberos authentication, edit Data Integration Service properties to enable communication between the Informatica domain and the cluster.
Analyst Service
If you use the Analyst tool to profile data in Hive data objects, configure properties on the Analyst Service to enable communication between the Analyst tool and the cluster, including testing of the Hive connection.

Configure the Data Integration Service

When the MapR cluster is secured with MapR Kerberos authentication, edit Data Integration Service properties to enable communication between the Informatica domain and the cluster.

Data Integration Service Properties

In the Administrator tool Domain Navigator, select the Data Integration Service to configure, then select the Properties tab.
In the Custom Properties area, define the following property value:
Property
Value
ExecutionContextOptions.JVMOption1
-Dmapr.library.flatclass

Data Integration Service Process Properties

In the Administrator tool Domain Navigator, select the Data Integration Service to configure, and then select the Processes tab.
In the Custom Properties area, define the following properties and values:
Property
Value
ExecutionContextOptions.JVMOption
-Djava.security.krb5.conf=<Informatica installation directory>/services/shared/security/krb5.conf
ExecutionContextOptions.JVMOption1
-Dmapr.library.flatclass
ExecutionContextOptions.JVMOption2
-Dhadoop.login=<MAPR_ECOSYSTEM_LOGIN_OPTS> -Dhttps.protocols=TLSv1.2
where <MAPR_ECOSYSTEM_LOGIN_OPTS> is the value of the MAPR_ECOSYSTEM_LOGIN_OPTS property in the file /opt/mapr/conf/env.sh.
For example, -Dhadoop.login=hybrid
ExecutionContextOptions.JVMOption7
-Dhttps.protocols=TLSv1.2
In the Environment Variables area, configure the following property to define the Kerberos authentication protocol:
Property
Value
JAVA_OPTS
-Dhadoop.login=<MAPR_ECOSYSTEM_LOGIN_OPTS> -Dhttps.protocols=TLSv1.2
where <MAPR_ECOSYSTEM_LOGIN_OPTS> is the value of the MAPR_ECOSYSTEM_LOGIN_OPTS property in the file /opt/mapr/conf/env.sh.
MAPR_HOME
Hadoop distribution directory location on the machine that runs the Data Integration Service.
For example,
<Informatica installation directory>/services/shared/hadoop/mapr_5.2.0
MAPR_TICKETFILE_LOCATION
Optional. Directory where an additional MapR Ticket file is stored on the machine that runs the Data Integration Service.
When the MapR cluster is configured to enable a user to use Kerberos authentication and MapR Ticket authentication, generate a MapR ticketfile for the user for each authentication mode. Save one ticketfile in /tmp. Save the other ticketfile in any directory on the Data Integration Service machine, and provide the location as the value for this property.
For example, for a user id 1234, save a MapR ticketfile named like maprticket_1234 in /tmp, and save another MapR ticketfile named like maprticket_1234 in the MAPR_TICKETFILE_LOCATION.
Note: The ticketfiles can have the same or different names. You must generate the MapR ticketfiles separately and save one to the MAPR_TICKETFILE_LOCATION.
Changes take effect when you restart the Data Integration Service.

Configure the Analyst Service

If you use the Analyst tool to profile data in Hive data objects, configure properties on the Analyst Service to enable communication between the Analyst tool and the cluster, including testing of the Hive connection.
In the Administrator tool Domain Navigator, select the Analyst Service to configure, then select the Processes tab.
In the Advanced Properties area, define the following property value:
Property
Value
ExecutionContextOptions.JVMOption1
-Dmapr.library.flatclass
In the Environment Variables area, configure the following property to define the Kerberos authentication protocol:
Property
Value
JAVA_OPTS
-Dhadoop.login=hybrid -Dhttps.protocols=TLSv1.2
MAPR_HOME
Hadoop distribution directory location on the machine that runs the Data Integration Service.
For example,
<Informatica installation directory>/services/shared/hadoop/mapr_5.2.0
MAPR_TICKETFILE_LOCATION
Directory where the MapR Ticket file is stored on the machine that runs the Analyst Service.
For example,
/export/home/username1/Keytabs_and_krb5conf/Tickets/project1/maprticket_30103
LD_LIBRARY_PATH
The location of Hadoop libraries.
For example,
<Informatica installation directory>/java/jre/lib:<Informatica installation directory>/services/shared/bin:<Informatica installation directory>/server/bin:<Informatica installation directory>/services/shared/hadoop/<MapR location>/lib/native/Linux-amd64-64
Changes take effect when you restart the Analyst Service.

Test the Hive Connection

After you configure users for MapR Ticket or Kerberos authentication on MapR clusters, you can test the Hive connection.
To test the Hive connection, or perform a metadata fetch task, use the following format for the connection string if the cluster is Kerberos-enabled:
jdbc:hive2://<hostname>:10000/default;principal=<SPN>
For example,
jdbc:hive2://myServer2:10000/default;principal=mapr/myServer2@clustername
Note: When the mapping performs a metadata fetch of a complex file object, the user whose maprticket is present at %TEMP% on the Windows machine must have read permission on the HDFS directory to list the files inside it and perform the import action. The metadata fetch operation ignores privileges of the user who is listed in the HDFS connection definition.