•When the connection is used as a target, the last batch of the insert load is not reflected in the record count. Refer the session logs for the record count of the last batch inserted.
•For example, if the commit interval is set to 1 million and the actual rows inserted are 1.1 million, the record count in the UI shows 1 million and the session logs reveal the row count of the reminder 100k records.
•Set the commit interval to the highest value possible before java.lang.OutofMemoryError is encountered.
•When the connection is used as a target to load data into Hadoop, ensure that all the fields are mapped.
•After a data load in Hive, Impala needs to be refreshed manually for the latest changes to the table to be reflected in Impala. In the current version, the connector does not automatically refresh Impala upon a Hive dataset insert.
•When you enable Apache Sentry authorization on a Cloudera cluster or Apache Ranger authorization on a Hortonworks cluster, you must enable the UseExternalTableToWrite property in the Custom Configuration Details section of the Secure Agent and restart the Secure Agent. Otherwise, the mapping tasks fails with the following error:
WRITER_1_*_1> JAVA PLUGIN_1762 [2018-03-13 11:15:17.884][ERROR]java.sql.SQLException: Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.MoveTask
Enabling the UseExternalTableToWrite Property
To enable the UseExternalTableToWrite property for the Secure Agent, perform the following steps:
1In Administrator, select Runtime Environments.
2Select the Secure Agent for which you want to set the custom configuration flag.
3On the upper-right corner of the page, click Edit.
4In the Custom Configuration Details section, select the Service as Data Integration Server.