Live Data Map Administrator Guide > Managing Resources > Resource Type
  

Resource Type

A resource type represents the type of source system from where scanners extract metadata. A resource type identifies the required and optional metadata types that you need to configure for each resource. Examples of the resource type are Business Objects, Oracle, PowerCenter, and Teradata.
Each resource type has different properties that you need to configure when you create a resource. The properties include the connection properties and additional properties, such as resource owner and resource classification.
You can create the following types of resources:
To run a scan job on resource types, such as Oracle, DB2, and SQL Server, use the native resource types available in Live Data Map. Informatica recommends that you do not configure the JDBC resource type to fetch data from the supported native metadata sources. However, you can use the JDBC resource type to fetch metadata from MySQL and i5/OS sources.

SQL Server Integration Services Resource Properties

The following tables list the properties that you must configure to add an SQL Server Integration Services (SSIS) resource:
The General tab includes the following properties:
Property
Description
SSIS Scanner Type
Select the type of SSIS resource:
  • - Repository Server. Select this option if you have a configured a repository in MSDB database to store all the packages in SSIS.
  • - File. Select this option if you have configured a file system to store all the packages in SSIS. Configure the following properties if you select the File option:
    • - Agent URL. The URL to access the SQL Server Agent that automates package runs.
    • - File. Click Choose to select the file that includes the packages that you want to upload to Live Data Map. Alternatively, you can place the file in the File text box from your file browser using a drag-and-drop operation. Make sure that the files have the extension .dtsx. If you want to upload multiple files, add the files in a .zip file and upload the .zip file.
    • - Encoding. Select the type of encoding you want for the extracted metadata from the drop-down list.
    • - Password. The password for the package.
    • - Variables value file. Click Choose to select the file that includes values for the SSIS variables. Alternatively, you can place the file in the Variable values file text box from your file browser using a drag-and-drop operation.
Agent URL
The URL to access the SQL Server Agent that automates package runs.
Note: Make sure that the SQL Server Agent runs on the same machine if you had selected the SSIS Scanner Type as Repository Server.
SQL Server Version
Select one of the following options from the SQL Server Version drop-down list:
  • - SQL Server 2008
  • - SQL Server 2012
  • - SQL Server 2014
Default is SQL Server 2008.
Host Name
The host name or IP address of the machine where SSIS is running.
Password
The password for the package.
Package/Repository Name
Specify the repository or package from which you want to import metadata. Click Select. The Select Package/Repository Name dialog box appears. Select the required package using one of the following options:
  • - Select from list. Select the required package or repository from a list of packages or repositories.
  • - Select using regex. Provide an SQL regular expression to select the package or the repository name that matches with the expression.
Variable values file
Click Choose to select the file that includes values for the SSIS variables. Alternatively, you can place the file in the Variable values file text box from your file browser using a drag-and-drop operation.
The Metadata Load Settings tab includes the following properties:
Property
Description
Auto assign connections
Select this option to specify that the connection must be assigned automatically.
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.

SAP R/3 Resource Properties

You must complete the prerequisites listed before configuring the properties for an SAP R/3 resource.

Prerequisites

  1. 1. Create an Event Details Record (EDR) connection for SAP R/3 using Informatica Administrator. See the Informatica Administrator Guide for more information about creating connections.
  2. 2. Disable the Catalog Service.
  3. 3. Remove all the resource binary files in HDFS from the /Informatica/LDM/<service cluster name>/scanner/* directory. You can use the command hdfs dfs -rm -R /Informatica/LDM/<service cluster name>/scanner/* to remove all the resource binary files.
  4. 4. Download the sapjco3.jar file and copy the file to the following location: <Install_directory>/services/CatalogService/access/WEB-INF/lib
  5. 5. If you want to enable profiling, copy the following files to the locations listed:
  6. 6. Copy the libsapjco3.so file to the following location: <Install_directory> /services/shared/bin
  7. 7. Include the libsapjco3.so file in the SAPJCO.zip file and copy the SAPJCO.zip file to the following location: <Install_directory>/services/CatalogService/ScannerBinaries
  8. 8. Restart Informatica domain.

Resource Configuration Properties

The General tab includes the following properties:
Property
Description
Username
The user name to access SAP R/3 system.
Password
The password associated with the user name.
Application server host
Host name or IP address of the system that hosts SAP R/3.
System number
System number of the SAP R/3 system to which you want to connect.
Client
The SAP R/3 client to access data from the SAP R/3 system.
Language
Specify the language to be used while importing metadata using the resource.
Encoding
Default is UTF-8 character encoding for metadata imported from the resource. You cannot change the default setting for this property.
The following table describes the properties that you can configure in the Source Metadata section of the Metadata Load Settings tab:
Property
Description
Repository Objects
Imports the repository objects such as resources, information, and activities from the SAP R/3 system.
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.
You can enable data discovery for an SAP R/3 resource. See the Enable Data Discovery section for more information.

Oracle Business Intelligence Enterprise Edition Resource Properties

The following tables list the properties that you must configure to add an Oracle Business Intelligence Enterprise Edition (OBIEE) resource:
The General tab includes the following properties:
Property
Description
Version
Select one of the following options to specify the version of OBIEE:
  • - Auto Detect. Select this option to let Live Data Map detect the version of OBIEE.
  • - OBIEE 11.x. Select this option to specify the OBIEE version as 11.x.
Server URL
The OBIEE Presentation Server URL. If you use SSL, you must make sure that Live Data Map trusts the server certificate of the OBIEE Presentation Server.
Login User
The username used to log on to the OBIEE Presentation Server. Make sure that the username you use has the necessary permissions to import metadata.
Login Password
The password associated with the username.
Optimize for large models
Select this option to optimize the import of metadata for large OBIEE repository models.
If you select this option, Live Data Map does not import metadata for the following assets:
  • - Foreign keys
  • - Joins
  • - Relationships
  • - Logical foreign keys
In addition, Live Data Map does not store expression tree objects with lineage links.
If you do not select this option, Live Data Map imports the entire repository model, resulting in a high consumption of memory.
Incremental import
Select this option to import only the changes made to the data source since the last metadata import. If you do not use this option, Live Data Map imports the entire metadata from the data source.
Worker threads
Specify the number of worker threads to process metadata asynchronously. You can leave the value empty if you want Live Data Map to calculate the value. Live Data Map assigns a value between one and six based on the JVM architecture and number of available CPU cores.
You can use the following points to decide the value to use:
  • - You can provide a value that is greater than or equal to one and lesser than six to specify the number of worker threads required.
  • - If you specify an invalid value, Live Data Map shows a warning and uses the value one.
  • - If your machine has more memory, you can specify a higher value to process more metadata asynchronously.
  • Note: Specifying a higher value might impact performance of the system.
File
The Oracle Business Intelligence Repository RPD file where the metadata is stored.
Click Choose to select the RPD file. Alternatively, you can place the RPD file in the Variable values file text box from your file browser using a drag-and-drop operation.
Variable values file
(Optional) The file that defines the list of RPD variable values.
Click Choose to select the file that includes the variable values. Alternatively, you can place the file in the Variable values file text box from your file browser using a drag-and-drop operation.
Auto assign connections
Select this option to specify that the connection must be assigned automatically.
The Metadata Load Settings tab includes the following properties:
Property
Description
Repository Subset
Click Select. The Select Repository subset dialog box appears. Select the folders from where you want to import metadata for reports from the Oracle Business Intelligence Presentation Server.
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.

MDM Resource Properties

The following tables list the properties that you must configure to add an MDM resource:
The General tab includes the following properties:
Property
Description
User
The user name used to access the MDM Hub Store.
Password
The password associated with the user name.
JDBC
The JDBC string to connect to the MDM Hub Store.
JDBCDriverClassName
The JDBC driver class name specific to the MDM Hub Store.
The Metadata Load Settings tab includes the following properties:
Property
Description
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.

Informatica Cloud Service Resource Connection Properties

Informatica Cloud is an on-demand subscription service that provides access to applications, databases, platforms, and flat files hosted on premise or on a cloud. Informatica Cloud runs at a hosting facility. Before you add an Informatica Cloud Service resource, perform the steps listed in the Prerequisites section.

Prerequisites

  1. 1. Create an organization for your company on the Informatica Cloud website, define the organization hierarchy, and configure the organization properties. You must perform this step before you can use Informatica Cloud.
  2. Note: To create an organization, you must have a REST API license. If you do not have a REST API license, contact Informatica Global Customer Support.
  3. 2. Create a subscription account on Informatica Cloud.
  4. 3. Verify that the machine where you install the Informatica Cloud Secure Agent meets the minimum system requirements. The Informatica Cloud Secure Agent is a lightweight program that runs all tasks and enables secure communication across the firewall between your organization and Informatica Cloud.
  5. 4. Download, install, and register the Informatica Cloud Secure Agent using the Informatica Cloud user name and password.
  6. 5. Create the following tasks on Informatica Cloud:
    1. a. Mapping tasks
    2. A mapping defines reusable data flow logic that you can use in Mapping Configuration tasks. Use a mapping to define data flow logic that is not available in Data Synchronization tasks, such as specific ordering of logic or joining sources from different systems. When you configure a mapping, you describe the flow of data from source and target. You can add transformations to transform data, such as an Expression transformation for row-level calculations or a Filter transformation to remove data from the data flow.
    3. b. PowerCenter tasks
    4. The PowerCenter task allows you to import PowerCenter workflows in to Informatica Cloud and run them as Informatica Cloud tasks.
    5. c. Data synchronization tasks
    6. The Data Synchronization task allows you to synchronize data between a source and target.
Note: An Informatica Cloud Service resource imports all the tasks to Live Data Map the first time metadata is extracted from the resource. During the subsequent extract operations, the resource imports only the updated tasks to Live Data Map.
For more information about the prerequisites, see the Informatica Cloud User Guide and the Informatica Cloud Administrator Guide.

Connection Properties

The General tab includes the following properties:
Property
Description
Cloud URL
The URL to access the Informatica Cloud Service.
Username
The user name to connect to the Informatica Cloud Service.
Password
The password associated with the user name.
Auto Assign Connections
Select this option to specify that the connection must be assigned automatically.
The Metadata Load Settings tab includes the following properties:
Property
Description
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.

Custom Lineage Resource Properties

You can create a custom lineage resource to view the data lineage information for the assets in your organization. A custom lineage resource uses CSV files provided by you that include lineage data for your enterprise. You can use this option if you do not have an ETL tool supported by Live Data Map. The following tables list the properties that you must configure to add a custom lineage resource:
The General tab includes the following properties:
Property
Description
File
The CSV file or the .zip file that includes the CSV files with the lineage data. Click Choose to select the required CSV file or .zip file that you want to upload. Ensure that the CSV files in the .zip file are not stored in a directory within the .zip file.
If you want to select multiple CSV files, you must include the required CSV files in a .zip file and then select the .zip file for upload.
Note: Make sure that the CSV file includes the following parameters in the header:
  • - From Connection
  • - To Connection
  • - From Object
  • - To Object
The Metadata Load Settings tab includes the following properties:
Property
Description
Auto Assign Connections
Specifies to automatically assign the connection.
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.

Amazon S3 Resource Properties

The following tables list the properties that you must configure to add an Amazon S3 resource:
The General tab includes the following properties:
Property
Description
Amazon Web Services Bucket URL
Amazon Web Services URL to access a bucket.
Amazon Web Services Access Key ID
Amazon Web Services access key ID to sign requests that you send to Amazon Web Services.
Amazon Web Services Secret Access Key
Amazon Web Services secret access key to sign requests that you send to Amazon Web Services.
Amazon Web Services Bucket Name
Amazon Web Services bucket name that Live Data Map needs to scan.
Source Directory
The source directory from where metadata must be extracted.
The Metadata Load Settings tab includes the following properties:
Property
Description
File Types
Select any or all of the following file types from which you want to extract metadata:
  • - CSV
  • - XML
  • - JSON
First Row as Column Header for CSV
Select this option to specify the first row as the column header for the CSV file.
First Level Directory
Use this option to specify a directory or a list of directories under the source directory. If you leave this option blank, Live Data Map imports all the files from the specified source directory.
To specify a directory or a list of directories, you can perform the following steps:
  1. 1. Click Select.... The Select First Level Directory dialog box appears.
  2. 2. Select the required directories using one of the following options:
    • - Select from list: select the required directories from a list of directories.
    • - Select using regex: provide an SQL regular expression to select schemas that match the expression.
Note: If you are selecting multiple directories, you must separate the directories using a semicolon (;).
Include Subdirectory
Select this option to import all the files in the subdirectories under the source directory.
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.
You can enable data discovery for an Amazon S3 resource. See the Enable Data Discovery section for more information.

HDFS Resource Connection Properties

The following tables list the properties that you must configure to add a Hadoop File System (HDFS) resource. Adding an HDFS resource allows you import metadata from CSV, XML, and JSON files.
The General tab includes the following properties:
Property
Description
Name Node URI 1
URI to the active HDFS NameNode. The active HDFS NameNode manages all the client operations in the cluster.
Name Node URI 2
URI to the secondary HDFS NameNode. The secondary HDFS NameNode stores modifications to HDFS as a log file appended to a native file system file.
HDFS Service Name
HDFS service name.
User Name/User Principal
User name to connect to HDFS. Specify the Kerberos Principal if the cluster is enabled for Kerberos.
Source Directory
The source location from where metadata must be extracted.
Kerberos Cluster
Select Yes if the cluster is enabled for Kerberos. If the cluster is enabled for Kerberos, provide the following details:
  • - HDFS Service Principal: the service principal name of HDFS service.
  • - Keytab File: the path to the Kerberos Principal keytab file. Make sure that the keytab file is present at the specified location on Informatica domain host and cluster hosts of the Catalog Service.
The Metadata Load Settings tab includes the following properties:
Property
Description
File Types
Select any or all of the following file types from which you want to extract metadata:
  • - CSV
  • - XML
  • - JSON
First Level Directory
Specifies that all the directories must be selected. If you want specific directories to be selected, use the Select Directory option. This option is disabled if you had selected the Include Subdirectories option on the General tab.
Include Subdirectory
Type the required directories in the text box or click Select... to choose the required directories. This option is disabled if you had selected the Include Subdirectories option on the General tab or the Select all Directories option listed above.
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.
You can enable data discovery for an HDFS resource. See the Enable Data Discovery section for more information.

Running an HDFS Resource on Kerberos-enabled Cluster

If you want to run an HDFS resource scanner on a Kerberos-enabled cluster, perform the following steps:
  1. 1. Copy the krb.conf file to the following location: <Install Directory>/data/ldmbcmev/Informatica/LDM20_309/source/services/shared/security/krb5.conf
  2. 2. Copy the krb.conf file to /etc location on all the clusters where the Catalog Service is running.
  3. 3. Copy the keytab file to the /opt directory in the following locations:
  4. 4. Add the machine details of the kdc host in the etc/hosts location of the domain machine and the cluster machine where the Catalog Service is running.

Amazon Redshift Resource Connection Properties

You must complete the prerequisites listed before configuring the properties for an Amazon Redshift resource:

Prerequisites

  1. 1. Download the redshiftJars.zip file and copy the file to the <INFA_HOME>/services/CatalogService/ScannerBinaries directory.
  2. 2. Open the <INFA_HOME>/services/CatalogService/ScannerBinaries/CustomDeployer/scannerDeployer.xml file and add the path to the redshift.jar file.
  3. 3. If you want to enable profiling for the Amazon Redshift resource, copy the redshift.jar file to the <INFA_HOME>/connectors/thirdparty/informatica.amazonredshift/common/ directory.
  4. 4. Restart the Informatica domain.

Resource Configuration Properties

The General tab includes the following properties:
Property
Description
User
The user name used to access the database.
Password
The password associated with the user name.
Host
Host name or IP address of Amazon Redshift service.
Port
Amazon Redshift server port number. Default is 5439.
Database
The name of the database instance.
The Metadata Load Settings tab includes the following properties:
Property
Description
Import System Objects
Select this option to specify that the system objects must be imported.
Schema
Click Select... to specify the Amazon Redshift schemas that you want to import. You can use one of the following options from the Select Schema dialog box to import the schemas:
  • - Select from List: Use this option to select the required schemas from a list of available schemas.
  • - Select using regex: Provide an SQL regular expression to select schemas that match the expression.
S3 Bucket Name
Provide a valid Amazon S3 bucket name for the Amazon Redshift data source. You must provide this value if you want to enable profiling for Amazon Redshift. If you do not want to enable profiling, retain the default value.
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.
You can enable data discovery for an Amazon Redshift resource. See the Enable Data Discovery section for more information.

MicroStrategy Resource Connection Properties

The following tables list the properties that you must configure to add a MicroStrategy resource:
The General tab includes the following properties:
Property
Description
Agent URL
URL to the Live Data Map agent that runs on a Microsoft Windows Server.
Version
Select the version of MicroStrategy from the drop-down list. You can select the Auto detect option if you want Live Data Map to automatically detect the version of the MicroStrategy resource.
Project Source
Name of the MicroStrategy project source to which you want to connect.
Login User
The user name used to connect to the project source.
Login Password
The password associated with the user name.
Default Language
Specify the language to be used while importing metadata from the resource.
Import Schema Only
Select this option to import the project schema without the reports and documents.
Data Model Tables Design Level
Select one of the following options to specify the design for the imported tables:
  • - Physical: the imported tables appear in the physical view of the model.
  • - Logical and Physical: the imported tables appear in the logical and physical view of the model.
Incremental Import
Select this option to import only the changes from the source. Clear this option to import the complete source every time.
Project(s)
Select the names of the projects to which you want to connect from the project source.
Auto Assign Connections
Specifies to automatically assign the connection.
The Metadata Load Settings tab includes the following properties:
Property
Description
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.

Business Glossary Classification Resource Type Properties

Business Glossary contains online glossaries of business terms and policies that define important concepts within an organization. Configure a Business Glossary Classification resource type to extract metadata from Business Glossary.
The following table describes the connection properties for the Business Glossary Classification resource type:
Property
Description
Username
Name of the user account used that connects to the Analyst tool.
Password
Password for the user account that connects to the Analyst tool.
Host
Name of the Analyst tool business glossary from which you want to extract metadata. Each resource can extract metadata from one business glossary.
Port
Port number on which the Analyst tool runs.
Namespace
Name of the security domain to which the Analyst tool user belongs.
If the domain uses LDAP authentication or Kerberos authentication, enter the security domain name. Otherwise, enter Native.
Enable Secure Communication
Enable secure communication from the Analyst tool to the Analyst Service.
Import Published Content Only
Select this option to specify that you want to import only the published content. If you do not select this option, Live Data Map imports all content.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Glossary
Name of the business glossary resource that you want to import.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.

IBM Cognos Connection Properties

The following tables describe the IBM Cognos connection properties:
The General tab includes the following properties:
Property
Description
Agent URL
URL to the Live Data Map agent that runs on a Microsoft Windows Server.
Version
Indicates the Cognos server version.
Dispatcher URL
URL used by the framework manager to send requests to Cognos.
Namespace
Defines a collection of user accounts from an authentication provider.
User
User name used to connect to the Cognos server.
Password
Password for the user account to connect to the Cognos server.
Add Dependent Objects
Use to import dependent objects to the selection. Selecting this option requires a complete scan of report dependencies on the Cognos server. You can select any of the following options for this property:
  • - None: only imports the selected Cognos objects.
  • - Packages referenced by selected reports: imports the reports and associated source packages.
  • - All: imports the source packages when a report is selected and imports the dependent reports when a source package is selected.
Incremental Import
You can specify one of the following values for this property:
  • - True: Imports only the changes in the source.
  • - False: Imports the complete source every time.
Folder Representation
Specifies how the folders from Cognos framework manager must be represented. You can select from the following options:
  • - Ignore: ignores the folders.
  • - Flat: represents the folders as diagrams, but does not retain the hierarchy.
  • - Hierarchical: represents folders as diagrams and retains the hierarchy.
Transformer Import Configuration
The XML file that describes mappings between Cognos Content Manager data sources and PowerPlay Transformer models.
Worker Threads
Number of worker threads required to retrieve metadata asynchronously.
Auto Assign Connections
Specifies to automatically assign the connection.
The Metadata Load Settings tab includes the following properties:
Property
Description
Content Browsing Mode
Specifies the content to be retrieved while searching the Cognos repository. You can select any of the following options:
  • - Packages Only: retrieves the packages and folders and does not retrieve the reports.
  • - Connections Only: retrieves the list of connections.
  • - All: retrieves the packages, folders, queries, and reports.
  • - Content: allows you to reduce the scope of import to a smaller set of objects than the whole set of objects on the server.
Content
Specifies the hierarchy for the content objects.
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.

Tableau Server Properties

The following tables describe the Tableau Server properties:
The General tab includes the following properties:
Property
Description
Server
The host name or the IP address where the Tableau server runs.
Site
Specify the site if the Tableau server has multiple sites installed. The value is case sensitive.
Username
The user name to connect to the Tableau server.
Password
The password associated with the user name.
Incremental Import
You can specify one of the following values for this property:
  • - True: Imports only the changes in the source.
  • - False: Imports the complete source every time.
Worker Threads
Number of worker threads required to retrieve metadata asynchronously.
Cache
Path to the folder with the Tableau repository cache.
Auto Assign Connections
Specifies to automatically assign the connection.
The Metadata Load Settings tab includes the following properties:
Property
Description
Group By
Specify to group workbooks in the following categories:
  • - Project
  • - My Workbooks
  • - All Workbooks
Repository Objects
Imports the repository objects such as workbooks and data sources. For any workbooks, the dependent data sources are also imported.
Memory
Specifies the memory required to run the scanner job. Select one of the following values based on the data set size imported:
  • - Low
  • - Medium
  • - High
See the Tuning Live Data Map Performance How-to-Library article for more information about memory values.

Cloudera Navigator Connection Properties

Configure the connection properties when you create or edit a Cloudera Navigator resource.
The following table describes the connection properties:
Property
Description
Navigator URL
URL of the Cloudera Navigator Server.
User
Name of the user account that connects to Cloudera Navigator.
Password
Password for the user account that connects to Cloudera Navigator.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Hive Database
Name of the Hive database or a schema from where you want to import a table.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.

Hive Resource Prerequisites and Connection Properties

Complete the prerequisites and configure the connection properties when you create a Hive resource.

Prerequisites

Complete the following prerequisites before you configure the connection properties for a Hive resource:
  1. 1. Create a <Filename>.zip file based on the Hadoop distribution type you require and include the driver .jar files for the distribution type in the .zip file.
  2. You can copy the required driver .jar files from the following locations, based on the Hadoop distribution type you require:
  3. 2. Update the scannerDeployer.xml file present under the $INFA_HOME/services/CatalogService/ScannerBinaries/CustomDeployer directory, based on the version of Live Data Map you are using:
  4. Note: In the lines that you add to scannerDeployer.xml, substitute <Filename>.zip with the name of the .zip file you created.

Connection Properties

The following table describes the connection properties:
Property
Description
Hadoop Distribution
Select one of the following Hadoop distribution types for the Hive resource:
  • - Cloudera
  • - Hortonworks
  • - MapR
URL
JDBC connection URL used to access the Hive server.
User
The Hive user name.
Password
The password for the Hive user name.
Keytab file
Path to the keytab file if Hive uses Kerberos for authentication.
User proxy
The proxy user name to be used if Hive uses Kerberos for authentication.
Kerberos Configuration File
Specify the path to the Kerberos configuration file if you use Kerberos-based authentication for Hive.
Enable Debug for Kerberos
Select this option to enable debugging options for Kerberos-based authentication.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Schema
Click Select... to specify the Hive schemas that you want to import. You can use one of the following options from the Select Schema dialog box to import the schemas:
  • - Select from List: Use this option to select the required schemas from a list of available schemas.
  • - Select using regex: Provide an SQL regular expression to select schemas that match the expression.
Table
Specify the name of the Hive table that you want to import. If you leave this property blank, Live Data Map imports all the Hive tables.
SerDe jars list
Specify the path to the Serializer/DeSerializer (SerDe) jar file list. You can specify multiple jar files by separating the jar file paths using a semicolon (;).
Worker Threads
Specify the number of worker threads to process metadata asynchronously. You can leave the value empty if you want Live Data Map to calculate the value. Live Data Map assigns a value between one and six based on the JVM architecture and number of available CPU cores.
You can use the following points to decide the value to use:
  • - You can provide a value that is greater than or equal to one and lesser than six to specify the number of worker threads required.
  • - If you specify an invalid value, Live Data Map shows a warning and uses the value one.
  • - If your machine has more memory, you can specify a higher value to process more metadata asynchronously.
  • Note: Specifying a higher value might impact performance of the system.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.
You can enable data discovery for a Hive resource. See the Enable Data Discovery section for more information.

Configure Hive Resource with Apache Knox Gateway

Live Data Map supports Knox if you configure Hive for Knox. Make sure that Informatica and Hive hosting service are installed on the same cluster.
Note: You cannot deploy Live Data Map on a cluster if all the services on the nodes are configured for Knox. Make sure that you configure Knox for Hive service and not for other services running on the nodes.

Informatica Platform Resource Type Properties

Create a resource based on the Informatica Platform resource type to extract metadata from the Model repository. You need to specify the Data Integration Service connection details when you configure the resource.
The following table describes the Data Integration Service connection properties:
Property
Description
Target version
The version number of the Informatica platform. You can choose any of the following Informatica versions:
  • - 9.6.1 HotFix 3
  • - 9.6.1 HotFix4
  • - 10.0
  • - 10.1
  • - 10.1.1
Domain Name
Name of the Informatica domain.
Data Integration Service Name
Name of the Data Integration Service.
Username
Username for the Data Integration Service connection.
Password
Password for the Data Integration Service connection.
Security Domain
Name of the LDAP security domain if the Informatica domain contains an LDAP security domain.
Host
Host name for the informatica domain.
Port
Port number of the Informatica domain.
Application Name
Name of the Data Integration Service application. Click Select... to select the name of the application from the Select Application Name dialog box.
Note: This property is applicable if you select Target version as 10.0, 10.1, or 10.1.1.
Param Set for Mappings in Application
Parameter set for mappings configured for the Data Integration Service application. Click Select... to select the parameter set from the Select Param Sets for Mappings in Application dialog box.
Note: This property is applicable if you select Target version as 10.0, 10.1, or 10.1.1.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Auto assign Connections
Specifies whether the connection must be automatically assigned.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.

PowerCenter Resource Type Properties

You can configure a PowerCenter resource type to extract metadata from PowerCenter repository objects. Use PowerCenter to extract data from multiple sources, transform the data according to business logic you build in the client application, and load the transformed data into file and relational targets.
The following table describes the properties for the PowerCenter resource type:
Property
Description
Gateway Host Name or Address
PowerCenter domain gateway host name or address.
Gateway Port Number
PowerCenter domain gateway port number.
Informatica Security Domain
LDAP security domain name if one exists. Otherwise, enter "Native."
Repository Name
Name of the PowerCenter repository.
Repository User Name
Username for the PowerCenter repository.
Repository User Password
Password for the PowerCenter repository.
PowerCenter Version
PowerCenter repository version.
PowerCenter Code Page
Code page for the PowerCenter repository.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Parameter File
Specify the parameter file that you want to attach from a local system.
Auto assign Connections
Specifies whether Live Data Map assigns the connection is automatically.
Repository subset
Enter the file path list separated by semicolons for the Informatica PowerCenter Repository object.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.

IBM DB2 Resource Type Properties

You can configure an IBM DB2 resource type to extract metadata from IBM DB2 databases.
The following table describes the connection properties for the IBM DB2 resource type:
Property
Description
User
Name of the user account that connects to IBM DB2 database.
Password
Password for the user account that connects to IBM DB2 database.
Host
Fully qualified host name of the machine where IBM DB2 database is hosted.
Port
Port number for the IBM DB2 database.
Database
The DB2 connection URL used to access metadata from the database.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Import system objects
Specifies the system objects to import.
Schema
Specifies a list of database schema.
Import stored procedures
Specifies the stored procedures to import.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.
You can enable data discovery for an IBM DB2 resource. See the Enable Data Discovery section for more information.

IBM DB2 for z/OS Resource Type Properties

Configure an IBM DB2 for z/OS resource type to extract metadata from IBM DB2 for z/OS databases.
The following table describes the properties for the IBM DB2 for z/OS resource type:
Property
Description
Location
Node name in the dbmover.cfg file on the machine where the Catalog Service runs that points to the PowerExchange Listener on the z/OS system.
Note: Live Data Map uses PowerExchange for DB2 for z/OS to access metadata from z/OS subsystems.
User
Name of the user account that connects to IBM DB2 for z/OS database.
Password
Password for the user account that connects to IBM DB2 for z/OS database.
Encoding
Code page for the IBM DB2 for z/OS subsystem.
Sub System ID
Name of the DB2 subsystem.
The following table describes the Additional property for source metadata settings on the Metadata Load Settings tab:
Property
Description
Schema
Specifies a list of database schema.
You can enable data discovery for an IBM DB2 for z/OS resource. See the Enable Data Discovery section for more information.

IBM Netezza Resource Type Properties

You need to set up multiple configuration properties when you create a resource to extract metadata from IBM Netezza databases.
The following table describes the connection properties for the IBM Netezza resource type:
Property
Description
Host
Host name or IP address of the machine where the database management server runs.
Port
Port number for the Netezza database.
User
Name of the user account used to connect to the Netezza database.
Password
Password for the user account used to connect to the Netezza database.
Database
ODBC data source connect string for a Netezza database. Enter the data source name of the Netezza DSN if you created one.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Schema
Specifies a list of semicolon-separated database schema.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.
You can enable data discovery for an IBM Netezza resource. See the Enable Data Discovery section for more information.

JDBC Resource Type Properties

You can use a JDBC connection to access tables in a database.
The following table describes the connection properties for the JDBC resource type:
Property
Description
Driver class
Name of the JDBC driver class.
URL
Connection string to connect to the database.
User
Database username.
Password
Password for the database user name.
Note: To extract metadata from multiple schemas of a source using the JDBC resource type, you can specify semicolon-separated schema names when you create the resource. You can type in the multiple schema names in the Schema field on the Metadata Load Settings page.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Catalog
Catalog name.
Note: You cannot use Catalog option for JDBC or ODBC sources.
Schema
Specifies a list of schemas to import.
Case sensitivity
Select one of the following options to specify if the database is configured for case sensitivity:
  • - Auto. If you select this option, the resource uses the JDBC API and attempts to find if the database is configured for case sensitivity. If the JDBC API is not available for the resource, the resource uses the case insensitive mode.
  • - Case Sensitive. Select this option to specify that the database is configured for case sensitivity.
  • - Case Insensitive. Select this option to specify that the database is configured as case insensitive.
View definition extracting SQL
Specifies the database specific SQL query to retrieve the view definition text.
Synonyms lineage SQL
Specifies the database specific SQL query to retrieve the synonym lineage. The following are the two columns that the query returns:
  • - Full Synonym Name
  • - Full Table Name
Optional Scope
Specifies the database object types to import, such as Tables and Views, Indexes, and Procedures.
Specify a list of optional database object types that you want to import. The list can have zero or more database object types, which are separated by semicolons. For example, Keys and Indexes, and Stored Procedures.
Import stored procedures
Specifies the stored procedures to import.
The default value is True or False whatever the case might be.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.
You can enable data discovery for a JDBC resource. See the Enable Data Discovery section for more information.

SQL Server Resource Type Properties

You can configure an SQL Server resource type to extract metadata from Microsoft SQL Server databases. Make sure that you configure the VIEW DEFINITION permission for the SQL database and configure the SELECT permission for the sys.sql_expression_dependencies of the database.
The following table describes the properties for the SQL Server resource type:
Property
Description
User
Name of the SQL Server user account that connects to the Microsoft SQL Server database. The Catalog Service uses SQL Server authentication to connect to the Microsoft SQL Server database.
Password
Password for the user account that connects to the Microsoft SQL Server database.
Host
Host name of the machine where Microsoft SQL Server runs.
Port
Port number for the SQL Server database engine service.
Database
Name of the SQL Server database.
Instance
SQL Server instance name.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Import system objects
Specifies the system objects to import.
The default value is True or False whatever the case might be.
Schema
Specifies a list of semicolon-separated database schema.
Import stored procedures
Specifies the stored procedures to be imported.
The default value is True or False whatever the case might be.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.
You can enable data discovery for an SQL Server resource. See the Enable Data Discovery section for more information.

Oracle Resource Type Properties

Configure an Oracle resource type to extract metadata from Oracle databases.
The following table describes the properties for the Oracle resource type:
Property
Description
User
Name of the user account that connects to the Oracle database.
Password
Password for the user account that connects to the Oracle database.
Host
Fully qualified host name of the machine where the Oracle database is hosted.
Port
Port number for the Oracle database engine service.
Service
Unique identifier or system identifier for the Oracle database server.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Import system objects
Specifies the system objects to import.
The default value is True or False whatever the case might be.
Schema
Specifies a list of semicolon-separated database schema.
Import stored procedures
Specifies the stored procedures to import.
The default value is True or False whatever the case might be.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.
You can enable data discovery for an Oracle resource. See the Enable Data Discovery section for more information.

Sybase Resource Type Properties

You can configure a Sybase resource type to extract metadata from Sybase databases.
The following table describes the properties for the Sybase resource type:
Property
Description
Host
Host name of the machine where Sybase database is hosted.
Port
Port number for the Sybase database engine service.
User
Database user name.
Password
The password for the database user name.
Database
Name of the database.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Schema
Specify a list of database or scheme to import.
Imported stored procedures
Specifies the stored procedures to import.
The default value is True or False whatever the case might be.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.
You can enable data discovery for a Sybase resource. See the Enable Data Discovery section for more information.

Teradata Resource Type Properties

Teradata is one of the ETL resource types in Live Data Map. Configure a Teradata resource type to extract metadata from Teradata databases.
The following table describes the properties for the Teradata resource type:
Property
Description
User
Name of the user account that connects to the Teradata database.
Password
Password for the user account that connects to the Teradata database.
Host
Fully qualified host name of the machine where the Teradata database is hosted.
Note: To connect to a Teradata resource using the LDAP server, you must specify /LOGMECH=LDAP to the host name.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Import system objects
Specifies the system objects to import.
The default value is True or False whatever the case might be.
Schema
Specifies a list of semicolon-separated database schema.
Import stored procedures
Specifies the stored procedures to import.
The default value is True or False whatever the case might be.
Fetch Views Data Types
Specifies that views data type must be imported.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.
You can enable data discovery for a Teradata resource. See the Enable Data Discovery section for more information.

SAP BusinessObjects Resource Type Properties

You can configure an SAP BusinessObjects resource type to extract metadata from SAP BusinessObjects. SAP BusinessObjects is a business intelligence tool that includes components for performance management, planning, reporting, analysis, and enterprise information management.
The following table describes the properties for the BusinessObjects resource type:
Property
Description
Agent URL
Host name and port number of the Live Data Map agent that runs on a Microsoft Windows Server.
Version
Version of the SAP Business Objects repository.
System
Name of the BusinessObjects repository. For BusinessObjects 11.x and 12.x, specify the name of the BusinessObjects Central Management Server.
Specify the server name in the following format:
<server name>:<port number>
If the Central Management Server is configured on a cluster, specify the cluster name in the following format:
<host name>:<port>@<cluster name>
Default port is 6400.
Note: If the version of the BusinessObjects repository is 14.0.6, do not specify a port number in the repository name. If you specify the port number, Live Data Map cannot extract the Web Intelligence reports.
Authentication mode
The authentication mode for the user account that logs in to the BusinessObjects repository.
Specify one of the following values:
  • - Enterprise. Log in using the BusinessObjects Enterprise authentication mode.
  • - LDAP. Log in using LDAP authentication configured to BusinessObjects.
Default is Enterprise.
User Name
User name to log in to the BusinessObjects repository.
Password
Password of the user account for the BusinessObjects repository.
Incremental import
Loads changes after the previous resource load or loads complete metadata.
Specify one of the following values:
  • - True. Loads only the recent changes.
  • - False. Performs a complete load of the metadata.
Add dependent objects
Choose the documents that depend on the universe you selected. Specify one of the following values:
  • - True. Imports the documents that depend on the specified universe.
  • - False. Ignores the documents that depend on the specified universe.
Note: Dependency information is retrieved from the Business Objects repository metadata cache. If the Live Data Map load does not reflect modified or moved reports, refresh the cache by loading these reports and refreshing the queries.
Add specific objects
Specifies additional objects to the universe. Specify one of the following values:
  • - None. Ignores all objects.
  • - Universe independent Documents. Imports documents that do not depend on any universe.
Default is none.
Crystal CORBA port
Specifies the client port number on which the Crystal SDK communicates with the Report Application Server (RAS).
The RAS server uses the port to send metadata to the local client computer. If you do not specify a port, the server randomly selects a port for each execution.
Class representation
Controls how the import of the tree structure of classes and sub classes occur.
The Live Data Map agent imports each class containing objects as a dimension or as a tree of packages. Specify one of the following values:
  • - As a flat structure. Creates no packages.
  • - As a simplified flat structure. Creates a package for each class with a sub class.
  • - As a full tree structure. Creates a package for each class.
Default is As a flat structure.
Worker Threads
Number of worker threads that the Live Data Map agent uses to extract metadata asynchronously. Leave blank or enter a positive integer value.
If left blank, the Live Data Map agent calculates the number of worker threads. The Live Data Map agent uses the JVM architecture and number of available CPU cores on the Live Data Map agent machine to calculate the number of threads. If you specify a value that is not valid, the Live Data Map agent uses one worker thread.
Reduce the number of worker threads if the Live Data Map agent generates out-of-memory errors during metadata extraction. Increase the number of worker threads if the Live Data Map agent machine has a large amount of available memory, for example, 10 GB or more. If you specify too many worker threads, performance can decrease.
Default is blank.
Auto Assign Connections
Choose to automatically assign the database schemas to the resource that you create for SAP BusinessObjects source.
The following table describes the Additional and Advanced properties for source metadata settings on the Metadata Load Settings tab:
Property
Description
Repository browsing mode
Specifies the available objects in the SAP BusinessObjects repository. select one of the following options:
  • - Universes only. Imports the metadata for the SAP BusinessObjects universes.
  • - Connection only. Imports the metadata for the connections to the databases that are published to the SAP BusinessObjects repository.
  • - All. Imports both the universes and connections from the SAP BusinessObjects repository.
Repository subset
Specifies the objects stored in a remote SAP BusinessObjects repository.
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.

Salesforce Resource Type Properties

Use a Salesforce connection to connect to a Salesforce object. The Salesforce connection is an application connection type.
The following table describes the connection properties for the Salesforce resource type:
Property
Description
Username
Salesforce username.
Password
Password for the Salesforce user name.
Service_URL
URL of the Salesforce service that you want to access.
The following table describes the Advanced property for source metadata settings on the Metadata Load Settings tab:
Property
Description
Memory
Specify the memory value required to run a scanner job.
Specify one of the following memory values:
  • - Low
  • - Medium
  • - High
Note: For details about the memory values, see the Tuning Live Data Map Performance How-To Library article.
You can enable data discovery for a Salesforce resource. See the Enable Data Discovery section for more information.