Application Services
Application services represent server-based functionality.
Application services include services that you create and system services that are created for you when you create the domain. A system service can have a single instance in the domain.
Application services include the following services:
- •Analyst Service
- •Catalog Service
- •Cluster Service
- •Data Integration Service
- •Model Repository Service
When you configure an application service, you designate a node to run the service process. When a service process runs, the Service Manager assigns a port number from the range of port numbers assigned to the node.
The service process is the runtime representation of a service running on a node. The service type determines how many service processes can run at a time.
If you have the high availability option, you can run a service on multiple nodes. Designate the primary node to run the service. All other nodes are back-up nodes for the service. If the primary node is not available, the service runs on a back-up node. You can subscribe to alerts to receive notification in the event of a service process failover.
If you do not have the high availability option, configure a service to run on one node. If you assign the service to multiple nodes, then the service will not start.
Analyst Service
The Analyst Service is an application service that runs the Informatica Analyst application in the Informatica domain. The Analyst Service manages the connections between service components and the users that log in to Informatica Analyst. The Analyst Service connects to a Data Integration Service, a Model Repository Service, a Metadata Manager Service, and a Search Service. The Analyst Service also specifies a flat file cache directory and a directory for business glossary export files.
When you configure the Analyst Service, connect it to a Data Integration Service to run profiles, scorecards, and mapping specifications. You can also connect the Analyst Service to a Data Integration Service that runs Human tasks. Connect the Analyst Service to a Model Repository Service to identify a Model repository.
Specify a flat file cache directory to store temporary data from flat files that you upload. Specify a business glossary directory to stores temporary files that you export from the Business Glossary.
Cluster Service
The Informatica Cluster Service is an application service that runs and manages all the associated services that are required to run Enterprise Data Catalog in the Informatica domain. The associated services include Mongo DB, Nomad, Solr, PostgreSQL, and ZooKeeper.
The following table lists the supported methods and algorithms:
Method | Algorithm |
---|
Key exchange | - - diffie-hellman-group-exchange-sha1
- - diffie-hellman-group1-sha1
- - diffie-hellman-group14-sha1
- - diffie-hellman-group-exchange-sha256
- - ecdh-sha2-nistp256
- - ecdh-sha2-nistp384
- - ecdh-sha2-nistp521
|
Cipher | - - blowfish-cbc
- - 3des-cbc
- - aes128-cbc
- - aes192-cbc
- - aes256-cbc
- - aes128-ctr
- - aes192-ctr
- - aes256-ctr
- - 3des-ctr
- - arcfour
- - arcfour128
- - arcfour256
|
MAC | - - hmac-md5
- - hmac-sha1
- - hmac-md5-96
- - hmac-sha1-96
|
Host key type | - - ssh-dss
- - ssh-rsa
- - ecdsa-sha2-nistp256
- - ecdsa-sha2-nistp384
- - ecdsa-sha2-nistp521
|
Catalog Service
The Catalog Service is an application service that runs Enterprise Data Catalog in the Informatica domain. The Catalog Service manages the connections between service components and the users that have access to Enterprise Data Catalog and Catalog Administrator.
The catalog represents an indexed inventory of all the configured data assets in an enterprise. You can find metadata and statistical information, such as profile statistics, data asset ratings, data domains, and data relationships, in the catalog.
Connect the MetaDex Server to the Catalog Service
Before you run MetaDex configurations, configure the MetaDex server to connect to the Catalog Service.
1. Log in to the MetaDex tool.
2. On the header of the MetaDex tool, click Administration > EDC Instance.
The EDC Instance page appears with the predefined connection variables and the information on the variable origin, as shown in the following image:
3. Click the Edit icon to the right of each variable.
The Edit global variable dialog box appears.
4. Enter the values for the following variables:
Name | Description |
---|
EDC_URL | The host name and port number of the Catalog Service in the following format: <host>:<port> |
EDC_USER | The user name configured for the catalog administrator user. |
EDC_PASSWORD | The password configured for the catalog administrator user. When you add the password, click the Hide password icon. |
5. Click Save.
The Edit global variable dialog box closes.
6. Test and reload the license.
Enable Data Preview, Provisioning and Preparation
To enable data preview, provisioning, and preparation, configure custom properties for the Catalog Service in the Informatica Administrator console.
Configure the following custom properties for the Catalog Service:
Name | Value |
---|
LdmCustomOptions.enableDataProvision | true |
LdmCustomOptions.provision.ics.master.app.url | Informatica Intelligent Cloud Services Organizations URL |
LdmCustomOptions.dataprep.app.url | Enterprise Data Preparation Service application URL |
Note: If you do not provide a valid Informatica Intelligent Cloud Services Organizations URL for the LdmCustomOptions.provision.ics.master.app.url custom property, Informatica Administrator console considers the following default Informatica Intelligent Cloud Services URL: https://dm-us.informaticacloud.com/.
Data Asset Analytics Repository Database Requirements
The Data Asset Analytics repository database stores the analytical information collected from the catalog. The analytical information is displayed as reports and statistical data in the Data Asset Analytics tab in Enterprise Data Catalog. You can access the Data Asset Analytics tab after you configure Data Asset Analytics in the Catalog Service and assign the required privileges for users to access the tab.
You can configure any of the following databases as the repository database for Data Asset Analytics:
- •Oracle
- •Microsoft SQL Server
- •PostgreSQL
Oracle Database Requirements
Perform the following steps before you configure Oracle as the repository database:
- 1. Create a user name and password to access the Oracle database using the following command: CREATE USER <Database schema>
- 2. Run the following commands to provide the required user privileges:
GRANT CONNECT TO <Database schema>;
GRANT RESOURCE TO <Database schema>;
GRANT CREATE VIEW TO <Database schema>;
GRANT CREATE MATERIALIZED VIEW TO <Database schema>;
GRANT UNLIMITED TABLESPACE TO <Database schema>;
Alternatively, if you do not want to grant the unlimited tablespace privilege, run the command ALTER USER <Database schema> QUOTA <Size> ON <Tablespace name>;
Microsoft SQL Server Database Requirements
Perform the following steps before you configure Microsoft SQL Server as the repository database:
- 1. Use the dbo schema for SQL Server.
- 2. Run the following commands to create the database, the user name, password and assign the required privileges for the user:
USE master
GO
CREATE DATABASE <New database name>;
GO
use <New database name>
CREATE LOGIN <New login name> WITH PASSWORD = '<Password>';
CREATE USER <New user name> FOR LOGIN <New login name> WITH DEFAULT_SCHEMA = [dbo];
GRANT CREATE TABLE TO <New user name>;
GRANT CREATE VIEW TO <New user name>;
GRANT SELECT, ALTER, INSERT, DELETE, UPDATE on schema::dbo to <New user name> ;
GRANT REFERENCES to <New user name>;
PostgreSQL Database Requirements
Perform the following steps before you configure PostgreSQL as the repository database:
- 1. Use the public schema for PostgreSQL.
- 2. Create the credentials to access the database using the following command: CREATE USER <New user name> WITH PASSWORD '<Password>';
Note: Make sure that you choose a user name that is different from the name of the database schema.
- 3. Create the database and assign the ownership to the user name that you created. Use the following command to complete this step: CREATE DATABASE <NEW db NAME> owner=<New user name>;
- 4. After you connect to the database, run the command SHOW search_path. The command must return the value <User>, public.
Data Flow Analytics Repository Database Requirements
Data Flow Analytics connects to the PowerCenter data source to retrieve mappings and performs analyses on the mappings. The statistical information from the analyses appears in the Data Flow Analytics tab in the Enterprise Data Catalog tool. Create a database user account and configure the required read privileges for the account. After you create the account, you can configure Data Flow Analytics for the PowerCenter resource using the Catalog Administrator tool.
Use the following guidelines when you set up the database user account for the PowerCenter repository.
Oracle Database Requirements
Perform the following steps before you configure Oracle as the PowerCenter data source for Data Flow Analytics:
- 1. Create a user name and password to access the database using the CREATE USER <Database schema> IDENTIFIED BY <database schema> command.
- 2. Run the following commands to provide the required privileges to the database user:
GRANT CONNECT TO <Database schema>;
GRANT SELECT ON <PowerCenter repository schema>.OPB_WFLOW_RUN TO <Database schema>;
GRANT SELECT ON <PowerCenter repository schema>.OPB_SUBJECT TO <Database schema>;
- 3. Run the following commands to create synonyms for the tables to query the workflow execution details:
CREATE synonym <Database schema>.OPB_WFLOW_RUN FOR <PCRS schema>.OPB_WFLOW_RUN;
CREATE synonym <Database schema>.OPB_SUBJECT FOR <PCRS schema>.OPB_SUBJECT;
Microsoft SQL Server Database Requirements
Perform the following steps before you configure Microsoft SQL Server as the PowerCenter data source for Data Flow Analytics:
- 1. Run the following commands to create a user name and password to access the database:
USE <PowerCenter repository name>
CREATE LOGIN <New login name> WITH PASSWORD='<Password>';
CREATE USER <New user name> FOR LOGIN <New login name>;
- 2. Run the following commands to provide the required privileges to the database user:
GRANT SELECT ON OPB_SUBJECT TO <New user name>;
GRANT SELECT ON OPB_WFLOW_RUN TO <New user name>
IBM DB2 Database Requirements
Perform the following steps before you configure IBM DB2 as the PowerCenter data source for Data Flow Analytics:
- 1. Log in to the machine that hosts the IBM DB2 database as a root user and create a new user account on the machine.
- 2. Run the following commands to provide the required privileges to the database user:
GRANT CONNECT ON <PowerCenter repository name> TO USER <Database user name>;
GRANT SELECT ON TABLE <PowerCenter repository name>."OPB_WFLOW_RUN" TO USER <Database user name>;
GRANT SELECT ON TABLE <PowerCenter repository name>."OPB_SUBJECT" TO USER <Database user name>;
- 3. Run the following commands to create aliases for the tables to query the workflow execution details:
CREATE ALIAS <Database user name>.OPB_WFLOW_RUN FOR <PowerCenter repository name>.OPB_WFLOW_RUN;
CREATE ALIAS <Database user name>.OPB_SUBJECT FOR <PowerCenter repository name>.OPB_SUBJECT;
Sybase
Perform the following steps before you configure Sybase as the PowerCenter data source for Data Flow Analytics:
- 1. Log in to a Sybase client as an administrator and run the following commands to create a user name and password to access the database:
sp_addlogin <New login name>, <Password>
use <PowerCenter repository name>
sp_adduser <New login name>, <New user name>
- 2. Run the following commands to provide the required privileges to the database user:
GRANT SELECT TO <New user name>
GRANT CREATE VIEW TO <New user name>
- 3. Run the following commands to create views for the tables to query the workflow execution details:
CREATE VIEW <Database user name>.OPB_WFLOW_RUN as select * from <PowerCenter repository schema>.OPB_WFLOW_RUN
CREATE VIEW <Database user name>.OPB_SUBJECT as select * from <PowerCenter repository schema>.OPB_SUBJECT
PostgreSQL
Perform the following steps before you configure PostgreSQL as the PowerCenter data source for Data Flow Analytics:
- 1. Create a user name and password to access the database using the following command: CREATE USER <Database user name> WITH PASSWORD '<Database user name>'
- 2. Create a role using the CREATE ROLE <Role name> command.
- 3. Run the following commands to provide the required privileges to the user:
GRANT <role name> TO <Database user name>;
GRANT CONNECT ON DATABASE "<PowerCenter repository name>" to <Role name>;
GRANT USAGE ON SCHEMA <PowerCenter repository schema> TO <Role name>;
GRANT SELECT ON OPB_WFLOW_RUN TO <Role name>;
GRANT SELECT ON OPB_SUBJECT TO <Role name>;
Content Management Service
The Content Management Service is an application service that manages reference data. A reference data object contains a set of data values that Enterprise Data Catalog searches while performing data domain discovery on source data. Data domain discovery finds inferred business semantics based on column data. Examples include Social Security number, phone number, and credit card number.
The Content Management Service uses the Data Integration Service to run mappings to transfer data between reference tables and external data sources.
Data Integration Service
The Data Integration Service is an application service that performs data integration tasks for Informatica Analyst, Informatica Developer, and external clients. Data integration tasks include previewing data and running profiles, SQL data services, web services, and mappings.
When you start a command from the command line or an external client to run SQL data services and mappings in an application, the command sends the request to the Data Integration Service.
You can configure the Data Integration Service to run on the following domain objects:
- On nodes
- If your license includes high availability, you can configure the service to run on multiple nodes. By default, the service runs on the primary node. If the primary node is not available, the service runs on a back-up node. If the service process fails or the node becomes unavailable, the service fails over to another node. If your license does not include high availability, you can configure the service to run on one node.
- On a grid
- If your license includes grid, you can configure the service to run on a grid. A grid is an alias assigned to a group of nodes. The Data Integration Service dispatches jobs to available nodes assigned to the grid. When the Data Integration Service runs on a grid, the service remains available if a service process fails or a node becomes unavailable.
Model Repository Service
The Model Repository Service is an application service that manages the Model repository. The Model repository is a relational database that stores the metadata for projects created in Informatica Analyst and Informatica Developer. The Model repository also stores run-time and configuration information for applications that are deployed to a Data Integration Service.
You can configure the Model Repository Service to run on one node. The Model Repository Service is not a highly available service. However, you can run multiple Model Repository Services on the same node. If the Model Repository Service fails, it automatically restarts on the same node.