Create and Configure the Data Integration Service
The Data Integration Service is an application service that performs data integration jobs for the Live Data Map tools, such Informatica Administrator, Enterprise Information Catalog, and Informatica Live Data Map Administrator.
When you run scans on resources and view the metadata and profiling statistics in Live Data Map, the client tool sends requests to the Data Integration Service to perform the data integration jobs.
Create the Data Integration Service
Use the service creation wizard in the Administrator tool to create the service.
Before you create the Data Integration Service, verify that you have created and enabled the Model Repository Service. You also need to verify that you have created a Model repository user that the Data Integration Service can use to access the Model Repository Service.
1. In the Administrator tool, click the Manage tab.
2. Click the Services and Nodes view.
3. In the Domain Navigator, select the domain.
4. Click Actions > New > Data Integration Service.
The New Data Integration Service wizard appears.
5. On the New Data Integration Service - Step 1 of 14 page, enter the following properties:
Property | Description |
---|
Name | Name of the service. The name is not case sensitive and must be unique within the domain. It cannot exceed 128 characters or begin with @. It also cannot contain spaces or the following special characters: ` ~ % ^ * + = { } \ ; : ' " / ? . , < > | ! ( ) ] [ |
Description | Description of the service. The description cannot exceed 765 characters. |
Location | Domain and folder where the service is created. Click Browse to choose a different folder. You can move the service after you create it. |
License | License object that allows use of the service. |
Assign | Select Node to configure the service to run on a node. If your license includes grid, you can create a grid and assign the service to run on the grid after you create the service. |
Node | Node on which the service runs. |
Backup Nodes | If your license includes high availability, nodes on which the service can run if the primary node is unavailable. |
Model Repository Service | Model Repository Service to associate with the service. |
Username | User name that the service uses to access the Model Repository Service. Enter the Model repository user that you created. Not available for a domain with Kerberos authentication. |
Password | Password for the Model repository user. Not available for a domain with Kerberos authentication. |
Security Domain | LDAP security domain for the Model repository user. The field appears when the Informatica domain contains an LDAP security domain. Not available for a domain with Kerberos authentication. |
6. Click Next.
The New Data Integration Service - Step 2 of 14 page appears.
7. Enter the HTTP port number to use for the Data Integration Service.
8. Accept the default values for the remaining security properties. You can configure the security properties after you create the Data Integration Service.
9. Select Enable Service.
The Model Repository Service must be running to enable the Data Integration Service.
10. Verify that the Move to plugin configuration page is not selected.
11. Click Next.
The New Data Integration Service - Step 3 of 14 page appears.
12. Set the Launch Job Options property to one of the following values:
- - In the service process. Configure when you run SQL data service and web service jobs. SQL data service and web service jobs typically achieve better performance when the Data Integration Service runs jobs in the service process.
- - In separate local processes. Configure when you run mapping, profile, and workflow jobs. When the Data Integration Service runs jobs in separate local processes, stability increases because an unexpected interruption to one job does not affect all other jobs.
If you configure the Data Integration Service to run on a grid after you create the service, you can configure the service to run jobs in separate remote processes.
13. Accept the default values for the remaining execution options and click Next.
The New Data Integration Service - Step 4 of 14 page appears.
14. If you created the data object cache database for the Data Integration Service, click Select to select the cache connection. Select the data object cache connection that you created for the service to access the database.
15. Accept the default values for the remaining properties on this page and click Next.
The New Data Integration Service - Step 5 of 14 page appears.
16. For optimal performance, enable the Data Integration Service modules that you plan to use.
The following table lists the Data Integration Service modules that you can enable:
Module | Description |
---|
Web Service Module | Runs web service operation mappings. |
Mapping Service Module | Runs mappings and previews. |
Profiling Service Module | Runs profiles and scorecards. |
SQL Service Module | Runs SQL queries from a third-party client tool to an SQL data service. |
Workflow Orchestration Service Module | Runs workflows. |
17. Click Next.
The New Data Integration Service - Step 6 of 14 page appears.
You can configure the HTTP proxy server properties to redirect HTTP requests to the Data Integration Service. You can configure the HTTP configuration properties to filter the web services client machines that can send requests to the Data Integration Service. You can configure these properties after you create the service.
18. Accept the default values for the HTTP proxy server and HTTP configuration properties and click Next.
The New Data Integration Service - Step 7 of 14 page appears.
The Data Integration Service uses the result set cache properties to use cached results for SQL data service queries and web service requests. You can configure the properties after you create the service.
19. Accept the default values for the result set cache properties and click Next.
The New Data Integration Service - Step 8 of 14 page appears.
20. If you created the profiling warehouse database for the Data Integration Service, select the Profiling Service module.
21. If you created the workflow database for the Data Integration Service, select the Workflow Orchestration Service module.
22. Verify that the remaining modules are not selected.
You can configure properties for the remaining modules after you create the service.
23. Click Next.
The New Data Integration Service - Step 11 of 14 page appears.
24. If you created the profiling warehouse database for the Data Integration Service, click Select to select the database connection. Select the profiling warehouse connection that you created for the service to access the database.
25. Select whether or not content exists in the profiling warehouse database.
If you created a new profiling warehouse database, select No content exists under specified connection string.
26. Click Next.
The New Data Integration Service - Step 12 of 14 page appears.
27. Accept the default values for the advanced profiling properties and click Next.
The New Data Integration Service - Step 14 of 14 page appears.
28. If you created the workflow database for the Data Integration Service, click Select to select the database connection. Select the workflow database connection that you created for the service to access the database.
29. Click Finish.
The domain creates and enables the Data Integration Service.
After you create the service through the wizard, you can edit the properties or configure other properties.
After You Create the Data Integration Service
After you create the Data Integration Service, perform the following tasks:
- •Verify the host file configuration on Linux.
- •Create other application services.
Verify the Host File Configuration on Linux
If you configured the Data Integration Service on Linux to launch jobs as separate processes, verify that the host file on the node that runs the service contains a localhost entry. Otherwise, jobs fail when the Launch Jobs as Separate Processes property for the Data Integration Service is enabled.
Create Other Services (DIS)
After you create the Data Integration Service, create the application services that depend on the Data Integration Service.
Create the dependent services in the following order:
- 1. Informatica Cluster Service if you choose internal Hadoop cluster for deploying Live Data Map.
- 2. Catalog Service
- 3. Content Management Service