Data Engineering Administrator Guide > Introduction to Data Engineering Administration > Application Services
  

Application Services

Informatica uses application services in the Informatica domain to process data.
Use the Administrator tool to create connections, monitor jobs, and manage application services.
Data Engineering Integration uses the following application services:
Analyst Service
The Analyst Service runs the Analyst tool in the Informatica domain. The Analyst Service manages the connections between service components and the users that have access to the Analyst tool.
Data Integration Service
The Data Integration Service can process mappings in the native environment or push the mapping for processing to a compute cluster in a non-native environment. The Data Integration Service also retrieves metadata from the Model repository when you run a Developer tool mapping or workflow. The Analyst tool and Developer tool connect to the Data Integration Service to run profile jobs and store profile results in the profiling warehouse.
Mass Ingestion Service
The Mass Ingestion Service manages and validates mass ingestion specifications that you create in the Mass Ingestion tool. The Mass Ingestion Service deploys specifications to the Data Integration Service. When a specification runs, the Mass Ingestion Service generates ingestion statistics.
Metadata Access Service
The Metadata Access Service allows the Developer tool to import and preview metadata from a Hadoop cluster.
The Metadata Access Service contains information about the Service Principal Name (SPN) and keytab information if the Hadoop cluster uses Kerberos authentication. You can create one or more Metadata Access Services on a node. Based on your license, the Metadata Access Service can be highly available.
HBase, HDFS, Hive, and MapR-DB connections use the Metadata Access Service when you import an object from a Hadoop cluster. Create and configure a Metadata Access Service before you create HBase, HDFS, Hive, and MapR-DB connections.
Model Repository Service
The Model Repository Service manages the Model repository. The Model Repository Service connects to the Model repository when you run a mapping, mapping specification, profile, or workflow.
REST Operations Hub
The REST Operations Hub Service is an application service in the Informatica domain that exposes Informatica product functionality to external clients through REST APIs.

Data Integration Service Process

The Data Integration Service starts one or more Data Integration Service processes to manage requests to run mapping jobs in the Hadoop environment.
When you run mappings in the Hadoop environment, the following components run within the Data Integration Service process:
The following diagram shows how the components interact with the client, the Hadoop cluster, and the Model Repository Service:
  1. 1. A client submits a mapping execution request to the Data Integration Service. The Mapping Service Module receives the request and stores the job in the queue.
  2. 2. The Mapping Service Module connects to the Model Repository Service to fetch mapping metadata from the Model repository.
  3. 3. The Mapping Service Module passes the mapping to the Logical Data Transformation Manager (LDTM).
  4. 4. The LDTM compiles the mapping and generates the Spark execution workflow. It stores the execution workflow in the Model repository.
  5. 5. The LTDM pushes the execution workflow through the Workflow Executor Service to the cluster for processing.
For more information about the architecture of a Data Integration Service, see the "Data Integration Service Architecture" chapter in the Informatica Application Service Guide.