Option | Description |
---|---|
Name | Enter a name for the profile. The profile name must be unique in the folder where you save it. The profile name cannot contain special characters. |
Description | Optionally, enter a description for the profile. |
Location | Choose a location to save the profile. If you do not choose a location, the profile is saved in the Default project. |
Option | Description |
---|---|
Connection | Choose an existing connection. You can create a connection in Administrator. |
Source object | Select a source object to run the profile on. When you click Select to browse for a source object, the Select a Source Object dialog box shows a maximum of 200 source objects. Use the Find field to search for a source object in the list. Optionally, you can use the copy ![]() |
Formatting Options | Optional. Define the file format options. Data Profiling supports CSV and TXT files that have UTF-8 encoding enabled. Appears when you select a file-based connection. |
Advanced Options | Mandatory. Configure the advanced options for the source objects. |
Option | Description |
---|---|
Delimiter | Indicates the boundary between two columns of data. Choose one of the following options:
If you use an escape character or a quote character as the delimiter, or if you use the same character as consecutive delimiter and qualifier, you might receive unexpected results. Default is comma. |
Text Qualifier | Character that defines the boundaries of text strings. If you select a quote character, Data Profiling ignores delimiters within quotes. Default is double quote ("). |
Escape Character | Character that immediately precedes a column delimiter character embedded in an unquoted string, or immediately precedes the quote character in a quoted string. When you specify an escape character, Data Profiling reads the delimiter character as a regular character. |
Field Labels | Choose one of the following options to display the column names in profile results:
|
First Data Row <row_number> | Row number from which Data Profiling starts to read when it imports the file. For example, if you enter 2, Data Profiling skips the first row. Note: Data Profiling sets the First Data Row automatically when you set the Import from Row option. For example, if you set the Import from Row option to 10, Data Profiling sets the First Data Row to 11. |
Option | Description |
---|---|
Schema Source | You must specify the schema of the source file. You can select one of the following options to specify a schema:
Default is Read from data file. |
Delimiter | Character used to separate columns of data. You can configure parameters such as comma, tab, colon, semicolon, or others. To set a tab as a delimiter, you must type the tab character in any text editor. Then, copy and paste the tab character in the Delimiter field.If you specify a multibyte character as a delimiter in the source object, the mapping fails. Default is comma (,). |
Escape Character | Character immediately preceding a column delimiter character embedded in an unquoted string, or immediately preceding the quote character in a quoted string. Default is backslash (\). |
Text Qualifier | Character that defines the boundaries of text strings. If you select a quote character, Data Profiling ignores delimiters within quotes. Default is double quote ("). |
Qualifier Mode | Specify the qualifier behavior for the target object. You can select one of the following options:
Default is Minimal. |
Code Page | UTF-8. Select for Unicode and non-Unicode data. Select the code page that the Secure Agent must use to read data. |
Header Line Number | Specify the line number that you want to use as the header when you read data from Amazon S3. You can also read data from a file that does not have a header. Default is 1. To read data from a file with no header, specify the value of the Header Line Number field as 0.To read data from a file with a header, set the value of the Header Line Number field to a value that is greater or equal to one. This property is applicable during runtime and data preview to read a file. This property is applicable during data preview to write a file. |
First Data Row | Specify the line number from where you want the Secure Agent to read data. You must enter a value that is greater or equal to one. To read data from the header, the value of the Header Line Number and the First Data Row fields should be the same. Default is 2. This property is applicable during runtime and data preview to read a file. This property is applicable during data preview to write a file. |
Row Delimiter | Character used to separate rows of data. You can set values as\r\n,\n, and \r. |
Option | Description |
---|---|
Schema Source | The schema of the source or target file. You can select one of the following options to specify a schema:
|
Schema File | Upload a schema definition file. You cannot upload a schema file when you create a target at runtime. |
Option | Description |
---|---|
Schema Source | The schema of the source or target file. You can select one of the following options to specify a schema:
|
Schema File | Upload a schema definition file. You cannot upload a schema file when you create a target at runtime. |
Sample Size | Specify the number of rows to read to find the best match to populate the metadata. |
Memory Limit | The memory that the parser uses to read the JSON sample schema and process it. The default value is 2 MB.If the file size is more than 2 MB, you might encounter an error. Set the value to the file size that you want to read. |
Option | Description |
---|---|
Schema Source | You must specify the schema of the source file. You can select one of the following options to specify a schema:
Default is Read from data file. |
Delimiter | Character used to separate columns of data. You can configure parameters such as comma, tab, colon, semicolon, or others. Note: You cannot set a tab as a delimiter directly in the Delimiter field. To set a tab as a delimiter, you must type the tab character in any text editor. Then, copy and paste the tab character in the Delimiter field. Default is comma (,). |
Escape Character | Character immediately preceding a column delimiter character embedded in an unquoted string, or immediately preceding the quote character in a quoted string. Default is backslash (\). |
Text Qualifier | Character that defines the boundaries of text strings. If you select a quote character, Data Profiling ignores delimiters within quotes. Default is double quote ("). |
Qualifier Mode | Specify the qualifier behavior for the target object. You can select one of the following options:
Default is Minimal. |
Code Page | Select the code page that the Secure Agent must use to read data. Microsoft Azure Data Lake Storage Gen2 Connector supports only UTF-8. Ignore rest of the code pages. |
Header Line Number | Specify the line number that you want to use as the header when you read data from Microsoft Azure Data Lake Storage Gen2. You can also read a data from a file that does not have a header. To read data from a file with no header, specify the value of the Header Line Number field as 0. Note: This property is applicable when you perform data preview. Default is 1. |
First Data Row | Specify the line number from where you want the Secure Agent to read data. You must enter a value that is greater or equal to one. To read data from the header, the value of the Header Line Number and the First Data Row fields should be the same. Default is 2. Note: This property is applicable when you perform data preview. |
Row Delimiter | Character used to separate rows of data. You can set values as\r\n,\n, and \r. |
Option | Description |
---|---|
Schema Source | The schema of the source or target file. You can select one of the following options to specify a schema:
|
Schema File | Upload a schema definition file. You cannot upload a schema file when you create a target at runtime. |
Option | Description |
---|---|
Schema Source | The schema of the source or target file. You can select one of the following options to specify a schema:
|
Schema File | Upload a schema definition file. You cannot upload a schema file when you create a target at runtime. |
Sample Size | Specify the number of rows to read to find the best match to populate the metadata. |
Memory Limit | The memory that the parser uses to read the JSON sample schema and process it. The default value is 2 MB.If the file size is more than 2 MB, you might encounter an error. Set the value to the file size that you want to read. |
Option | Description |
---|---|
Schema Source | You must specify the schema of the source file. You can select one of the following options to specify a schema:
Default is Read from data file. |
Delimiter | Character used to separate columns of data. You can configure parameters such as comma, tab, colon, semicolon, or others. To set a tab as a delimiter, you must type the tab character in any text editor. Then, copy and paste the tab character in the Delimiter field. If you specify a multibyte character as a delimiter in the source object, the mapping fails. Note: To set a tab as a delimiter, you must type the tab character in any text editor. Then, copy and paste the tab character in the Delimiter field. |
Escape Character | Character immediately preceding a column delimiter character embedded in an unquoted string, or immediately preceding the quote character in a quoted string. |
Text Qualifier | Character that defines the boundaries of text strings. If you select a quote character, Data Profiling ignores delimiters within quotes. Default is double quote ("). |
Qualifier Mode | Specify the qualifier behavior for the target object. You can select one of the following options:
|
Code Page | Select the code page that the Secure Agent must use to read or write data.Google Cloud Storage V2 Connector supports the following code pages:
|
Header Line Number | Specify the line number that you want to use as the header when you read data from Google Cloud Storage. You can also read a file that doesn't have a header. Default is 1. To read data from a file with no header, specify the value of the Header Line Number field as 0. To read data from a file with a header, set the value of the Header Line Number field to a value that is greater than or equal to one. Ensure that the value of the Header Line Number field is lesser than or equal to the value of the First Data Row field. This property is applicable during runtime and data preview to read a file. When you create a mapping in advanced mode, set the value of the header line number to 0, 1, or empty to run the mapping successfully. |
First Data Row | Specify the line number from where you want the Secure Agent to read data. You must enter a value that is greater or equal to one. To read data from the header, the value of the Header Line Number and the First Data Row fields should be the same. Default is 1. This property is applicable during runtime and data preview to read a file. This property is applicable during data preview to write a file. |
Row Delimiter | Not applicable. Character used to separate rows of data. You can set values as\r\n,\n, and \r. |
Option | Description |
---|---|
Schema Source | The schema of the source or target file. You can select one of the following options to specify a schema:
|
Schema File | Upload a schema definition file. You cannot upload a schema file when you create a target at runtime. |
Option | Description |
---|---|
Schema Source | The schema of the source or target file. You can select one of the following options to specify a schema:
|
Schema File | Upload a schema definition file. You cannot upload a schema file when you create a target at runtime. |
Sample Size | Specify the number of rows to read to find the best match to populate the metadata. |
Memory Limit | The memory that the parser uses to read the JSON sample schema and process it. The default value is 2 MB.If the file size is more than 2 MB, you might encounter an error. Set the value to the file size that you want to read. |
Read multiple-line JSON files | Not applicable. |
Option | Description |
---|---|
Schema Source | You must specify the schema of the source file. You can select one of the following options to specify a schema:
Default is Read from data file. |
Delimiter | Character used to separate columns of data. You can configure parameters such as comma, tab, colon, semicolon, or others. Note: You cannot set a tab as a delimiter directly in the Delimiter field. To set a tab as a delimiter, you must type the tab character in any text editor. Then, copy and paste the tab character in the Delimiter field. Default is comma (,). |
Escape Character | Character immediately preceding a column delimiter character embedded in an unquoted string, or immediately preceding the quote character in a quoted string. Default is backslash (\). |
Text Qualifier | Character that defines the boundaries of text strings. If you select a quote character, Data Profiling ignores delimiters within quotes. Default is double quote ("). |
Qualifier Mode | Specify the qualifier behavior for the target object. You can select one of the following options:
Default is Minimal. |
Code Page | Select the code page that the Secure Agent must use to read data. Oracle Cloud Object Storage Connector supports only UTF-8. Ignore rest of the code pages. |
Header Line Number | Specify the line number that you want to use as the header when you read data from Oracle Cloud Object Storage. You can also read a data from a file that does not have a header. To read data from a file with no header, specify the value of the Header Line Number field as 0. Note: This property is applicable when you perform data preview. Default is 1. |
First Data Row | Specify the line number from where you want the Secure Agent to read data. You must enter a value that is greater or equal to one. To read data from the header, the value of the Header Line Number and the First Data Row fields should be the same. Default is 2. Note: This property is applicable when you perform data preview. |
Row Delimiter | Character used to separate rows of data. You can set values as\r\n,\n, and \r. |
Property | Description |
---|---|
Retain Athena Query Result On S3 File | Specifies whether you want to retain the Amazon Athena query result on the Amazon S3 file. Select the check box to retain the Amazon Athena query result on the Amazon S3 file. The Amazon Athena query result in stored in the CSV file format. By default, the Retain Athena Query Result on S3 File check box is not selected. |
S3OutputLocation | Specifies the location of the Amazon S3 file that stores the result of the Amazon Athena query. You can also specify the Amazon S3 file location in the S3OutputLocation parameter in the JDBC URL connection property. If you specify the Amazon S3 output location in both the connection and the advanced source properties, the Secure Agent uses the Amazon S3 output location specified in the advanced source properties. |
Fetch Size | Determines the number of rows to read in one result set from Amazon Athena. Default is 10000. |
Encryption Type | Encrypts the data in the Amazon S3 staging directory. You can select the following encryption types:
Default is None. |
Schema Name | Overrides the schema name of the source object. |
Source Table Name | Overrides the table name used in the metadata import with the table name that you specify. |
SQL Query | Overrides the default SQL query. Enclose column names in double quotes. The SQL query is case sensitive. Specify an SQL statement supported by the Amazon Athena database. When you specify the columns in the SQL query, ensure that the column name in the query matches the source column name in the mapping. |
Option | Description |
---|---|
Source Type | Type of the source from which you want to read data. You can select the following source types:
Default is File. For more information about the source type, see Source types in Amazon S3 V2 sources. |
Folder Path | Optional. Overwrites the bucket name or folder path of the Amazon S3 source file.If applicable, include the folder name that contains the source file in the <bucket_name>/<folder_name> format. If you do not provide the bucket name and specify the folder path starting with a slash (/) in the /<folder_name> format, the folder path appends with the folder path that you specified in the connection properties. For example, if you specify the /<dir2> folder path in this property and <my_bucket1>/<dir1> folder path in the connection property, the folder path appends with the folder path that you specified in the connection properties in <my_bucket1>/<dir1>/<dir2> format. |
File Name | Optional. Overwrites the Amazon S3 source file name. |
Allow Wildcard Characters | Use the ? and * wildcard characters to specify the folder path or file name if you run a mapping in advanced mode to read data from an Avro, flat, JSON, ORC, or Parquet file. |
Enable Recursive Read | Use the recursive read option for flat, Avro, JSON, ORC, and Parquet files. The files that you read using recursive read must have the same metadata. Enable recursive read when you specify wildcard characters in a folder path or file name. To enable recursive read, select the source type as Directory. |
Incremental File Load | Incrementally load source files in a directory to read and process only the files that have changed since the last time the mapping task ran. |
Staging Directory | Optional. Path of the local staging directory. Ensure that the user has write permissions on the directory. In addition, ensure that there is sufficient space to enable staging of the entire file. Default staging directory is the /temp directory on the machine that hosts the Secure Agent. When you specify the directory path, the Secure Agent create folders depending on the number of partitions that you specify in the following format: InfaS3Staging<00/11><timestamp>_<partition number> where, 00 represents read operation and 11 represents write operation. For example, InfaS3Staging000703115851268912800_0 The temporary files are created within the new directory. The staging directory in the source property does not apply to an advanced cluster. However, you must specify a staging directory on Amazon S3 in the advanced configuration. For more information, see Administrator. |
Hadoop Performance Tuning Options | Optional. This property is not applicable for Amazon S3 V2 Connector. |
Compression Format | Decompresses data when you read data from Amazon S3. You can choose to decompress the data in the following formats:
Default is None. Note: Amazon S3 V2 Connector does not support the Lzo and Bzip2 compression format even though the option appears in this property. For more information about the compression format, see Data compression in Amazon S3 V2 sources and targets. |
Download Part Size | Downloads the part size of an Amazon S3 object in bytes. Default is 5 MB. Use this property when you run a mapping to read a file of flat format type. |
Multipart Download Threshold | Minimum threshold size to download an Amazon S3 object in multiple parts. To download the object in multiple parts in parallel, ensure that the file size of an Amazon S3 object is greater than the value you specify in this property. Default is 10 MB. |
Temporary Credential Duration | The time duration during which an IAM user can use the dynamically generated temporarily credentials to access the AWS resource. Enter the time duration in seconds. Default is 900 seconds. If you require more than 900 seconds, you can set the time duration maximum up to 12 hours in the AWS console and then enter the same time duration in this property. |
Option | Description |
---|---|
Concurrent Threads | Optional. Number of concurrent connections to load data from the Microsoft Azure Data Lake Storage Gen2. When writing a large file, you can spawn multiple threads to process data. Configure Block Size to divide a large file into smaller parts. Default is 4. Maximum is 10. |
Filesystem Name Override | Optional. Overrides the default file name. |
Source Type | Type of the source from which you want to read data. You can select the following source types:
Default is File. For more information about the source type, see Directory Source in Microsoft Azure Data Lake Storage Gen2 Sources. |
Allow Wildcard Characters | Use the ? and * wildcard characters to specify the folder path or file name if you run a mapping in advanced mode to read data from an Avro, flat, JSON, ORC, or Parquet file. |
Directory Override | Optional. Microsoft Azure Data Lake Storage Gen2 directory that you use to write data. Default is root directory. The Secure Agent creates the directory if it does not exist. The directory path specified at run time overrides the path specified while creating a connection. |
File Name Override | Optional. Target object. Select the file from which you want to write data. The file specified at run time overrides the file specified in Object. |
Block Size | Optional. Divides a large file or object into smaller parts each of specified block size. When writing a large file, consider dividing the file into smaller parts and configure concurrent connections to spawn required number of threads to process data in parallel. Default is 8 MB. |
Compression Format | Optional. Compresses and writes data to the target. Select Gzip to write flat files. |
Timeout Interval | Optional. The number of seconds to wait when attempting to connect to the server. A timeout will occur if the connection cannot be established in the specified amount of time. |
Interim Directory | Optional. Path to the staging directory in the Secure Agent machine. Specify the staging directory where you want to stage the files when you read data from Microsoft Azure Data Lake Store. Ensure that the directory has sufficient space and you have write permissions to the directory. Default staging directory is /tmp. You cannot specify an interim directory for an advanced cluster. |
Incremental File Load | Incrementally load source files in a directory to read and process only the files that have changed since the last time the mapping task ran. |
Enable Recursive Read | Use the recursive read option for flat, Avro, JSON, ORC, and Parquet files. The files that you read using recursive read must have the same metadata. Enable recursive read when you specify wildcard characters in a folder path or file name. To enable recursive read, select the source type as Directory. |
Option | Description |
---|---|
Google Cloud Storage Path | Overrides the Google Cloud Storage path that you specified in the connection. This property is required when the source is not a flat file. Use the following format: gs://<bucket name> or gs://<bucket name>/<folder name> |
Source File Name | Optional. Overrides the Google Cloud Storage source file name that you specified in the Source transformation. Note: Does not apply when you configure Is Directory option to read multiple files from a directory. |
Is Directory | Select this property to read all the files available in the folder specified in the Google Cloud Storage Path property. |
Encryption Type | Method to decrypt data. You can select one of the following encryption types:
Default is None . |
Option | Description |
---|---|
Database | Overrides the database specified in the connection. |
Schema | Overrides the schema specified in the connection. |
Warehouse | Overrides the Snowflake warehouse name specified in the connection. |
Role | Overrides the Snowflake role assigned to user specified in the connection. |
Table Name | Overrides the table name of the imported Snowflake Data Cloud source table. |
Option | Description |
---|---|
S3 Bucket Name | Amazon S3 bucket name for staging the data. You can also specify the bucket name with the folder path. |
Enable Compression | Compresses the staging files into the Amazon S3 staging directory. The task performance improves when the Secure Agent compresses the staging files. Default is selected. |
Staging Directory Location | Location of the local staging directory. When you run a task in Secure Agent runtime environment, specify a directory path that is available on the corresponding Secure Agent machine in the runtime environment.Specify the directory path in the following manner: <staging directory> For example, C:\Temp. Ensure that you have the write permissions on the directory. Does not apply to an advanced cluster. |
Temporary Credential Duration | The time duration during which an IAM user can use the dynamically generated temporarily credentials to access the AWS resource. Enter the time duration in seconds. Default is 900 seconds. If you require more than 900 seconds, you can set the time duration up to a maximum of 12 hours in the AWS console and then enter the same time duration in this property. |
Encryption Type | Encrypts the data in the Amazon S3 staging directory. You can select the following encryption types:
You can only use SSE-S3 encryption in a mapping that runs on an advanced cluster. Default is None. |
Download S3 Files in Multiple Parts | Downloads large Amazon S3 objects in multiple parts. When the file size of an Amazon S3 object is greater than 8 MB, you can choose to download the object in multiple parts in parallel. Default is 5 MB. Does not apply to an advanced cluster. |
Multipart Download Threshold Size | The maximum threshold size to download an Amazon S3 object in multiple parts. Default is 5 MB. Does not apply to an advanced cluster. |
Schema Name | Overrides the default schema name. Note: You cannot configure a custom query when you use the schema name. |
Source Table Name | Overrides the default source table name. Note: When you select the source type as Multiple Objects or Query , you cannot use the Source Table Name option. |
Option | Description |
---|---|
Database Name | Overrides the database name provided in connection and the database name provided during metadata import. Note: To read from multiple objects. ensure that you have specified the database name in the connection properties. |
Table Name | Overrides the table name used in the metadata import with the table name that you specify. |
SQL Override | Overrides the default SQL query used to read data from a Databricks Delta custom query source. The column names in the SQL override query should match with the column names in the custom query in a SQL transformation. Note: The metadata of the source should be the same as SQL override to override the query. You can use the option when you run the profiling task on a Data Integration Server. |
Staging Location | Relative directory path to the staging file storage.
Note: When you use the unity catalog, a pre-existing location on the user's cloud storage must be provided in the Staging Location. The Staging Location is not required for the Unity Catalog when you run the profiling task on a Data Integration Server. |
Job Timeout | Maximum time in seconds that is taken by the Spark job to complete processing. If the job is not completed within the time specified, the Databricks cluster terminates the job and the mapping fails. If the job timeout is not specified, the mapping shows success or failure based on the job completion. |
Job Status Poll Interval | Poll interval in seconds at which the Secure Agent checks the status of the job completion. Default is 30 seconds. |
DB REST API Timeout | The Maximum time in seconds for which the Secure Agent retries the REST API calls to Databricks when there is an error due to network connection or if the REST endpoint returns 5xx HTTP error code. Default is 10 minutes. |
DB REST API Retry Interval | The time Interval in seconds at which the Secure Agent must retry the REST API call, when there is an error due to network connection or when the REST endpoint returns 5xx HTTP error code. This value does not apply to the Job status REST API. Use job status poll interval value for the Job status REST API. Default is 30 seconds. |
Option | Description |
---|---|
Azure Blob Container Name | Microsoft Azure Blob Storage container name. Required if you select Azure Blob storage in the connection properties. |
ADLS FileSystem Name | The name of the file system in Microsoft Azure Data Lake Storage Gen2. Required if you select ADLS Gen2 storage in the connection properties. You can also provide the path of the directory under given file system. |
Schema Name Override | Overrides the schema specified in the connection. |
Table Name Override | Overrides the table name of the imported Microsoft Azure Synapse SQL source table. |
Field Delimiter | Character used to separate fields in the file. Default is 0x1e. You can specify 'TAB' or 0-256 single-char printable and non-printable ASCII characters. Non-printable characters must be specified in hexadecimal. |
Number of concurrent connections to Blob Store | Number of concurrent connections to extract data from the Microsoft Azure Blob Storage. When reading a large-size blob, you can spawn multiple threads to process data. Configure Blob Part Size to partition a large-size blob into smaller parts. Default is 4. Maximum is 10. |
Blob Part Size | Partitions a blob into smaller parts each of specified part size. When reading a large-size blob, consider partitioning the blob into smaller parts and configure concurrent connections to spawn required number of threads to process data in parallel. Default is 8 MB. |
Quote Character | The Secure Agent skips the specified character when you read data from Microsoft Azure Synapse SQL. Default is 0x1f . |
Interim Directory | Optional. Path to the staging directory in the Secure Agent machine.Specify the staging directory where you want to stage the files when you read data from Microsoft Azure Synapse SQL. Ensure that the directory has sufficient space and you have write permissions to the directory. Default staging directory is /tmp. You cannot specify an interim directory for an advanced cluster. |
Option | Description |
---|---|
Schema Name Override | Overrides the schema specified in the connection. |
Table Name Override | Overrides the table name of the Microsoft Fabric Data Warehouse source table. |
SQL Override | The SQL statement to override the default query and the object name that is used to read data from the Microsoft Fabric Data Warehouse source. |
pre SQL | The SQL query that the Secure Agent runs before reading data from the source. |
post SQL | The SQL query that the Secure Agent runs after reading data from the source. |
On Pre-Post SQL Error | Required if the you use pre SQL or post SQL options. If you select Stop, the SQL query does not run if an error occurs. If you select Continue, the SQL query continues to run regardless of errors. |
Option | Description |
---|---|
Pre SQL | The SQL query that the Secure Agent runs before reading data from the source. |
Post SQL1 | The SQL query that the Secure Agent runs after reading data from the source. |
Fetch Size | The number of rows that the Secure Agent fetches from the database in a single call. |
Table Name | Overrides the table name used in the metadata import with the table name that you specify. |
Schema Name | Overrides the schema name of the source object. If you specify the schema name both in the connection and the source properties, the Secure Agent uses the schema name specified in the source properties. |
SQL Override | The SQL statement to override the default query and the object name that is used to read data from the JDBC V2 source. |
1Doesn't apply to mappings in advanced mode. |
Option | Description |
---|---|
Folder Path | Overrides the folder path value in the Oracle Cloud Object Storage connection. |
File Name | Overrides the Oracle Cloud Object Storage source file name. |
Staging Directory | Path of the local staging directory. Ensure that the user has write permissions on the directory. In addition, ensure that there is sufficient space to enable staging of the entire file. Default staging directory is the /temp directory on the machine that hosts the Secure Agent. The temporary files are created within the new directory. |
Multipart Download Threshold | Minimum threshold size to download an Oracle Cloud Object Storage object in multiple parts. To download the object in multiple parts in parallel, ensure that the file size of an Oracle Cloud Object Storage object is greater than the value you specify in this property. Range :
Default is 64 MB. |
Download Part Size | Downloads the part size of an Oracle Cloud Object Storage object in bytes. Range:
Default is 32 MB. |
Option | Description |
---|---|
Pre-SQL | The SQL query that the Secure Agent runs before reading data from the source. |
Post-SQL | The SQL query that the Secure Agent runs after reading data from the source. |
Fetch Size | Determines the number of rows to read in one result set from the PostgreSQL database. Default is 100000. |
Schema Name | Overrides the schema name of the source object. |
Source Table Name | Overrides the default source table name. |
SQL Override | The SQL statement to override the default query and the object name that is used to read data from the PostgreSQL source. |
Option | Description |
---|---|
Packet size in MB | Size of the HTTP packet that SAP sends to the Secure Agent. The unit is MB. Default is 10 MB. |
Package size in ABAP in rows | Number of rows that are read and buffered in SAP at a time. Default is 1000 rows. |
Enable Compression | When selected, the ABAP program compresses the data in the gzip format before it sends the data to the Secure Agent. If the Secure Agent and the SAP system are not on the same network, you might want to enable the compression option to optimize performance. Default is not selected. |
Option | Description |
---|---|
Number of rows to be fetched | The number of rows that are randomly retrieved from the SAP Table. Default value of zero retrieves all the rows in the table. |
Number of rows to be skipped | The number of rows to be skipped. |
Packet size in MB | Packet size. Default is 10 MB. |
Data extraction mode | You can use one of the following modes to read data from an SAP Table:
For more information about the data extraction mode, see the Data Extraction mode section in the Performance Tuning Guidelines for SAP Table Reader Connector How-To Library article. |
Enable Compression | Enables compression. If the Secure Agent and the SAP System are not located in the same network, you may want to enable the compression option to optimize performance. |
Update Mode | When you read data from SAP tables, you can configure a mapping to perform delta extraction. You can use one of the following options based on the update mode that you want to use:
Default is 0- Full. For more information about the update mode, see the Update modes for delta extraction section in the SAP connector help. |
Parameter Name for Update Mode | The parameter name that you defined for update mode in the parameter file. |
Override Table Name for Delta Extraction | Overrides the SAP table name with the SAP structure name from which you want to extract delta records that are captured with the structure name in the CDPOS table. |
Property | Description |
---|---|
Run profile on | Choose one of the following sampling options to run the profile:
|
Drilldown | Choose one of the following drill-down options:
To drill down and to query the source object, you need Data Preview privileges in Data Profiling. Note: You cannot perform drill down on the profile results or queries if you select the Avro or Parquet source object for Amazon S3 and Azure Data Lake Store connections. |
Connection | Sampling Option |
---|---|
Amazon Athena | All Rows First N Rows |
Amazon Redshift V2 | All Rows Random N Rows |
Amazon S3 v2 | All Rows |
Azure Data Lake Store Gen2 | All Rows |
Databricks Delta | All Rows (Data Integration Server and advanced mode execution) Sample N Rows (Data Integration Server execution) |
Flat File | All Rows |
Google Big Query v2 | All Rows |
Google Cloud Storage V2 | All Rows |
JDBC V2 | All Rows First N Rows |
Mapplets | All Rows |
Microsoft Azure Synapse SQL | All Rows First N Rows Random N Rows |
Microsoft Fabric Data Warehouse | All Rows First N Rows |
ODBC | All Rows First N Rows. For Postgres and IBM DB2 data sources over an ODBC connection. |
Oracle | All Rows First N Rows |
PostgreSQL | All Rows First N Rows |
SAP BW Reader | All Rows |
SAP Table | All Rows To retrieve random number of rows from the data source, you can configure the Number of rows to be fetched option in the advanced options for the source connection. |
SQL Server | All Rows First N Rows |
Salesforce | All Rows First N Rows |
Snowflake Data Cloud | All Rows First N Rows Random N Rows |
Property | Description |
---|---|
Columns | The column names in the selected source object. |
Type | The data type that appears in the transformations. They are internal data types based on ANSI SQL-92 generic data types, which the Secure Agent uses to move data across platforms. Transformation data types appear in all transformations in a mapping |
Precision | The number of characters in a column. |
Native Data Type | The data type specific to the source database or flat files. |
Scale | The number of numeric characters after the decimal point. |
Nullable | Indicates whether the column can accommodate a NULL value. |
Key | Indicates whether the column has been designated as a primary key in the data source. |