IBM Watson Data Platform Data Flows Service - Data Asset and Connection Properties

The following information describes the possible values of the 'properties' section of a connection within a data flow. The properties that follow can also be used in the 'properties' section of a data asset definition, although care has to be taken to confirm the type of connection referenced by the data asset. Different types of connections will support different properties (termed 'interaction properties' below). Some of these properties are only applicable when the connection is used as a source, some when the connection is used as a target, and some are applicable in either case.

For example, if an 'IBM Db2 Warehouse on Cloud' connection is used as a source both the table_name and schema_name properties can be set as follows:
	{  
	   "id":"source1",
	   "type":"binding",
	   "output":{  
	      "id":"source1Output"
	   },
	   "connection":{  
	      "properties":{  
	         "schema_name":"GOSALESHR",
	         "table_name":"EMPLOYEE"
	      },
	      "ref":"{connection_id}"
	   }
	}
alternatively the 'IBM Db2 Warehouse on Cloud' connection used as a source also allows just a SQL select statement to be provided:
	{  
	   "id":"source1",
	   "type":"binding",
	   "output":{  
	      "id":"source1Output"
	   },
	   "connection":{  
	      "properties":{  
	         "select_statement":"select * from GOSALES.PRODUCT_NAME_LOOKUP"
	      },
	      "ref":"{connection_id}"
	   }
	}

Note that in the tables below a * character next to a property name denotes the property is required to be assigned a value.


Table of contents

Amazon RDS for MySQL
Amazon RDS for Oracle
Amazon RDS for PostgreSQL
Amazon Redshift
Amazon S3
Apache Cassandra
Apache Cassandra (optimized)
Apache Derby
Apache HDFS
Apache Hive
Apache Kafka
Box
Cloudera Impala
Custom No-op
Dremio
Dropbox
Elasticsearch
FTP
Generic S3
Google BigQuery
Google Cloud Pub/Sub
Google Cloud Storage
Greenplum
HTTP
IBM Cloud Data Engine
IBM Cloud Databases for DataStax
IBM Cloud Databases for MongoDB
IBM Cloud Databases for MySQL
IBM Cloud Databases for PostgreSQL
IBM Cloud Object Storage
IBM Cloud Object Storage (infrastructure)
IBM Cloudant
IBM Cognos Analytics
IBM Data Virtualization Manager for z/OS
IBM Db2
IBM Db2 (optimized)
IBM Db2 Big SQL
IBM Db2 Warehouse
IBM Db2 for i
IBM Db2 for z/OS
IBM Db2 on Cloud
IBM Informix
IBM MQ
IBM Match 360
IBM Netezza Performance Server
IBM Netezza Performance Server (optimized)
IBM Planning Analytics
IBM Product Master
IBM Watson Query
Looker
MariaDB
Microsoft Azure Blob Storage
Microsoft Azure Cosmos DB
Microsoft Azure Data Lake Store
Microsoft Azure File Storage
Microsoft Azure SQL Database
Microsoft SQL Server
MongoDB
MySQL
ODBC
OData
Oracle
Oracle (optimized)
PostgreSQL
Presto
SAP ASE
SAP IQ
SAP OData
Salesforce.com
Salesforce.com (optimized)
SingleStoreDB
Snowflake
Tableau
Teradata
Teradata (optimized)




Amazon RDS for MySQL


Description: Amazon RDS for MySQL is a MySQL relational database that runs on the Amazon Relational Database Service (RDS).
Data source type ID: 9aa630f2-efc4-4d54-b8cb-254f31405b78
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: true
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [none, random]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Amazon RDS for Oracle


Description: Amazon RDS for Oracle is an Oracle relational database that runs on the Amazon Relational Database Service (RDS).
Data source type ID: 8b8fcd6d-8f95-49c7-8195-c72c95c9a84b
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
connection_mode. Values: [sid, service_name]
sid *The unique name of the database instance. If you provide a SID, do not provide a service name
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
metadata_discoveryDetermines what types of metadata can be discovered, 'No Remarks' option will be set as default. Values: [no_remarks, no_remarks_or_synonyms, no_synonyms, remarks_and_synonyms]
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
service_name *The name of the service. If you provide a service name, do not provide a SID
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Amazon RDS for PostgreSQL


Description: Amazon RDS for PostgreSQL is a PostgreSQL relational database that runs on the Amazon Relational Database Service (RDS).
Data source type ID: 9493d830-882b-445e-96c7-8e4c635a1a5b
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
database *The name of the database
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: true
query_timeoutSets the default query timeout in seconds for all statements created by a connection. If not specified the default value of 300 seconds will be used.. Default: 300
retry_limitSpecify the maximum number of retry connection attempts to be made by the connector with an increasing delay between each retry. If no value is provided, two attempts will be made by default if necessary.
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Amazon Redshift


Description: Amazon Redshift is a fully managed, scalable cloud data warehouse product that is part of Amazon Web Services (AWS).
Data source type ID: 31170994-f54c-4148-9c5a-807832fa1d07
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
database *The name of the database
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: true
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
time_typeChoose the required Time Type for time values in the data source . Values: [time, timestamp, varchar]. Default: timestamp
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [none, random]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
access_key *The access key ID (username) for authorizing access to AWS
bucket *The name of the bucket that contains the files to access
create_statementThe Create DDL statement for recreating the target table
file_name *Name of a temporary file to be stored in the S3 bucket.
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
region *Amazon Web Services (AWS) region
schema_nameThe name of the schema that contains the table to write to
secret_key *The password associated with the access key ID for authorizing access to AWS
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, load, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Amazon S3


Description: Amazon S3 (Amazon Simple Storage Service) is an object storage service that is offered by Amazon Web Services (AWS). For other types of S3-compliant connections, you can use the Generic S3 connection.
Data source type ID: a0b1d14a-4767-404c-aac1-4ce0e62818c3
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
access_keyThe access key ID (username) for authorizing access to AWS
auth_method. Values: [basic_credentials, temporary_credentials, trusted_role_credentials]
bucketThe name of the bucket that contains the files to access
duration_secondsThe duration in seconds of the temporary security credentials
urlThe endpoint URL to use for access to AWS S3
external_idThe external ID of the organization that is attempting to assume a role
proxy_host *The server proxy host
proxy_passwordThe password used to authenticate with the server proxy
proxy_port *The server proxy port
proxy_userThe name of the user used to connect to the server proxy
regionAmazon Web Services (AWS) region
role_arnThe Amazon Resource Name (ARN) of the role that the connection should assume
role_session_name *A name such as your IAM user name to identify the session to S3 administrators
secret_key *The password associated with the access key ID for authorizing access to AWS
proxyUse server proxy. Default: false
session_tokenThe session token (only needed with temporary credentials)

Interaction properties (when used as a source)

Name Type Description
bucketThe name of the bucket that contains the files to read
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
_file_format._delimited_syntax._data_formatSelect how binary data is represented. Binary data includes data that is of integer, float, double, or binary data types. If variable length binary fields are written as binary, they are prefixed with a 4-byte integer that represents the size of the field.. Values: [1, 0]. Default: 0
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
_file_format._delimited_syntax._record_def._record_def_sourceIf the record definition is a delimited string, enter a delimited string that specifies the names and data types of the files. Use the format name:data_type, and separate each field with the delimiter specified as the >B<Field delimiter>/B< property. If the record definition is in a delimited string file or Osh schema file, specify the full path of the file.
display_value_labelsDisplay the value labels
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
_file_format._delimited_syntax._escapeSpecify the character to use to escape field and row delimiters. If an escape character exists in the data, the escape character is also escaped. Because escape characters require additional processing, do not specify a value for this property if you do not need to include escape characters in the data.
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
_exclude_filesSpecify a comma-separated list of file prefixes to exclude from the files that are read. If a prefix includes a comma, escape the comma by using a backslash (\).
exclude_missing_valuesSet values that have been defined as missing values to null
_file_format._delimited_syntax._field_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default: ,
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
_file_format. Values: [2, 4, 1, 0, 6, 7]. Default: 0
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_nameThe name of the file to read
_filename_columnSpecify the name of the column to write the source file name to.
first_lineIndicates at which row start reading. Default: 0
_first_line_headerIndicates whether the row where reading starts is the header. Default: false
first_line_headerIndicates whether the row where reading starts is the header. Default: false
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
_recurseSpecify whether to read files that are in child folders of the prefix that is specified for the File name property. If you exclude child folders, the prefix that is specified must include a trailing forward slash (/).. Default: true
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
null_valueThe value that represents null (a missing value) in the file, for example, NULL
_file_format._avro_source._output_j_s_o_n#PROP_DESC_AVROOUTPUTJSON#. Default: false
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
_read_modeSelect Read single file to read from a single file or Read multiple files to read from the files that match a specified file prefix. Select List buckets to list the buckets for your account in the specified region. Select List files to list files files for your account in the specified bucket.. Values: [2, 3, 1, 0]. Default: 0
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
_file_format._delimited_syntax._record_defSelect whether the record definition is provided to the Amazon S3 connector from the source file, a delimited string, a file that contains a delimited string, or a schema file. When runtime column propagation is enabled, this metadata provides the column definitions. If a schema file is provided, the schema file overrides the values of formatting properties in the stage and the column definitions that are specified on the Columns page of the output link.. Values: [1, 2, 3, 0, 4]. Default: 0
_reject_modeSpecify what the connector does when a record that contains invalid data is found in the source file. Select Continue to read the rest of the file, Fail to stop the job with an error message, or Reject to send the rejected data to a reject link.. Values: [0, 1, 2]. Default: 0
_file_format._delimited_syntax._row_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default:
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
_file_format._o_r_c_source._temp_staging_areaSpecify a directory on the engine tier with write permission for the user running the job. This directory will be used to create the temporary files during the job run.
_file_format._parquet_source._temp_staging_areaSpecify a directory on the engine tier with write permission for the user running the job. This directory will be used to create the temporary files during the job run.
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
_file_format._delimited_syntax._trace_fileSpecify the full path to a file to contain trace information from the parser for delimited files. Because writing to a trace file requires additional processing, specify a value for this property only during job development.
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
_create_bucket._append_u_u_i_dSelect whether to append a unique set of characters to identify the bucket to the bucket name that is created.. Default: false
append_uidUse this property to choose if a unique identifier is to be appended to the file name. When the value of this property is set to yes, then the file name gets appended with the unique identifier, and a new file would be written for every wave of data that is streamed into the stage. When the value of this property is set to No, then the file would be overwritten on every wave.. Default: false
_file_attributes._life_cycle_rule._transitionSpecify whether to archive the file in Amazon Glacier. You can specify the date when the file is set to be archived or the number of days before the file is set to be archived.. Default: false
_file_format._avrotarget._avro_array_keysIf the file format is Avro in a target stage, normalization is controlled via array keys.
_file_format._avrotarget._avro_schemaSpecify the fully qualified path for a JSON file that defines the schema for the Avro file.
_file_format._parquet_target._parquet_block_sizeSpecify the blocksize, default is 100000. Default: 10000000
bucketThe name of the bucket that contains the files to write
_file_format._o_r_c_target._orc_buffer_sizeBuffer size. Default: 10000
_codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
_file_format._o_r_c_target._orc_compressSpecify compression mechanism. Values: [0, 2, 1]. Default: 2
_file_format._parquet_target._parquet_compressSpecify compression mechanism. Values: [2, 3, 0, 1]. Default: 1
_file_attributes._content_typeSpecify the content type of the file to write, for example, text/xml or application/x-www-form-urlencoded; charset=utf-8.
create_bucketCreate the bucket that contains the files to write to. Default: false
_file_format._delimited_syntax._data_formatSelect how binary data is represented. Binary data includes data that is of integer, float, double, or binary data types. If variable length binary fields are written as binary, they are prefixed with a 4-byte integer that represents the size of the field.. Values: [1, 0]. Default: 0
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
_file_attributes._life_cycle_rule._transition._transition_date *Specify the date when the file is set to be archived in Amazon Glacier in the format "YYYY-MM-DD".
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
_file_attributes._life_cycle_ruleSpecify whether you want to define one or more rules for when a file is set to expire or be archived.. Default: false
_file_attributes._life_cycle_rule._expiration._expiration_duration *Specify the number of days that the file will exist in Amazon Web Services before it expires.
_file_attributes._life_cycle_rule._transition._transition_duration *Specify the number of days that the file will exist in Amazon S3 before it is set to be archived in Amazon Glacier.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
_file_format._delimited_syntax._escapeSpecify the character to use to escape field and row delimiters. If an escape character exists in the data, the escape character is also escaped. Because escape characters require additional processing, do not specify a value for this property if you do not need to include escape characters in the data.
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
_file_attributes._life_cycle_rule._expirationSpecify whether you want the file to expire. When a file expires, it is deleted from Amazon Web Services. You can specify the date when the file is set to expire or the number of days that the file will exist in Amazon Web Services before it is set to expire.. Default: false
_file_attributes._life_cycle_rule._expiration._expiration_date *Specify the date when the file is set to expire in the format "YYYY-MM-DD".
_file_format._delimited_syntax._field_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default: ,
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
_file_format. Values: [2, 4, 1, 0, 6, 7]. Default: 0
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_nameThe name of the file to write to or delete
file_size_thresholdSpecify the threshold for the file size in megabytes. Processing nodes will start a new file each time the size exceeds the value specified in the threshold.. Default: 1
_first_line_headerIndicates whether the row where reading starts is the header. Default: false
first_line_headerIndicates whether the row where reading starts is the header. Default: false
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
_file_existsSpecify what the connector does when it tries to write a file that already exists. Select Overwrite file to overwrite a file if it already exists, Do not overwrite file to not overwrite the file and stop the job, or Fail to stop the job with an error message.. Values: [1, 2, 0]. Default: 0
_file_format._delimited_syntax._encoding._output_b_o_mSpecify whether to include a byte order mark in the file when the file encoding is a Unicode encoding such as UTF-8, UTF-16, or UTF-32.. Default: false
include_typesInclude data types in the first line of the file. Default: false
_file_format._avrotarget._input_j_s_o_n#PROP_DESC_AVROINPUTJSON#. Default: false
_log_intervalSpecify the amount of data in MB that the connector writes to Amazon S3 before the connector writes a progress message to the job log. For example, if the interval is 20 MB, the connector writes a progress message to the log after the connector writes 20 MB of data, 40 MB of data, and so on. If you do not specify an interval, progress messages are not written.
names_as_labelsSet column labels to the value of the column name
null_valueThe value that represents null (a missing value) in the file, for example, NULL
_thread_countSpecify the number of writers that will write parts of the file at the same time.. Default: 5
_file_format._parquet_target._parquet_page_sizeSpecify the pagesize, default is 10000. Default: 10000
partitionedWrite the file as multiple partitions. Default: false
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
_file_format._delimited_syntax._row_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default:
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
_file_attributes._life_cycle_rule._l_c_rule_scopeSpecify whether to apply the rule to the file only or to all of the files in the folder that contains the file. If the connector runs in parallel, this property is ignored, and the rule is applied to all of the files in the folder.. Values: [0, 1]. Default: 0
_file_attributes._encryption. Values: [1, 2, 0]. Default: 0
_file_attributes._storage_classSpecify the storage class for the file. The reduced redundancy storage class provides less redundancy for files than the standard class. For more information, see the Amazon S3 documentation.. Values: [1, 0]. Default: 0
_file_format._o_r_c_target._orc_stripe_sizeStripe size. Default: 100000
_file_format._o_r_c_target._temp_staging_areaSpecify a directory on the engine tier with write permission for the user running the job. This directory will be used to create the temporary files during the job run.
_file_format._parquet_target._temp_staging_areaSpecify a directory on the engine tier with write permission for the user running the job. This directory will be used to create the temporary files during the job run.
time_formatThe format of time values, for example, HH:mm:ss[.f]
_file_attributes._life_cycle_rule._l_c_rule_formatSpecify whether the lifecycle rule is based on the number of days from the date that the file is created or based on a specific date.. Values: [0, 1]. Default: 0
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
_file_format._avrotarget._use_schemaSpecify if you would like to provide the Avro schema using a schema file. It is recommended to use No for primitive datatypes and Yes for complex datatypes.. Default: false
_file_attributes._user_metadataSpecify metadata in a list of name-value pairs. Separate each name-value pair with a semicolon, for example, Topic=News;SubTopic=Sports. All characters that you specify must be in the US-ASCII character set.
sheet_nameThe name of the Excel worksheet to write to
_write_modeSelect Write to write a file per node, or select Delete to delete files.. Values: [1, 0]. Default: 0
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




Apache Cassandra


Description: Apache Cassandra is an open source, distributed, NoSQL database.
Data source type ID: e6ff8c10-4199-4b58-9a93-749411eafacd
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
keyspace *The name of the keyspace
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Apache Cassandra (optimized)


Description: Connect to an Apache Cassandra database for the DataStage Apache Cassandra (optimized) connector. The Apache Cassandra (optimized) connector has more features for DataStage.
Data source type ID: 123e4263-dd25-44e5-8282-cf1b2eeea9bd
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
authenticator_typeChose authentication method used in connection to Cassandra cluster. Values: [allow_all_authenticator, password_authentication]. Default: allow_all_authenticator
cluster_contact_points *Multiple contact points (host names or IPs) of the target Cassandra cluster, separated by semicolon
compressionThe type of compression of transport-level requests and responses - you need to provide third-party libraries and make them available on the connector classpath. Values: [lz4, no_compression, snappy]. Default: no_compression
ssl_keystore_passwordProvide the password that was used when generating the keystore
ssl_keystore_pathThe path to your keystore file
local_datacenter *The name of the datacenter local to defined contact points.. Default: datacenter1
password *The user's password used to connect to Cassandra cluster
protocol_versionChose CQL native protocol version that should be used to connect to the target Cassandra cluster. Values: [dse_v1, dse_v2, newest_supported, newest_beta, v1, v2, v3, v4, v5, v9]. Default: newest_supported
ssl_truststore_passwordProvide the password that was used when generating the truststore
ssl_truststore_pathThe path to your truststore file
use_sslUse SSL/TLS to secure connection between client and Cassandra cluster. Default: false
use_ssl_client_cert_authWith this option Cassandra nodes verify the identity of the client
use_ssl_client_encryptionThe traffic between client and cluster nodes is encrypted and the client verifies the identity of the Cassandra nodes it connects to
username *The name of the user used to connect to Cassandra cluster

Interaction properties (when used as a source)

Name Type Description
check_schema_agreementCheck if schema is exactly the same on all cluster nodes. Default: true
read_consistency_levelThe level of consistency used in the read or write operation. Values: [all_nodes, each_data_center_quorum, local_one, local_quorum, one_node, quorum, three_nodes, two_nodes]. Default: quorum
custom_typecodecsYou can provide list of type codec classes that can be used to support your custom mappings between Cassandra and DataStage (semicolon separated list of classes)
tracing_statementsWith tracing enabled the connector provides execution plan for each CQL statement (SELECT, INSERT, UPDATE, DELETE). Default: false
enable_quoted_identifiersSpecifies whether or not to enclose database object names in quotes when generating CQL statements. Default: false
ignore_blob_truncation_errorsShould we ignore BLOB truncation errors when value's length is bigger than column field length provided in link column definition. Default: false
ignore_string_truncation_errorsShould we ignore string truncation errors when value's length is bigger than column field length provided in link column definition. Default: false
cassandra_keyspace *The name of the keyspace in the target Cassandra database
lookup_typeLookup Type. Values: [empty]. Default: empty
page_sizeThe size of page that is used to retrive a subset of data. Default: 10
parallel_read_strategyParallel read strategy determines how workload is distributed among players. Values: [equal_splitter, host_aware]. Default: equal_splitter
prefetching_thresholdStart pre-fetching when current page contains less rows than the value set in this property. Default: 2
cassandra_table *The name of the table in the target Cassandra database
use_json_mapped_rowsEnables selecting and inserting a single row as a JSON encoded map. Default: false
use_parallel_readSplit reading data to all available nodes to speed up the process. Default: false

Interaction properties (when used as a target)

Name Type Description
check_schema_agreementCheck if schema is exactly the same on all cluster nodes. Default: true
write_consistency_levelThe level of consistency used in the read or write operation. Values: [all_nodes, any_node, each_data_center_quorum, local_one, local_quorum, one_node, quorum, three_nodes, two_nodes]. Default: quorum
custom_typecodecsYou can provide list of type codec classes that can be used to support your custom mappings between Cassandra and DataStage (semicolon separated list of classes). Default: com.ibm.is.cc.cassandra.codec.UuidToStringCodec;com.ibm.is.cc.cassandra.codec.TimeUuidToStringCodec;com.ibm.is.cc.cassandra.codec.VarIntToStringCodec;com.ibm.is.cc.cassandra.codec.InetToStringCodec
tracing_statementsWith tracing enabled the connector provides execution plan for each CQL statement (SELECT, INSERT, UPDATE, DELETE). Default: false
enable_quoted_identifiersSpecifies whether or not to enclose database object names in quotes when generating CQL statements. Default: false
cassandra_keyspace *The name of the keyspace in the target Cassandra database
mutation_typeChose the type of modification that you would like perform. Values: [delete_columns, delete_entire_rows, insert, update]. Default: insert
save_null_valuesShould we save null values in the target table (this will create tombstones for each null value). Default: true
cassandra_table *The name of the table in the target Cassandra database
use_json_mapped_rowsEnables selecting and inserting a single row as a JSON encoded map. Default: false




Apache Derby


Description: Apache Derby is an open-source relational database.
Data source type ID: 82696f1d-600c-4f78-a03c-d8349ea1976f
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Apache HDFS


Description: Connect to Apache HDFS with the WebHDFS API. Apache Distributed File System (HDFS) is a distributed file system designed to run on commodity hardware. Apache HDFS was formerly named Hortonworks HDFS.
Data source type ID: c10e5224-f17d-4524-844f-e97b1305e489
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
connect_to_apache_hiveEnter the Hive properties if you plan to write tables to the Hive data source using this connection.. Default: false
hive_sslDetermines whether use SSL protocol for Hive connection. Default: true
hive_dbThe database in Apache Hive
hive_http_pathThe path of the endpoint such as gateway/default/hive when the Apache Hive server is configured for HTTP transport mode
hive_hostThe hostname or IP address of the Apache Hive server
hive_passwordThe password associated with the username for connecting to Apache Hive
hive_portThe port of the Apache Hive server
hive_userThe username for connecting to Apache Hive
passwordThe password associated with the username for accessing the data source
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
use_home_as_rootUser home directory is used as the root of browsing. Default: true
username *The username for accessing the data source
url *The WebHDFS URL for accessing HDFS

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
_file_format.impl_syntax.binarySpecify the type of implicit file. Values: [binary]. Default: binary
_file_format.delimited_syntax.field_formats.date_formatSpecify a string that defines the format for fields that have the Date data type.
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
_file_format.delimited_syntax.field_formats.decimal_formatSpecify a string that defines the format for fields that have the Decimal or Numeric data type.
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
_file_format.delimited_syntax.record_def.record_def_sourceIf the record definition is a delimited string, enter a delimited string that specifies the names and data types of the files. Use the format name:data_type, and separate each field with the delimiter specified as the >B/B< property. If the record definition is in a delimited string file or Osh schema file, specify the full path of the file.
_file_format.impl_syntax.record_def.record_def_sourceEnter a delimited string that specifies the names and data types and length of each fields. Use the format name:data_type[length], and separate each field with the delimiter specified as the >B/B< property. If the record definition is in a delimited string file or Osh schema file, specify the full path of the file.
display_value_labelsDisplay the value labels
_file_format.delimited_syntax.encodingSpecify the encoding of the files to read or write, for example, UTF-8.
_file_format.impl_syntax.encodingSpecify the encoding of the files to read or write, for example, UTF-8.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
_file_format.delimited_syntax.escapeSpecify the character to use to escape field and row delimiters. If an escape character exists in the data, the escape character is also escaped. Because escape characters require additional processing, do not specify a value for this property if you do not need to include escape characters in the data.
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
_exclude_filesSpecify a comma-separated list of file prefixes to exclude from the files that are read. If a prefix includes a comma, escape the comma by using a backslash (\\).
exclude_missing_valuesSet values that have been defined as missing values to null
_file_format.delimited_syntax.field_delimiterSpecify a string or one of the following values: , , , . The string can include Unicode escape strings in the form \\uNNNN where NNNN is the Unicode character code.. Default: ,
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
_file_formatSpecify the format of the files to read or write. The implicit file format specifies that the input to the file connector is in binary or string format without a delimiter.. Values: [avro, comma-separated_value_csv, delimited, implicit, orc, parquet, sequencefile]. Default: delimited
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
_filename_source *Specify the name of the file to read from, or specify a pattern to read from multiple files.
file_name *The name of the file to read
_filename_columnSpecify the name of the column to write the source file name to.
first_lineIndicates at which row start reading. Default: 0
_file_format.delimited_syntax.headerSelect Yes if the first row of the file contains field headers and is not part of the data. If you select Yes, when the connector writes data, the field names will be the first row of the output. If runtime column propagation is enabled, metadata can be obtained from the first row of the file.. Default: false
_file_format.impl_syntax.headerSelect Yes if the first row of the file contains field headers and is not part of the data. If you select Yes, when the connector writes data, the field names will be the first row of the output. If runtime column propagation is enabled, metadata can be obtained from the first row of the file.. Default: false
first_line_headerIndicates whether the row where reading starts is the header. Default: false
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
_file_format.delimited_syntax.null_valueSpecify the character or string that represents null values in the data. For a source stage, input data that has the value that you specify is set to null on the output link. For a target stage, in the output file that is written to the file system, null values are represented by the value that is specified for this property. To specify that an empty string represents a null value, specify "" (two double quotation marks).
null_valueThe value that represents null (a missing value) in the file, for example, NULL
_file_format.avro_source.output_jsonSpecify if each rows in the avro file should be exported as JSON to a string column.. Default: false
_file_format.delimited_syntax.quotes. Values: [double, none, single]. Default: none
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
_read_modeSelect Read single file to read from a single file or Read multiple files to read from the files that match a specified file prefix. Select List buckets to list the buckets for your account in the specified region. Select List files to list files files for your account in the specified bucket.. Values: [read_multiple_files, read_single_file]. Default: read_single_file
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
_file_format.delimited_syntax.record_defSelect whether the record definition is provided to the connector from the source file, a delimited string, a file that contains a delimited string, or a schema file. When runtime column propagation is enabled, this metadata provides the column definitions. If a schema file is provided, the schema file overrides the values of formatting properties in the stage and the column definitions that are specified on the Columns page of the output link.. Values: [delimited_string, delimited_string_in_a_file, file_header, infer_schema, none, schema_file]. Default: none
_file_format.impl_syntax.record_defSelect whether the record definition is provided to the connector from the source file, a delimited string, a file that contains a delimited string, or a schema file. When runtime column propagation is enabled, this metadata provides the column definitions. If a schema file is provided, the schema file overrides the values of formatting properties in the stage and the column definitions that are specified on the Columns page of the output link.. Values: [delimited_string, delimited_string_in_a_file, file_header, none, schema_file]. Default: none
_reject_modeSpecify what the connector does when a record that contains invalid data is found in the source file. Select Continue to read the rest of the file, Fail to stop the job with an error message, or Reject to send the rejected data to a reject link.. Values: [continue, fail, reject]. Default: continue
_file_format.delimited_syntax.row_delimiterSpecify a string or one of the following values: , , , . The string can include Unicode escape strings in the form \\uNNNN where NNNN is the Unicode character code.. Default:
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
_file_format.delimited_syntax.record_limitSpecify the maximum number of records to read from the file per node. If a value is not specified for this property, the entire file is read.
_file_format.impl_syntax.record_limitSpecify the maximum number of records to read from the file per node. If a value is not specified for this property, the entire file is read.
row_limitThe maximum number of rows to return
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
_file_format.delimited_syntax.field_formats.time_formatSpecify a string that defines the format for fields that have the Time data type.
time_formatThe format of time values, for example, HH:mm:ss[.f]
_file_format.delimited_syntax.field_formats.timestamp_formatSpecify a string that defines the format for fields that have the Timestamp data type.
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
_file_format.trace_fileSpecify the full path to a file to contain trace information from the parser for delimited files. Because writing to a trace file requires additional processing, specify a value for this property only during job development.
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
_create_hive_table.additional_driver_paramsSpecify any additional driver-specific connection attributes. Enter the attributes in the name=value format, separated by semi-colon if multiple attributes needs to be specified. For information about the supported driver-specific attributes, refer to the Progress DataDirect driver documentation.
_wave_handling.append_uidUse this property to choose if a unique identifier is to be appended to the file name. When the value of this property is set to yes, then the file name gets appended with the unique identifier, and a new file would be written for every wave of data that is streamed into the stage. When the value of this property is set to No, then the file would be overwritten on every wave. Default: false
_file_format.avro_target.avro_array_keysIf the file format is Avro in a target stage, then normalization is controlled through array keys. Specify ''ITERATE()'' in the description for the corresponding array field in column definition in the input tab of file connector.
_file_format.avro_target.avro_codecSpecify the compression algorithm that will be used to compress the data.. Values: [bzip2, deflate, none, snappy]. Default: none
_file_format.avro_target.avro_schema *Specify the fully qualified path for a JSON file that defines the schema for the Avro file.
_file_format.parquet_target.parquet_block_sizeBlock size. Default: 10000000
_file_format.orc_target.orc_buffer_sizeBuffer Size. Default: 10000
_split_on_key.case_sensitiveSelect Yes to make the key value case sensitive.. Default: false
_cleanupIf a job fails, select whether the connector deletes the file or files that have been created.. Default: true
_file_format.orc_target.orc_compressSpecify Compression mechanism. Values: [none, snappy, zlib]. Default: snappy
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
_file_format.parquet_target.parquet_compressSpecify compression mechanism. Values: [gzip, lzo, none, snappy]. Default: snappy
create_hive_tableCreate a table in the database. Default: false
_create_hive_tableSelect Yes to create or use an existing Hive table after data has been loaded to HDFS.. Default: false
_create_hive_table.create_hive_schemaSpecify Yes to create the schema indicated in the fully qualified table name if it does not already exist. If Yes is specified and the table name does not contain a schema, the job will fail. If Yes is specified and the schema already exists, the job will not fail.. Default: false
_file_format.impl_syntax.binarySpecify the type of implicit file. Values: [binary]. Default: binary
_file_format.delimited_syntax.field_formats.date_formatSpecify a string that defines the format for fields that have the Date data type.
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
_file_format.delimited_syntax.field_formats.decimal_formatSpecify a string that defines the format for fields that have the Decimal or Numeric data type.
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
_create_hive_table.drop_hive_tableSpecify Yes to drop the Hive table if it already exists. No to append to existing Hive table.. Default: true
_create_hive_table.use_staging_table.hive_target_table_properties.hive_drop_staging_tableUse this property to drop the staging table. By default, the staging table would be dropped once the target table has been created. In case, the user do not want the staging table to be removed, set the value of this property to No. Default: true
_file_format.delimited_syntax.encodingSpecify the encoding of the files to read or write, for example, UTF-8.
_file_format.impl_syntax.encodingSpecify the encoding of the files to read or write, for example, UTF-8.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
_file_format.delimited_syntax.escapeSpecify the character to use to escape field and row delimiters. If an escape character exists in the data, the escape character is also escaped. Because escape characters require additional processing, do not specify a value for this property if you do not need to include escape characters in the data.
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
_split_on_key.exclude_part_stringSelect Yes to exclude the partition string each processing node appends to the file name.. Default: false
_file_format.delimited_syntax.field_delimiterSpecify a string or one of the following values: , , , . The string can include Unicode escape strings in the form \\uNNNN where NNNN is the Unicode character code.. Default: ,
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
_file_formatSpecify the format of the files to read or write. The implicit file format specifies that the input to the file connector is in binary or string format without a delimiter.. Values: [avro, comma-separated_value_csv, delimited, implicit, orc, parquet, sequencefile]. Default: delimited
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
_filename_target *Specify the name of the file to write to.
file_name *The name of the file to write to or delete
_wave_handling.file_size_thresholdSpecify the threshold for the file size in megabytes. Processing nodes will start a new file each time the size exceeds the value specified in the threshold and on reaching the wave boundary. The file will be written only on the wave boundary and hence the threshold value specified is only a soft limit. The actual size of the file can be higher than the specified threshold depending on the size of the wave. Default: 1
_file_format.delimited_syntax.headerSelect Yes if the first row of the file contains field headers and is not part of the data. If you select Yes, when the connector writes data, the field names will be the first row of the output. If runtime column propagation is enabled, metadata can be obtained from the first row of the file.. Default: false
_file_format.impl_syntax.headerSelect Yes if the first row of the file contains field headers and is not part of the data. If you select Yes, when the connector writes data, the field names will be the first row of the output. If runtime column propagation is enabled, metadata can be obtained from the first row of the file.. Default: false
first_line_headerIndicates whether the row where reading starts is the header. Default: false
_force_sequentialSelect Yes to run the connector sequentially on one node.. Default: false
hive_tableThe name of the table to create
_create_hive_table.hive_table_typeSpecify Hive table type, as external (default) or internal.. Values: [external, internal]. Default: external
_file_existsSpecify what the connector does when it tries to write a file that already exists. Select Overwrite file to overwrite a file if it already exists, Do not overwrite file to not overwrite the file and stop the job, or Fail to stop the job with an error message.. Values: [do_not_overwrite_file, fail, overwrite_file]. Default: overwrite_file
_file_format.delimited_syntax.encoding.output_bomSpecify whether to include a byte order mark in the file when the file encoding is a Unicode encoding such as UTF-8, UTF-16, or UTF-32.. Default: false
_file_format.impl_syntax.encoding.output_bomSpecify whether to include a byte order mark in the file when the file encoding is a Unicode encoding such as UTF-8, UTF-16, or UTF-32.. Default: false
_file_format.delimited_syntax.header.include_typesSelect Yes to append the data type to each field name that the connector writes in the first row of the output.. Default: false
_file_format.impl_syntax.header.include_typesSelect Yes to append the data type to each field name that the connector writes in the first row of the output.. Default: false
include_typesInclude data types in the first line of the file. Default: false
_file_format.avro_target.input_jsonSpecify if each rows in avro file should be imported from a JSON string.. Default: false
_split_on_key.key_columnSpecify the key column to use for splitting files. If not specified, the connector will use the first key column on the link.
_create_hive_table.use_staging_table.hive_target_table_properties.hive_target_table_locationUse this property to set the location of the HDFS files serving as storage for the Hive table
_max_file_sizeSpecify the maximum file size in megabytes. Processing nodes will start a new file each time the size exceeds this value.. Default: 0
_create_hive_table.use_staging_table.load_existing_table.max_dynamic_partitionsUse this property to set the maximum number of Dynamic paritions to be created while loading into a partitioned table.. Default: 1000
names_as_labelsSet column labels to the value of the column name
_file_format.delimited_syntax.null_valueSpecify the character or string that represents null values in the data. For a source stage, input data that has the value that you specify is set to null on the output link. For a target stage, in the output file that is written to the file system, null values are represented by the value that is specified for this property. To specify that an empty string represents a null value, specify "" (two double quotation marks).
null_valueThe value that represents null (a missing value) in the file, for example, NULL
_create_hive_table.use_staging_table.hive_target_table_properties.hive_orc_compressUse this property to set the compression type for the target table when the table format is ORC. Values: [none, snappy, zlib]. Default: zlib
_file_format.parquet_target.parquet_page_sizePage size. Default: 10000
_create_hive_table.use_staging_table.hive_target_table_properties.hive_parquet_compressUse this property to set the compression type for the target table when the table format is Parquet. Values: [gzip, lzo, none, snappy]. Default: snappy
partitionedWrite the file as multiple partitions. Default: false
_file_format.delimited_syntax.quotes. Values: [double, none, single]. Default: none
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
_file_format.delimited_syntax.row_delimiterSpecify a string or one of the following values: , , , . The string can include Unicode escape strings in the form \\uNNNN where NNNN is the Unicode character code.. Default:
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
_split_on_keySelect Yes to create a new file when key column value changes. Data must be sorted and partitioned for this to work properly.. Default: false
_file_format.orc_target.orc_stripe_sizeStripe Size. Default: 100000
_create_hive_table.use_staging_table.hive_target_table_properties.hive_orc_stripe_sizeStripe size. Default: 64
_create_hive_table.hive_table *Enter the name of the table to create.
_create_hive_table.use_staging_table.hive_target_table_properties.hive_target_table_formatUse this property to set the format of the target table. Values: [orc, parquet]. Default: parquet
_create_hive_table.use_staging_table.hive_target_table_properties.hive_target_table_typeUse this property to set the format of the type of the target table.. Values: [external, internal]. Default: external
_file_format.delimited_syntax.field_formats.time_formatSpecify a string that defines the format for fields that have the Time data type.
time_formatThe format of time values, for example, HH:mm:ss[.f]
_file_format.delimited_syntax.field_formats.timestamp_formatSpecify a string that defines the format for fields that have the Timestamp data type.
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
_split_on_key.key_in_filenameSelect Yes to use the key value in the generated file name.. Default: false
_create_hive_table.use_staging_tableSet Yes to use staging table. This option will be enabled only when the FileFormat is Delimited. Default: false
sheet_nameThe name of the Excel worksheet to write to
_write_modeSelect Write single file to write a file per node, select Write multiple files to write multiple files per node (based on size and/or key value), or select Delete to delete files.. Values: [delete, write_multiple_files, write_single_file]. Default: write_single_file
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




Apache Hive


Description: Apache Hive is a data warehouse software project that provides data query and analysis and is built on top of Apache Hadoop.
Data source type ID: 0fd83fe5-8995-4e2e-a1be-679bb8813a6d
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
http_pathThe path of the endpoint such as gateway/default/hive when the server is configured for HTTP transport mode
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
row_limit_supportEnable if connector should append limit operator to the queries.. Default: true
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
_before_after._after_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node after all of the data is processed on that node.
_before_after._after_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once after all of the data is processed.
_before_after._before_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node before any data is processed on that node.
_before_after._before_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once before any data is processed.
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
_session._character_set_for_non_unicode_columnsSelect the character set option for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode. If you select the Default option, the character set encoding of the engine host system locale is used. If you select the Custom option, you must provide the character set name to be used.. Values: [_custom, _default]. Default: _default
_session._character_set_for_non_unicode_columns._character_set_name *Specify the name of the character set encoding for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
_session._default_length_for_columnsEnter the default length for the Char, NChar, Binary, VarChar, NVarChar, and VarBinary link columns for which the Length attribute is not set.. Default: 200
_session._default_length_for_long_columnsEnter the default length for the LongVarChar, LongNVarChar and LongVarBinary link columns for which the Length attribute is not set.. Default: 20000
_enable_partitioned_readsSelect Yes to run the statement on each processing node. When using 'Database partition' as Partitioned read method the statement should consist of a where clause = [[part-value]], where partition-column is the partition from which you want to query and the [[part-value]] is the placeholder which is replaced on each processing node with the actual partition values to prepare the query. A simple example is SELECT c1,c2 from part_table where pc1=[[part-value]]. When using 'Modulus', [[mod-part-expr]] should be used. Example: SELECT * from part_table where [[mod-part-value]]. With 'Minimum and Maximum range' the placeholder is [[range-part-expr]]. Example: SELECT * from part_table where [[range-part-value]] . Default: false
_enable_quoted_idsSelect Yes to enclose the specified table name and column names on the links in quoting strings when SQL statements are generated. The connector queries the driver to determine the quoting string. If it fails to obtain this information from the driver, the connector uses the backtick (`) character as the quoting string. The default is No.. Default: false
_session._fetch_sizeSpecify the number of rows that the driver must try to fetch from the data source when the connector requests a single row. Fetching rows in addition to the row requested by the connector can improve performance because the driver can complete the subsequent requests for more rows from the connector locally without a need to access the data source. The default value is 0, which indicates that the driver optimizes the fetch operation based on its internal logic.. Default: 0
_generate_sqlSelect Yes to automatically generate the SQL statements at run time.. Default: true
_session._generate_all_columns_as_unicodeAlways generate columns as NChar, NVarChar and LongNVarChar columns instead of Char, VarChar and LongVarChar columns.. Default: false
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
_session._keep_conductor_connection_aliveSelect Yes to keep the connection alive in the conductor process while the player processes are processing records. Select No to close the connection in the conductor process before player processes start processing records, and to connect again if necessary after the player processes complete processing the records.. Default: true
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
_transaction._end_of_waveSelect Yes to generate an end-of-wave record after each wave of records, where the number of records in each wave is specified in the Record count property. When the Record count property is set to 0, the end-of-wave records are not generated.. Values: [_no, _yes]. Default: _no
_enable_partitioned_reads._partition_methodUse this property to set the type of the partitioned to be used when the partitioned reads is enabled.. Values: [_hive_partition, _minimum_and_maximum_range, _modulus]. Default: _hive_partition
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
_transaction._record_countSpecify the number of rows that the stage reads from or writes to the data source in a single transaction. When this property is set to 0, the transaction is committed only once on each processing node of the stage after the stage processes all the rows on that node. When rows arrive on the input link of the stage in waves, the Record count value applies to each wave separately.. Default: 2000
_session._report_schema_mismatchSelect Yes to perform early comparison of the column definitions on the link with the column definitions in the data source and to issue warning messages for any detected discrepancies which can result in data corruption. Depending on the environment and the usage scenario the early detection of discrepancies may not be possible in which case the error messages are reported only when the actual data corruption is detected. . Default: false
_limit_rows._limitEnter the maximum number of rows to be returned by the connector or each node when Partition Read is enabled.
row_limitThe maximum number of rows to return
_before_afterSelect Yes to run SQL statements before and after data is accessed in the database.. Default: false
sampling_percentagePercentage for each row or block to be included in the sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none]. Default: none
schema_nameThe name of the schema that contains the table to read from
_select_statement *Enter a SELECT statement or the fully qualified name of the file that contains the SELECT statement. The statement is used to read rows from the database.
select_statement *The SQL SELECT statement for retrieving data from the table
_hive_parametersEnter the statement to set the database parameters.
_before_after._after_sql_node._fail_on_errorSelect Yes to stop the job if the After SQL (node) statement fails.. Default: true
_before_after._after_sql._fail_on_errorSelect Yes to stop the job if the After SQL statement fails.. Default: true
_before_after._before_sql_node._fail_on_errorSelect Yes to stop the job if the Before SQL (node) statement fails.. Default: true
_before_after._before_sql._fail_on_errorSelect Yes to stop the job if the Before SQL statement fails.. Default: true
_hive_parameters._fail_on_errorSelect Yes to stop the job if the database parameters are not set.. Default: false
_table_name *Enter the fully qualified name of the table that you want to access in the data source.
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
_before_after._after_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node after all of the data is processed on that node.
_before_after._after_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once after all of the data is processed.
_session._batch_sizeEnter the number of records to include in the batch of records for each statement execution. The value 0 indicates that all input records are passed to the statements in a single batch.. Default: 2000
_before_after._before_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node before any data is processed on that node.
_before_after._before_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once before any data is processed.
_session._character_set_for_non_unicode_columnsSelect the character set option for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode. If you select the Default option, the character set encoding of the engine host system locale is used. If you select the Custom option, you must provide the character set name to be used.. Values: [_custom, _default]. Default: _default
_session._character_set_for_non_unicode_columns._character_set_name *Specify the name of the character set encoding for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode.
_table_action._generate_create_statement._create_statement *Enter the CREATE TABLE statement to run to create the target database table.
_custom_statementsCustom statements to be run for each input row
_table_action._generate_create_statement._storage_formatUse this property to specify the storage format of the file that stores the data in the table.. Values: [_avro, _orc, _parquet, _rc_file, _sequence_file, _text_file]. Default: _text_file
_session._default_length_for_columnsEnter the default length for the Char, NChar, Binary, VarChar, NVarChar, and VarBinary link columns for which the Length attribute is not set.. Default: 200
_session._default_length_for_long_columnsEnter the default length for the LongVarChar, LongNVarChar and LongVarBinary link columns for which the Length attribute is not set.. Default: 20000
_delete_statement *Enter a DELETE statement or the fully qualified name of the file that contains a DELETE statement. The statement is used to delete rows from the database.
_table_action._generate_drop_statement._drop_statement *Enter the DROP TABLE statement to run to drop the target database table.
_session._drop_unmatched_fieldsSelect Yes to drop any fields from the input link for which there are no matching parameters in the statements configured for the stage. Select No to issue error message when an unmatched field is present on the link.. Default: false
_enable_partitioned_writeSelect Yes to insert data into partitioned table. In the insert query, ORCHESTRATE. would be replaced by the column data passed from input link.. Default: false
_enable_quoted_idsSelect Yes to enclose the specified table name and column names on the links in quoting strings when SQL statements are generated. The connector queries the driver to determine the quoting string. If it fails to obtain this information from the driver, the connector uses the backtick (`) character as the quoting string. The default is No.. Default: false
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
_table_action._generate_create_statement._row_format._field_terminatorUse this property to specify the field terminator to create the table.
file_formatThe format of the file to write to. Values: [avro, csv, delimited, orc, parquet]. Default: delimited
file_nameThe name of the file to write to or delete
_generate_sqlSelect Yes to automatically generate the SQL statements at run time.. Default: true
_table_action._generate_create_statementSelect Yes to automatically generate the CREATE TABLE statement at run time. Depending on the input link column data types, the driver, and the data source, the connector might not be able to determine the corresponding native data types and produce a valid statement.. Default: true
_table_action._generate_drop_statementSelect Yes to automatically generate the DROP TABLE statement at run time.. Default: true
_table_action._generate_truncate_statementSelect Yes to automatically generate the TRUNCATE TABLE statement at run time.. Default: true
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
_insert_statement *Enter an INSERT statement or the fully qualified name of the file that contains an INSERT statement. The statement is used to insert rows into the database.
_session._keep_conductor_connection_aliveSelect Yes to keep the connection alive in the conductor process while the player processes are processing records. Select No to close the connection in the conductor process before player processes start processing records, and to connect again if necessary after the player processes complete processing the records.. Default: true
_table_action._generate_create_statement._row_format._line_terminatorUse this property to specify the line terminator to create the table.
null_valueThe value that represents null (a missing value) in the file, for example, NULL
_table_action._table_action_firstSelect Yes to perform the table action first. Select No to run the Before SQL statements first.. Default: true
_transaction._record_countSpecify the number of rows that the stage reads from or writes to the data source in a single transaction. When this property is set to 0, the transaction is committed only once on each processing node of the stage after the stage processes all the rows on that node. When rows arrive on the input link of the stage in waves, the Record count value applies to each wave separately.. Default: 2000
_session._report_schema_mismatchSelect Yes to perform early comparison of the column definitions on the link with the column definitions in the data source and to issue warning messages for any detected discrepancies which can result in data corruption. Depending on the environment and the usage scenario the early detection of discrepancies may not be possible in which case the error messages are reported only when the actual data corruption is detected. . Default: false
_table_action._generate_create_statement._row_formatSelect row format option for table creation.. Values: [_delimited, _ser_de, _storage_format]. Default: _storage_format
_before_afterSelect Yes to run SQL statements before and after data is accessed in the database.. Default: false
schema_nameThe name of the schema that contains the table to write to
_table_action._generate_create_statement._row_format._serde_library *Use this property to specify the library name for SerDe for creating the table
_hive_parametersEnter the statement to set the database parameters.
static_statement *The SQL used for setup operations, for example a CREATE statement
_before_after._after_sql_node._fail_on_errorSelect Yes to stop the job if the After SQL (node) statement fails.. Default: true
_before_after._after_sql._fail_on_errorSelect Yes to stop the job if the After SQL statement fails.. Default: true
_before_after._before_sql_node._fail_on_errorSelect Yes to stop the job if the Before SQL (node) statement fails.. Default: true
_before_after._before_sql._fail_on_errorSelect Yes to stop the job if the Before SQL statement fails.. Default: true
_table_action._generate_drop_statement._fail_on_errorSelect Yes to stop the job if the DROP TABLE statement fails.. Default: false
_table_action._generate_create_statement._fail_on_errorSelect Yes to stop the job if the CREATE TABLE statement fails.. Default: true
_table_action._generate_truncate_statement._fail_on_errorSelect Yes to stop the job if the TRUNCATE TABLE statement fails.. Default: true
_hive_parameters._fail_on_errorSelect Yes to stop the job if the database parameters are not set.. Default: false
_table_action *Select the action to complete before writing data to the table.. Values: [_append, _create, _replace, _truncate]. Default: _append
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace]. Default: append
_table_action._generate_create_statement._table_locationUse this property to specify the location of the file that serves as storage for the table.
_table_name *Enter the fully qualified name of the table that you want to access in the data source.
table_nameThe name of the table to write to
_table_action._generate_truncate_statement._truncate_statement *Enter the TRUNCATE TABLE statement to run to truncate the target database table.
_update_statement *Enter an UPDATE statement or the fully qualified name of the file that contains an UPDATE statement. The statement is used to update rows in the database.
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
_write_modeSelect the mode that you want to use to write to the data source.. Values: [_custom, _delete, _insert, _update]. Default: _insert
write_modeThe mode for writing records to the target table. Values: [insert, static_statement, update_statement, update_statement_table_action]. Default: insert




Apache Kafka


Description: Apache Kafka is an open-source distributed event streaming platform.
Data source type ID: f13bc9b7-4a46-48f4-99c3-01d943334ba7
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
schema_registry_authenticationSelect option to provide credentials for authentication. Values: [none, reuse_sasl_credentials, user_credentials]. Default: none
server_name *Specify the host name and port for the virtual Kafka operational server [HostName:Port]. Use comma to separate multiple servers.
key_pemPrivate key in the PEM format. Only PKCS#8 keys are supported. If the key is encrypted, key password must be specified using 'Key password' property.
registry_key_pemPrivate key in the PEM format. Only PKCS#8 keys are supported. If the key is encrypted, key password must be specified using 'Key password' property.
key_chain_pemCertificates chain for private key in the PEM format. Only X.509 certificates are supported.
registry_key_chain_pemCertificates chain for private key in the PEM format. Only X.509 certificates are supported.
key_passwordPassword for key file
registry_key_passwordPassword for key
passwordSpecify the password to use to connect to the virtual Kafka operational server.
registry_passwordPassword
schema_registry_urlSchema Registry service URL
schema_registry_secureSelect type of secure connection to schema registry. Values: [none, ssl, reuse_ssl]. Default: none
secure_connectionType of secure connection to Kafka operational server. Values: [None, SASL_PLAIN, SASL_SSL, SCRAM-SHA-256, SCRAM-SHA-512, SSL]. Default: None
registry_truststore_pemTrusted certificates in the PEM format. Only X.509 certificates are supported.
truststore_pemTrusted certificates in the PEM format. Only X.509 certificates are supported.
use_schema_registryUse Schema Registry service for message format definition. Default: false
registry_usernameUser name
usernameSpecify the user name to be use to connect to the kerberized kafka server/cluster.

Interaction properties (when used as a source)

Name Type Description
_advanced_kafka_config_optionsAdvanced Kafka Client configuration. Default: false
_advanced_client_loggingAdvanced Kafka Client logging. Default: false
schema_registry_authenticationSelect option to provide credentials for authentication. Values: [none, reuse_sasl_credentials, user_credentials]. Default: none
_kafka_config_optionsAdditional Kafka Client configuration options. Depending on the context in which Kafka Connector stage is used, either Kafka Producer or Kafka Consumer properties should be provided. The value of this multiline property must conform Java Properties class requirements.
advanced_kafka_config_optionsAdditional Kafka Client configuration options. Depending on the context in which Kafka Connector stage is used, either Kafka Producer or Kafka Consumer properties should be provided. The value of this multiline property must conform Java Properties class requirements.
_consumer_group_nameConsumer group name to be used when reading messages from Kafka topic.
consumer_group_nameConsumer group name to be used when reading messages from Kafka topic.
continuous_modeChoose continuous mode on or off. Default: false
end_of_dataSpecify whether to insert EOW marker for the final set of records when their number is smaller than the value specified for the transaction record count. Note that if the specified transaction record count value is 0 (representing all available records), there is only one transaction wave which consists of all the records, and so the End of data value should be set to Yes in order for EOW marker to be inserted for that transaction wave. Default: false
end_of_waveSpecify whether to insert EOW marker. Default: false
heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
_isolation_levelKafka isolation level for messages written transactionally. Read commited will only fetch commited messages, Read uncommitted will fetch all of them.. Values: [read_committed, read_uncommitted]. Default: read_uncommitted
isolation_levelKafka isolation level for messages written transactionally. Read commited will only fetch commited messages, Read uncommitted will fetch all of them.. Values: [read_committed, read_uncommitted]. Default: read_uncommitted
_start_offsetStarting offset per each partition. [partition:offset,partition:offset,...] or [offset,offset,offset]
start_offsetStarting offset per each partition. [partition:offset,partition:offset,...] or [offset,offset,offset]
registry_key_pemPrivate key in the PEM format. Only PKCS#8 keys are supported. If the key is encrypted, key password must be specified using 'Key password' property.
registry_key_chain_pemCertificates chain for private key in the PEM format. Only X.509 certificates are supported.
_key_serializer_typeSpecify the type of data key to use appropriate serializer or deserializer. Values: [avro, avro_to_json, byte, double, integer, small_integer, string]. Default: string
key_serializer_typeSpecify the type of data key to use appropriate serializer or deserializer. Values: [avro, avro_to_json, byte, double, integer, small_integer, string]. Default: string
_client_logging_levelLogging level. Values: [debug, error, fatal, info, off, trace, warn]. Default: off
kafka_client_logging_levelMinimum logging level of messages from Kafka Client that will be written in the job log. Each entry that is read from Kafka Client has special prefix [KAFKA].. Values: [debug, error, fatal, info, off, trace, warn]. Default: off
max_messagesMaximum number of messages to be produced to the topic on a per player process basis. This should be a multiple of max poll records.. Default: 100
max_poll_recordsMaximum records to be fetched in a single poll.. Default: 100
_max_poll_recordsMaximum records to be fetched in a single poll.. Default: 100
_timeoutSpecify the time in seconds after which the consumer would not poll for records.
registry_passwordPassword
record_countNumber of records per transaction. The value 0 means all available records. Default: 0
schema_registry_urlSpecify schema registry service URL
_reset_policySet reset policy (must be either 'earliest', 'latest' (default: latest). Values: [earliest, latest]. Default: latest
reset_policySet reset policy (must be either 'earliest', 'latest' (default: latest). Values: [earliest, latest]. Default: earliest
schema_registry_secureSelect type of secure connection to schema registry. Values: [none, ssl, reuse_ssl]. Default: none
stop_messageRegular expression, which if matched, will stop continuous mode
_stop_messageRegular expression, which if matched, will stop continuous mode
time_intervalTime interval for transaction. Default: 0
timeoutSpecify the time in seconds after which the consumer would not poll for records.. Default: 30
_timeout_after_last_messageTimeout after last message (secs). Default: 30
topic_name *Kafka topic name.
_max_messagesTotal number of messages
registry_truststore_pemTrusted certificates in the PEM format. Only X.509 certificates are supported.
registry_usernameUser name
_value_serializer_typeSpecify the type of data to use appropriate serializer or deserializer. Values: [avro, avro_to_json, byte, double, integer, small_integer, string]. Default: string
value_serializer_typeSpecify the type of data to use appropriate serializer or deserializer. Values: [avro, avro_to_json, byte, double, integer, small_integer, string]. Default: string
kafka_warning_and_error_logsDefines how messages with severity higher than WARN (WARN, ERROR, FATAL) will be written in the job log.. Values: [keep_severity, log_as_informational, log_as_warning]. Default: log_as_informational

Interaction properties (when used as a target)

Name Type Description
_advanced_kafka_config_optionsAdvanced Kafka Client configuration. Default: false
_advanced_client_loggingAdvanced Kafka Client logging. Default: false
schema_registry_authenticationSelect option to provide credentials for authentication. Values: [none, reuse_sasl_credentials, user_credentials]. Default: none
_kafka_config_optionsAdditional Kafka Client configuration options. Depending on the context in which Kafka Connector stage is used, either Kafka Producer or Kafka Consumer properties should be provided. The value of this multiline property must conform Java Properties class requirements.
advanced_kafka_config_optionsAdditional Kafka Client configuration options. Depending on the context in which Kafka Connector stage is used, either Kafka Producer or Kafka Consumer properties should be provided. The value of this multiline property must conform Java Properties class requirements.
heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
registry_key_pemPrivate key in the PEM format. Only PKCS#8 keys are supported. If the key is encrypted, key password must be specified using 'Key password' property.
registry_key_chain_pemCertificates chain for private key in the PEM format. Only X.509 certificates are supported.
_key_serializer_typeSpecify the type of data key to use appropriate serializer or deserializer. Values: [avro, avro_to_json, byte, double, integer, small_integer, string]. Default: string
key_serializer_typeSpecify the type of data key to use appropriate serializer or deserializer. Values: [avro, avro_to_json, byte, double, integer, small_integer, string]. Default: string
_client_logging_levelLogging level. Values: [debug, error, fatal, info, off, trace, warn]. Default: off
kafka_client_logging_levelMinimum logging level of messages from Kafka Client that will be written in the job log. Each entry that is read from Kafka Client has special prefix [KAFKA].. Values: [debug, error, fatal, info, off, trace, warn]. Default: off
registry_passwordPassword
schema_registry_urlSpecify schema registry service URL
schema_registry_secureSelect type of secure connection to schema registry. Values: [none, ssl, reuse_ssl]. Default: none
topic_name *Kafka topic name.
registry_truststore_pemTrusted certificates in the PEM format. Only X.509 certificates are supported.
registry_usernameUser name
_value_serializer_typeSpecify the type of data to use appropriate serializer or deserializer. Values: [avro, avro_to_json, byte, double, integer, small_integer, string]. Default: string
value_serializer_typeSpecify the type of data to use appropriate serializer or deserializer. Values: [avro, avro_to_json, byte, double, integer, small_integer, string]. Default: string
kafka_warning_and_error_logsDefines how messages with severity higher than WARN (WARN, ERROR, FATAL) will be written in the job log.. Values: [keep_severity, log_as_informational, log_as_warning]. Default: log_as_informational




Box


Description: The Box platform is a cloud content management and file sharing service.
Data source type ID: 99c3c67b-2133-4006-81f6-2b375a0048a3
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
client_id *The client ID (username) for authorizing access to Box
client_secret *The password associated with the client ID for authorizing access to Box
enterprise_id *The ID for your organization
private_key *The private key that was generated and provided to you by Box
private_key_password *The password associated with the private key that was generated and provided to you by Box
public_key *The public key that was generated and provided to you by Box
usernameThe username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
display_value_labelsDisplay the value labels
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_name *The name of the file to read
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
null_valueThe value that represents null (a missing value) in the file, for example, NULL
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_name *The name of the file to write to or delete
first_line_headerIndicates whether the row where reading starts is the header. Default: false
include_typesInclude data types in the first line of the file. Default: false
names_as_labelsSet column labels to the value of the column name
null_valueThe value that represents null (a missing value) in the file, for example, NULL
partitionedWrite the file as multiple partitions. Default: false
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
sheet_nameThe name of the Excel worksheet to write to
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




Cloudera Impala


Description: Cloudera Impala provides SQL queries directly on your Apache Hadoop data stored in HDFS or HBase.
Data source type ID: 05c58384-862e-4597-b19a-c71ea7e760bc
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
_before_after._after_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node after all of the data is processed on that node.
_before_after._after_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once after all of the data is processed.
_before_after._before_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node before any data is processed on that node.
_before_after._before_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once before any data is processed.
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
_session._character_set_for_non_unicode_columnsSelect the character set option for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode. If you select the Default option, the character set encoding of the engine host system locale is used. If you select the Custom option, you must provide the character set name to be used.. Values: [_custom, _default]. Default: _default
_session._character_set_for_non_unicode_columns._character_set_name *Specify the name of the character set encoding for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
_session._default_length_for_columnsEnter the default length for the Char, NChar, Binary, VarChar, NVarChar, and VarBinary link columns for which the Length attribute is not set.. Default: 200
_session._default_length_for_long_columnsEnter the default length for the LongVarChar, LongNVarChar and LongVarBinary link columns for which the Length attribute is not set.. Default: 20000
_enable_partitioned_readsSelect Yes to run the statement on each processing node. When using 'Database partition' as Partitioned read method the statement should consist of a where clause = [[part-value]], where partition-column is the partition from which you want to query and the [[part-value]] is the placeholder which is replaced on each processing node with the actual partition values to prepare the query. A simple example is SELECT c1,c2 from part_table where pc1=[[part-value]]. When using 'Modulus', [[mod-part-expr]] should be used. Example: SELECT * from part_table where [[mod-part-value]]. With 'Minimum and Maximum range' the placeholder is [[range-part-expr]]. Example: SELECT * from part_table where [[range-part-value]] . Default: false
_enable_quoted_idsSelect Yes to enclose the specified table name and column names on the links in quoting strings when SQL statements are generated. The connector queries the driver to determine the quoting string. If it fails to obtain this information from the driver, the connector uses the backtick (`) character as the quoting string. The default is No.. Default: false
_session._fetch_sizeSpecify the number of rows that the driver must try to fetch from the data source when the connector requests a single row. Fetching rows in addition to the row requested by the connector can improve performance because the driver can complete the subsequent requests for more rows from the connector locally without a need to access the data source. The default value is 0, which indicates that the driver optimizes the fetch operation based on its internal logic.. Default: 0
_generate_sqlSelect Yes to automatically generate the SQL statements at run time.. Default: true
_session._generate_all_columns_as_unicodeAlways generate columns as NChar, NVarChar and LongNVarChar columns instead of Char, VarChar and LongVarChar columns.. Default: false
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
_session._keep_conductor_connection_aliveSelect Yes to keep the connection alive in the conductor process while the player processes are processing records. Select No to close the connection in the conductor process before player processes start processing records, and to connect again if necessary after the player processes complete processing the records.. Default: true
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
_transaction._end_of_waveSelect Yes to generate an end-of-wave record after each wave of records, where the number of records in each wave is specified in the Record count property. When the Record count property is set to 0, the end-of-wave records are not generated.. Values: [_no, _yes]. Default: _no
_enable_partitioned_reads._partition_methodUse this property to set the type of the partitioned to be used when the partitioned reads is enabled.. Values: [_hive_partition, _minimum_and_maximum_range, _modulus]. Default: _hive_partition
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
_transaction._record_countSpecify the number of rows that the stage reads from or writes to the data source in a single transaction. When this property is set to 0, the transaction is committed only once on each processing node of the stage after the stage processes all the rows on that node. When rows arrive on the input link of the stage in waves, the Record count value applies to each wave separately.. Default: 2000
_session._report_schema_mismatchSelect Yes to perform early comparison of the column definitions on the link with the column definitions in the data source and to issue warning messages for any detected discrepancies which can result in data corruption. Depending on the environment and the usage scenario the early detection of discrepancies may not be possible in which case the error messages are reported only when the actual data corruption is detected. . Default: false
_limit_rows._limitEnter the maximum number of rows to be returned by the connector or each node when Partition Read is enabled.
row_limitThe maximum number of rows to return
_before_afterSelect Yes to run SQL statements before and after data is accessed in the database.. Default: false
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none]. Default: none
schema_nameThe name of the schema that contains the table to read from
_select_statement *Enter a SELECT statement or the fully qualified name of the file that contains the SELECT statement. The statement is used to read rows from the database.
select_statement *The SQL SELECT statement for retrieving data from the table
_before_after._after_sql_node._fail_on_errorSelect Yes to stop the job if the After SQL (node) statement fails.. Default: true
_before_after._after_sql._fail_on_errorSelect Yes to stop the job if the After SQL statement fails.. Default: true
_before_after._before_sql_node._fail_on_errorSelect Yes to stop the job if the Before SQL (node) statement fails.. Default: true
_before_after._before_sql._fail_on_errorSelect Yes to stop the job if the Before SQL statement fails.. Default: true
_table_name *Enter the fully qualified name of the table that you want to access in the data source.
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
_before_after._after_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node after all of the data is processed on that node.
_before_after._after_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once after all of the data is processed.
_session._batch_sizeEnter the number of records to include in the batch of records for each statement execution. The value 0 indicates that all input records are passed to the statements in a single batch.. Default: 2000
_before_after._before_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node before any data is processed on that node.
_before_after._before_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once before any data is processed.
_session._character_set_for_non_unicode_columnsSelect the character set option for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode. If you select the Default option, the character set encoding of the engine host system locale is used. If you select the Custom option, you must provide the character set name to be used.. Values: [_custom, _default]. Default: _default
_session._character_set_for_non_unicode_columns._character_set_name *Specify the name of the character set encoding for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode.
_table_action._generate_create_statement._create_statement *Enter the CREATE TABLE statement to run to create the target database table.
_custom_statementsCustom statements to be run for each input row
_table_action._generate_create_statement._storage_formatUse this property to specify the storage format of the file that stores the data in the table.. Values: [_avro, _orc, _parquet, _rc_file, _sequence_file, _text_file]. Default: _text_file
_session._default_length_for_columnsEnter the default length for the Char, NChar, Binary, VarChar, NVarChar, and VarBinary link columns for which the Length attribute is not set.. Default: 200
_session._default_length_for_long_columnsEnter the default length for the LongVarChar, LongNVarChar and LongVarBinary link columns for which the Length attribute is not set.. Default: 20000
_delete_statement *Enter a DELETE statement or the fully qualified name of the file that contains a DELETE statement. The statement is used to delete rows from the database.
_table_action._generate_drop_statement._drop_statement *Enter the DROP TABLE statement to run to drop the target database table.
_session._drop_unmatched_fieldsSelect Yes to drop any fields from the input link for which there are no matching parameters in the statements configured for the stage. Select No to issue error message when an unmatched field is present on the link.. Default: false
_enable_partitioned_writeSelect Yes to insert data into partitioned table. In the insert query, ORCHESTRATE. would be replaced by the column data passed from input link.. Default: false
_enable_quoted_idsSelect Yes to enclose the specified table name and column names on the links in quoting strings when SQL statements are generated. The connector queries the driver to determine the quoting string. If it fails to obtain this information from the driver, the connector uses the backtick (`) character as the quoting string. The default is No.. Default: false
_table_action._generate_create_statement._row_format._field_terminatorUse this property to specify the field terminator to create the table.
_generate_sqlSelect Yes to automatically generate the SQL statements at run time.. Default: true
_table_action._generate_create_statementSelect Yes to automatically generate the CREATE TABLE statement at run time. Depending on the input link column data types, the driver, and the data source, the connector might not be able to determine the corresponding native data types and produce a valid statement.. Default: true
_table_action._generate_drop_statementSelect Yes to automatically generate the DROP TABLE statement at run time.. Default: true
_table_action._generate_truncate_statementSelect Yes to automatically generate the TRUNCATE TABLE statement at run time.. Default: true
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
_insert_statement *Enter an INSERT statement or the fully qualified name of the file that contains an INSERT statement. The statement is used to insert rows into the database.
_session._keep_conductor_connection_aliveSelect Yes to keep the connection alive in the conductor process while the player processes are processing records. Select No to close the connection in the conductor process before player processes start processing records, and to connect again if necessary after the player processes complete processing the records.. Default: true
_table_action._generate_create_statement._row_format._line_terminatorUse this property to specify the line terminator to create the table.
_table_action._table_action_firstSelect Yes to perform the table action first. Select No to run the Before SQL statements first.. Default: true
_transaction._record_countSpecify the number of rows that the stage reads from or writes to the data source in a single transaction. When this property is set to 0, the transaction is committed only once on each processing node of the stage after the stage processes all the rows on that node. When rows arrive on the input link of the stage in waves, the Record count value applies to each wave separately.. Default: 2000
_session._report_schema_mismatchSelect Yes to perform early comparison of the column definitions on the link with the column definitions in the data source and to issue warning messages for any detected discrepancies which can result in data corruption. Depending on the environment and the usage scenario the early detection of discrepancies may not be possible in which case the error messages are reported only when the actual data corruption is detected. . Default: false
_table_action._generate_create_statement._row_formatSelect row format option for table creation.. Values: [_delimited, _ser_de, _storage_format]. Default: _storage_format
_before_afterSelect Yes to run SQL statements before and after data is accessed in the database.. Default: false
schema_nameThe name of the schema that contains the table to write to
_table_action._generate_create_statement._row_format._serde_library *Use this property to specify the library name for SerDe for creating the table
static_statement *The SQL used for setup operations, for example a CREATE statement
_before_after._after_sql_node._fail_on_errorSelect Yes to stop the job if the After SQL (node) statement fails.. Default: true
_before_after._after_sql._fail_on_errorSelect Yes to stop the job if the After SQL statement fails.. Default: true
_before_after._before_sql_node._fail_on_errorSelect Yes to stop the job if the Before SQL (node) statement fails.. Default: true
_before_after._before_sql._fail_on_errorSelect Yes to stop the job if the Before SQL statement fails.. Default: true
_table_action._generate_drop_statement._fail_on_errorSelect Yes to stop the job if the DROP TABLE statement fails.. Default: false
_table_action._generate_create_statement._fail_on_errorSelect Yes to stop the job if the CREATE TABLE statement fails.. Default: true
_table_action._generate_truncate_statement._fail_on_errorSelect Yes to stop the job if the TRUNCATE TABLE statement fails.. Default: true
_table_action *Select the action to complete before writing data to the table.. Values: [_append, _create, _replace, _truncate]. Default: _append
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace]. Default: append
_table_action._generate_create_statement._table_locationUse this property to specify the location of the file that serves as storage for the table.
_table_name *Enter the fully qualified name of the table that you want to access in the data source.
table_name *The name of the table to write to
_table_action._generate_truncate_statement._truncate_statement *Enter the TRUNCATE TABLE statement to run to truncate the target database table.
_update_statement *Enter an UPDATE statement or the fully qualified name of the file that contains an UPDATE statement. The statement is used to update rows in the database.
_write_modeSelect the mode that you want to use to write to the data source.. Values: [_custom, _delete, _insert, _update]. Default: _insert
write_modeThe mode for writing records to the target table. Values: [insert, static_statement, update_statement, update_statement_table_action]. Default: insert




Custom No-op


Description: Base data source type for a customizable no-op data source type
Data source type ID: dca613ef-5e34-4eca-9a80-fedcf9122834
Can be used as a source: No
Can be used as a target: No
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description

Interaction properties (when used as a source)

Name Type Description

Interaction properties (when used as a target)

Name Type Description




Dremio


Description: Dremio is an open data lake platform. It supports all the major third-party data sources.
Data source type ID: cca8fcca-71f9-4e4f-a8fb-89fd2c0072cf
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
project_idThe ID of the Dremio Cloud project
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Dropbox


Description: Dropbox is a cloud storage service for storing, sharing, and synchronizing files across devices.
Data source type ID: 507b850c-f4a1-41d7-ad64-4182a1264014
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
access_token *The OAuth2 access token that you obtained by following the instructions at https://www.dropbox.com/developers/reference/oauth-guide

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
display_value_labelsDisplay the value labels
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_name *The name of the file to read
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
null_valueThe value that represents null (a missing value) in the file, for example, NULL
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_name *The name of the file to write to or delete
first_line_headerIndicates whether the row where reading starts is the header. Default: false
include_typesInclude data types in the first line of the file. Default: false
names_as_labelsSet column labels to the value of the column name
null_valueThe value that represents null (a missing value) in the file, for example, NULL
partitionedWrite the file as multiple partitions. Default: false
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
sheet_nameThe name of the Excel worksheet to write to
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




Elasticsearch


Description: Elasticsearch is a distributed, open-source search and analytics engine. Use the Elasticsearch connector to access JSON documents in Elasticsearch indexes.
Data source type ID: 200d71ab-24a5-4b3d-85a4-a365bdd0d4cb
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
password *The password associated with the username for accessing the data source
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
url *The URL to access Elasticsearch
use_anonymous_accessConnect without providing logon credentials. Default: false
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
file_name *The name of the file to read
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
query_bodyJSON containing the body of a search request
query_stringSearch query in the Lucene query string syntax

Interaction properties (when used as a target)

Name Type Description
create_index_bodyJSON containing the body of a create index request
document_typeThe type of the document
file_actionThe action to take on the target file to handle the new data set. Values: [append, replace, truncate]. Default: append
file_name *The name of the file to write to or delete
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




FTP


Description: FTP is a standard communication protocol that is used to transfer files from a server to a client on a computer network.
Data source type ID: d5dbc62f-7c4c-4d49-8eb2-dab6cef2969c
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
mvs_datasetCheck, if MVS Dataset is to be accessed.
auth_methodAuthentication method. If you use an encrypted private key, you will need a key passphrase.. Values: [username_password, username_password_key, username_key]
connection_mode *Connection mode. Values: [anonymous, basic, mvssftp, sftp, ftps]
ftadvSpecify File Transfer Advice strings as comma-delimited key-value pairs.
host *The hostname or IP address of the remote FTP server
key_passphraseIf the private key is encrypted, this passphrase is needed to decrypt/encrypt it
passwordThe password associated with the username for connecting to the FTP server
portThe port of the FTP server
private_keyThe private key for your account. The key must be an RSA private key that is generated by the ssh-keygen tool and it must be in the PEM format. If the private key is encrypted, you will need a key passphrase.
username *The username for connecting to the FTP server. Default: anonymous

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
display_value_labelsDisplay the value labels
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_name *The name of the file to read
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
null_valueThe value that represents null (a missing value) in the file, for example, NULL
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_name *The name of the file to write to or delete
first_line_headerIndicates whether the row where reading starts is the header. Default: false
include_typesInclude data types in the first line of the file. Default: false
names_as_labelsSet column labels to the value of the column name
null_valueThe value that represents null (a missing value) in the file, for example, NULL
partitionedWrite the file as multiple partitions. Default: false
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
sheet_nameThe name of the Excel worksheet to write to
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




Generic S3


Description: Access data from a storage service that is compatible with the Amazon S3 API. To connect directly to Amazon S3, use the Amazon S3 connector.
Data source type ID: 38714ac2-8f66-4a8c-9b40-806ffb61c759
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
access_key *The access key ID (username) for authorizing access to S3
bucketThe name of the bucket that contains the files to access
disable_chunked_encodingSet this property if the storage doesn't support chunked encoding. Default: false
enable_global_bucket_accessWhether global bucket access should be enabled. Default: true
enable_path_style_accessWhether path style access should be enabled
url *The endpoint URL to use for access to S3
regionS3 region
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
secret_key *The password associated with the access key ID for authorizing access to S3
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
bucketThe name of the bucket that contains the files to read
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
display_value_labelsDisplay the value labels
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_name *The name of the file to read
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
null_valueThe value that represents null (a missing value) in the file, for example, NULL
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
bucketThe name of the bucket that contains the files to write
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
create_bucketCreate the bucket that contains the files to write to. Default: false
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_name *The name of the file to write to or delete
first_line_headerIndicates whether the row where reading starts is the header. Default: false
include_typesInclude data types in the first line of the file. Default: false
names_as_labelsSet column labels to the value of the column name
null_valueThe value that represents null (a missing value) in the file, for example, NULL
partitionedWrite the file as multiple partitions. Default: false
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
sheet_nameThe name of the Excel worksheet to write to
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




Google BigQuery


Description: Google BigQuery is a fully managed, serverless data warehouse that enables scalable analysis over petabytes of data.
Data source type ID: 933152db-99e1-453a-8ce5-ae0e6714d1a9
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
access_token *An access token that can be used to connect to BigQuery
auth_method. Values: [credentials, credentials_oauth2, workload_identity_federation_token, workload_identity_federation_token_url]
client_id *The OAuth client ID
client_secret *The OAuth client secret
credentials *The contents of the Google service account key (JSON) file
token_url_headersHTTP headers for the token URL request in JSON or as a JSON body: "Key1"="Value1","Key2"="Value2"
token_url_methodHTTP method that will be used for the token URL request. Values: [get, post, put]. Default: get
project_idThe ID of the Google project
refresh_token *A refresh token to be used to refresh the access token
token_url_bodyThe body of the HTTP request to retrieve a token
sts_audience *The Security Token Service audience containing the project ID, pool ID, and provider ID
service_account_email *The e-mail address of the service account
service_account_token_lifetimeThe lifetime in seconds of the service account access token
token_url *The URL to retrieve a token
token_field_name *The name of the field in the JSON response that contains the token
token_formatThe format of the token. Values: [json, text]. Default: text
token_typeThe type of access token. Values: [aws4_request, access_token, id_token, jwt, saml2]. Default: id_token

Interaction properties (when used as a source)

Name Type Description
bucket *The name of the Google Cloud Storage bucket to be used for staging temporary files
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
schema_name *Specify the name of the BigQuery dataset that contains the table. When GCS staging area is selected the temporary staging table will be created in this schema.
database_nameSpecify the Google project id where the the table resides. This is an optional property. If this property is not specified, project id from the connection will be used for operations. When GCS staging area is selected the temporary staging table will be created in this project id.
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from
use_gcs_stagingSpecify whether you want to use Google Cloud Storage as staging area while executing the select statement, to improve performance.. Default: false

Interaction properties (when used as a target)

Name Type Description
bucketThe name of the Google Cloud Storage bucket to be used for staging temporary files
schema_name *Specify the name of the BigQuery dataset that contains the table.
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
database_nameSpecify the Google project id where the the table resides. This is an optional property. If this property is not specified, project id from the connection will be used for operations.
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace]. Default: append
table_name *The name of the table to write to
update_statement *The SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [delete, delete_insert, insert, merge, static_statement, update, update_statement]. Default: insert




Google Cloud Pub/Sub


Description: Google Cloud Pub/Sub is used for streaming analytics and data integration pipelines that ingest and distribute data.
Data source type ID: 693c2a02-39d1-4394-9426-fcdcfc4f3d7a
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
access_token *An access token that can be used to connect to Google Pub/Sub
auth_method. Values: [credentials, credentials_oauth2]
client_id *The OAuth client ID
client_secret *The OAuth client secret
credentials *The contents of the Google service account key (JSON) file
project_idThe ID of the Google project
refresh_token *A refresh token to be used to refresh the access token

Interaction properties (when used as a source)

Name Type Description
row_limitThe maximum number of rows to return
subscription_idThe id of the Google Pub/Sub subscription
timeout_after_last_messageNumber of seconds after last received message.. Default: 300

Interaction properties (when used as a target)

Name Type Description
topic_idThe id of the Google Pub/Sub topic




Google Cloud Storage


Description: Google Cloud Storage is an online file storage web service for storing and accessing data on Google Cloud Platform Infrastructure.
Data source type ID: 05b7f0ea-6ae4-45e2-a455-cc280f110825
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
access_token *An access token that can be used to connect to Google Cloud Storage
auth_method. Values: [credentials, credentials_oauth2]
client_id *The OAuth client ID
client_secret *The OAuth client secret
credentials *The contents of the Google service account key (JSON) file
project_idThe id of the Google project
refresh_token *A refresh token to be used to refresh the access token

Interaction properties (when used as a source)

Name Type Description
bucketThe name of the bucket that contains the files to read
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
display_value_labelsDisplay the value labels
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_name *The name of the file to read
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
null_valueThe value that represents null (a missing value) in the file, for example, NULL
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
bucketThe name of the bucket that contains the files to write
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
create_bucketCreate the bucket that contains the files to write to. Default: false
database_nameThe name of the database that contains the table to write to
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_name *The name of the file to write to or delete
first_line_headerIndicates whether the row where reading starts is the header. Default: false
include_typesInclude data types in the first line of the file. Default: false
create_bigquery_tableWhether to load data into BigQuery table from the GCS file. Default: false
locationA region, dual-region or multi-region location where your data will be stored
names_as_labelsSet column labels to the value of the column name
null_valueThe value that represents null (a missing value) in the file, for example, NULL
partitionedWrite the file as multiple partitions. Default: false
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
schema_name *The name of the schema that contains the table to write to
storage_classThe storage class for the created bucket. Values: [coldline, multi_regional, nearline, regional, standard]. Default: standard
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace]. Default: append
table_name *The name of the table to write to
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
sheet_nameThe name of the Excel worksheet to write to
write_modeWhether to write to, or delete, the target. Values: [delete, delete_multiple_prefix, write, write_raw]. Default: write




Greenplum


Description: Greenplum is a massively parallel processing (MPP) database server that supports next generation data warehousing and large-scale analytics processing.
Data source type ID: e278eff1-a7c4-4d60-9a02-bde1bb1d26ef
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
prepare_statement_supportEnable if connector should allow prepare statements.. Default: false
query_timeoutSets the default query timeout in seconds for all statements created by a connection. If not specified the default value of 300 seconds will be used.. Default: 300
retry_limitSpecify the maximum number of retry connection attempts to be made by the connector with an increasing delay between each retry. If no value is provided, two attempts will be made by default if necessary.
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




HTTP


Description: To access your data from a URL, create an HTTP connection asset for it.
Data source type ID: 4210c294-8b0f-46b4-bcdc-1c6ada2b7e6b
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
url *The URL of the file to be accessed
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
display_value_labelsDisplay the value labels
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_nameThe name of the file to read
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
null_valueThe value that represents null (a missing value) in the file, for example, NULL
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
read_modeThe method for reading files. Values: [read_single, read_raw]. Default: read_single
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description




IBM Cloud Data Engine


Description: IBM Data Engine is a service on IBM Cloud that you use to build, manage, and consume data lakes and their table assets in IBM Cloud Object Storage. IBM Cloud Data Engine was formerly named IBM Cloud SQL Query.
Data source type ID: 6bcaf300-30b3-11eb-adc1-0242ac120002
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
crn *To find the CRN, go to IBM Cloud Data Engine service. Copy the value of CRN from deployment details
password *The IAM API key for accessing the data source
target_cos_url *Target Cloud Object Storage, where the results should be stored

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [none, random, row]. Default: none
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description




IBM Cloud Databases for DataStax


Description: IBM Cloud Databases for DataStax is a scale-out NoSQL database in IBM Cloud that is built on Apache Cassandra.
Data source type ID: 0bd5946b-6fcb-4253-bf76-48b362d24a89
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
keyspace *The name of the keyspace
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: true
private_keyThe private key
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
key_certificateThe certificate that will be stored along with the private key
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Cloud Databases for MongoDB


Description: IBM Cloud Databases for MongoDB is a MongoDB database that is managed by IBM Cloud. It uses a JSON document store with a rich query and aggregation framework. For the MongoDB connector that is not on IBM Cloud, select the MongoDB connector.
Data source type ID: 8e65204d-6156-49e7-96e5-d635b2aa05f6
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
auth_databaseThe name of the database in which the user was created
column_discovery_sample_sizeThe number of rows sampled per collection to determine table schemas. The default is 1000.
database *The name of the database
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: true
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
schema_filterA comma-separated list of database:collection pairs for which the driver should fetch metadata. For more information look into DataDirect driver documentation.
special_char_behaviorSpecifies whether special characters in names that do not conform to SQL identifier syntax should be stripped (the default), included, or replaced with underscores. Values: [include, replace, strip]
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Cloud Databases for MySQL


Description: IBM Cloud Databases for MySQL extends the capabilities of MySQL by offering an easy, auto-scaling deployment system managed on IBM Cloud that delivers high availability and redundancy, and automated backups.
Data source type ID: 0cd4b64c-b485-47ed-a8c4-329c25412de3
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [none, random]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Cloud Databases for PostgreSQL


Description: IBM Cloud Databases for PostgreSQL is an open-source object-relational database that is highly customizable. It's a feature-rich enterprise database with JSON support. For the PostgreSQL connector that is not on IBM Cloud, select the PostgreSQL connector.
Data source type ID: 048ed1bf-516c-46f0-ae90-fa3349d8bc1c
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
query_timeoutSets the default query timeout in seconds for all statements created by a connection. If not specified the default value of 300 seconds will be used.. Default: 300
retry_limitSpecify the maximum number of retry connection attempts to be made by the connector with an increasing delay between each retry. If no value is provided, two attempts will be made by default if necessary.
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Cloud Object Storage


Description: IBM Cloud Object Storage on IBM Cloud provides unstructured data storage for cloud applications. Cloud Object Storage offers S3 API and application binding with regional and cross-regional resiliency.
Data source type ID: 193a97c1-4475-4a19-b90c-295c4fdc6517
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
api_key *To find the API key, go to https://cloud.ibm.com/resources, expand the Storage resource, click the Cloud Object Storage service, and then click Service credentials in the left pane. Expand the desired Key name. Copy the value of apikey without the quotation marks
access_keyTo find the Access key, go to https://cloud.ibm.com/resources, expand the Storage resource, click the Cloud Object Storage service, and then click Service credentials in the left pane. Expand the desired Key name. Copy the value of access_key_id without the quotation marks
auth_method. Values: [accesskey_secretkey, instanceid_apikey, instanceid_apikey_accesskey_secretkey, credentials]
bucketThe name of the bucket that contains the files to access
url *To find this URL, go to https://cloud.ibm.com/resources, expand the Storage resource, click the Cloud Object Storage service, and then click Endpoint in the left pane. Copy the value of the public endpoint that you want to use
resource_instance_idTo find the Resource instance ID, go to https://cloud.ibm.com/resources, expand the Storage resource, click the Cloud Object Storage service, and then click Service credentials in the left pane. Expand the desired Key name. Copy the value of resource_instance_id without the quotation marks
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
secret_key *To find the Secret key, go to https://cloud.ibm.com/resources, expand the Storage resource, click the Cloud Object Storage service, and then click Service credentials in the left pane. Expand the desired Key name. Copy the value of secret_access_key without the quotation marks
credentialsThe contents of the Cloud Object Storage service credentials (JSON) file. Find JSON content by going to "Service credentials" tab and expanding selected credentials. Copy whole content including {} brackets.

Interaction properties (when used as a source)

Name Type Description
bucketThe name of the bucket that contains the files to read
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
display_value_labelsDisplay the value labels
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_nameThe name of the file to read
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
null_valueThe value that represents null (a missing value) in the file, for example, NULL
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
append_uidUse this property to choose if a unique identifier is to be appended to the file name. When the value of this property is set to yes, then the file name gets appended with the unique identifier, and a new file would be written for every wave of data that is streamed into the stage. When the value of this property is set to No, then the file would be overwritten on every wave.. Default: false
bucketThe name of the bucket that contains the files to write
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
create_bucketCreate the bucket that contains the files to write to. Default: false
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_nameThe name of the file to write to or delete
file_size_thresholdSpecify the threshold for the file size in megabytes. Processing nodes will start a new file each time the size exceeds the value specified in the threshold.. Default: 1
first_line_headerIndicates whether the row where reading starts is the header. Default: false
include_typesInclude data types in the first line of the file. Default: false
names_as_labelsSet column labels to the value of the column name
null_valueThe value that represents null (a missing value) in the file, for example, NULL
partitionedWrite the file as multiple partitions. Default: false
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
storage_classThe storage class for the created bucket. Values: [cold_vault, flex, standard, vault]. Default: standard
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
sheet_nameThe name of the Excel worksheet to write to
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




IBM Cloud Object Storage (infrastructure)


Description: The IBM Cloud Object Storage (infrastructure) connector is for object storage that was formerly on SoftLayer. SoftLayer was replaced by IBM Cloud. You cannot provision a new instance for Cloud Object Storage (infrastructure). This connector is for users who set up an earlier instance on SoftLayer.
Data source type ID: 4bf2dedd-3809-4443-96ec-b7bc5726c07b
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
access_keyTo find the Access key, go to https://cloud.ibm.com/resources, expand the Storage resource, click the Cloud Object Storage service, and then click Service credentials in the left pane. Expand the desired Key name. Copy the value of access_key_id without the quotation marks
auth_method. Values: [accesskey_secretkey, credentials]
bucketThe name of the bucket that contains the files to access
url *To find this URL, go to https://cloud.ibm.com/resources, expand the Storage resource, click the Cloud Object Storage service, and then click Endpoint in the left pane. Copy the value of the public endpoint that you want to use
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
secret_key *To find the Secret key, go to https://cloud.ibm.com/resources, expand the Storage resource, click the Cloud Object Storage service, and then click Service credentials in the left pane. Expand the desired Key name. Copy the value of secret_access_key without the quotation marks
credentialsThe contents of the Cloud Object Storage service credentials (JSON) file. Find JSON content by going to "Service credentials" tab and expanding selected credentials. Copy whole content including {} brackets.

Interaction properties (when used as a source)

Name Type Description
bucketThe name of the bucket that contains the files to read
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
display_value_labelsDisplay the value labels
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_nameThe name of the file to read
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
null_valueThe value that represents null (a missing value) in the file, for example, NULL
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
bucketThe name of the bucket that contains the files to write
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
create_bucketCreate the bucket that contains the files to write to. Default: false
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_nameThe name of the file to write to or delete
first_line_headerIndicates whether the row where reading starts is the header. Default: false
include_typesInclude data types in the first line of the file. Default: false
names_as_labelsSet column labels to the value of the column name
null_valueThe value that represents null (a missing value) in the file, for example, NULL
partitionedWrite the file as multiple partitions. Default: false
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
storage_classThe storage class for the created bucket. Values: [cold_vault, flex, standard, vault]. Default: standard
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
sheet_nameThe name of the Excel worksheet to write to
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




IBM Cloudant


Description: IBM Cloudant is a JSON document database available in IBM Cloud.
Data source type ID: 44e904b5-0cb2-4d8e-a5c0-c48bc3e24fdd
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
custom_urlThe URL to the Cloudant database
databaseThe database to connect to
password *The password associated with the username for accessing the data source
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
database *The database to connect to
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_limitThe maximum number of rows to return

Interaction properties (when used as a target)

Name Type Description
blob_truncation_sizeThe maximum size for BLOB values. Values larger than this will be truncated. Default: 8000
batch_sizeThe number of documents to send per request. Default: 100
clob_truncation_sizeThe maximum size for CLOB values. Values larger than this will be truncated. Default: 8000
create_databaseCreate the database to connect to. Default: false
database *The database to connect to
document_typeThe type of the document
input_formatThe format of the source data. Values: [json, relational]. Default: relational
write_modeWhether to write to, or delete, the target. Values: [delete, write]. Default: write




IBM Cognos Analytics


Description: IBM Cognos Analytics is an AI-fueled business intelligence platform that supports the entire analytics cycle.
Data source type ID: 11f3029d-a1cf-4c4d-b8e7-64422fa54a94
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
auth_method. Values: [anonymous, username_password_namespace]
url *The gateway URL to access Cognos
namespace_id *The identifier of the authentication namespace
password *The password associated with the username for accessing the data source
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
file_name *The name of the file to read
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data

Interaction properties (when used as a target)

Name Type Description




IBM Data Virtualization Manager for z/OS


Description: IBM Data Virtualization Manager for z/OS provides virtual, integrated views of data residing on IBM Z.
Data source type ID: 39a78d59-ef34-4108-8e46-4460433a3b99
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Db2


Description: IBM Db2 is a database that contains relational data.
Data source type ID: 8c1a4480-1c29-4b33-9086-9cb799d7b157
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
username_password_encryptionThe encryption algorithm for username and password credentials. Values: [aes_256_bit, des_56_bit, default]. Default: default
username_password_securityThe DRDA security mechanism for username and password credentials. Values: [clear_text, default, encrypted_password, encrypted_username, encrypted_username_password, encrypted_username_password_data]. Default: default

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Db2 (optimized)


Description: Connect to a Db2 database for the DataStage Db2 (optimized) connector. The IBM Db2 (optimized) connector has more features for DataStage.
Data source type ID: fa31fba9-10e9-32d7-968c-f677fffd1e3b
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
advanced.api_keyAPI key to connect to the database.
authentication_typeSelect type of credentials. Values: [api_key, username_and_password]. Default: username_and_password
database *Specifies the name of the database to connect to.
advanced.hostname *A database hostname to connect to.
keep_conductor_connection_aliveSelect to keep the connector conductor process connected during the job run. If you do not select this property, the connector conductor process will disconnect from the database while the player processes run, and then reconnect when the player processes complete.. Default: false
advanced.optionsAdditional connection options passed as parameters to ODBC connection string.
password *Specifies the password to use for connecting to the database.
advanced.port *Port that database process is listening on.. Default: 50000
advanced.ssl_certificateOptional database SSL certificate (arm format) for establishing secure connection.
advanced.ssl_connectionUse of SSL connection.. Default: false
username *Specifies the user name to use for connecting to the database.

Interaction properties (when used as a source)

Name Type Description
before_after.afterEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once after all data is processed.
before_after.after_nodeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once one each node after all data is processed on that node.
session.array_sizeThe array size to be used for all read and write database operations. Default: 2000
session.autocommit_modeSpecifies whether the connector commits transactions manually, or allows the database to commit transactions automatically at its discretion.. Values: [off, on]. Default: off
before_after.beforeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once before any data is processed.
before_after.before_nodeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once on each node before any data is processed on that node.
before_afterBefore/After SQL properties. Default: false
sql.enable_partitioning.partitioning_method.key_field *Specifies the key column that is used by the selected partitioned reads method. This column must be a numeric data type.
session.pass_lob_locator.column *Use to choose columns containing LOBs to be passed by locator (reference)
session.use_external_tables.directory_for_named_pipeSpecifies the location where the named pipe used by the load operation should be created. This property applies to Unix systems only.. Default: /tmp
session.pass_lob_locatorEnables/disables the ability to specify LOB columns to be passed using locator (reference) information. LOB columns not specified will be passed inline. Default: false
sql.enable_partitioningEnable or disable partitioned reads by using the selected partitioning method.. Default: false
enable_quoted_i_dsSpecifies whether or not to enclose database object names in quotes when generating DDL and DML. Default: true
transaction.end_of_wave.end_of_dataSpecifies whether to insert an EOW marker for the last set of records when the number is less than the specified transaction record count value. Default: false
transaction.end_of_waveSpecify settings for the end of wave handling. None means EOW markers are never inserted, Before means EOW markers are inserted before committing the transaction, After means EOW markers are inserted after committing the transaction. Values: [after, before, none]. Default: none
before_after.after.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.after_node.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.before.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.before_node.fail_on_errorAbort the job if there is an error executing a command. Default: true
session.schema_reconciliation.fail_on_size_mismatchFail if the sizes of numeric and string fields are not compatible when validating the design schema against the database. Default: true
session.schema_reconciliation.fail_on_type_mismatchFail if the types of fields are not compatible when validating the design schema against the database. Default: true
generate_sqlSpecifies whether to generate SQL statements at run time.. Default: false
sql.enable_partitioning.partitioning_method.gen_partitioning_sqlSpecifies whether the connector should modify the SELECT statement at run-time and generate the required partitioning clause. Default: true
session.isolation_levelSpecifies the isolation level that is used for all database transactions.. Values: [cursor_stability, read_stability, read_uncommitted, repeatable_read]. Default: cursor_stability
limit_rows.limitEnter the maximum number of rows that will be returned by the connector.. Default: 1000
limit_rowsSelect Yes to limit the number of rows that are returned by the connector.. Default: false
lock_wait_modeSpecifies the lock wait strategy that is used when a lock cannot be obtained immediately.. Values: [return_an_sqlcode_and_sqlstate, use_the_lock_timeout_database_configuration_parameter, user_specified, wait_indefinitely]. Default: use_the_lock_timeout_database_configuration_parameter
lock_wait_mode.lock_wait_mode_time *Time to wait for a lock
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
sql.other_clauseThe other clause predicate of the SQL statement
session.use_external_tables.other_optionsAdditional options to be passed to the external table statement
pad_characterSpecifies the pad character that is used in the WHERE clause for string columns that are smaller than the column size.
sql.enable_partitioning.partitioning_methodThe method to use for partitioned reads.. Values: [db2_connector, minimum_and_maximum_range, modulus]. Default: minimum_and_maximum_range
prefix_for_expression_columnsSpecifies the prefix for columns that contain the result of expressions.. Default: EXPR
before_after.after_node.read_from_file_after_sql_nodeSelect Yes to read the SQL statement from the file that is specified in the After SQL (node) statement property.. Default: false
before_after.after.read_from_file_after_sqlSelect Yes to read the SQL statement from the file that is specified in the After SQL statement property.. Default: false
before_after.before_node.read_from_file_before_sql_nodeSelect Yes to read the SQL statement from the file that is specified in the Before SQL (node) statement property.. Default: false
before_after.before.read_from_file_before_sqlSelect Yes to read the SQL statement from the file that is specified in the Before SQL statement property.. Default: false
sql.select_statement.read_from_file_selectSelect YES to read the SELECT statement from the file that is specified in the SELECT statement property.. Default: false
transaction.record_countNumber of records per transaction. The value 0 means all available records. Default: 2000
re_optimizationSpecifies the type of reoptimization that is done by Db2.. Values: [always, none, once]. Default: none
sql.select_statement *Statement to be executed when reading rows from the database or the fully qualified name of the file that contains the statement.
sql.enable_partitioning.partitioning_method.table_name *Specifies the table that is used by the selected partitioned reads method.
table_name *The table name to be used in generated SQL. The table name must be schema qualified in order to preview data.
session.use_external_tablesIndicates whether external tables are used.. Default: false
sql.where_clauseThe where clause predicate of the SQL statement

Interaction properties (when used as a target)

Name Type Description
load_to_zos.data_file_attributes.discard_data_set.file_disposition.abnormal_terminationThis option specifies the abnormal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.data_file_attributes.error_data_set.file_disposition.abnormal_terminationThis option specifies the abnormal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.data_file_attributes.input_data_files.file_disposition.abnormal_terminationThis option specifies the abnormal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: keep
load_to_zos.data_file_attributes.map_data_set.file_disposition.abnormal_terminationThis option specifies the abnormal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.data_file_attributes.work1_data_set.file_disposition.abnormal_terminationThis option specifies the abnormal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: delete
load_to_zos.data_file_attributes.work2_data_set.file_disposition.abnormal_terminationThis option specifies the abnormal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: delete
load_to_zos.image_copy_function.image_copy_backup_file.file_disposition.abnormal_terminationThis option specifies the abnormal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.image_copy_function.image_copy_file.file_disposition.abnormal_terminationThis option specifies the abnormal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.image_copy_function.recovery_backup.file_disposition.abnormal_terminationThis option specifies the abnormal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.image_copy_function.recovery_file.file_disposition.abnormal_terminationThis option specifies the abnormal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
before_after.afterEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once after all data is processed.
before_after.after_nodeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once one each node after all data is processed on that node.
load_control.allow_access_modeSpecifies the level of access on the table that is to be loaded.. Values: [no_access, read]. Default: no_access
load_to_zos.image_copy_function.allow_changesIndicates whether other programs can update the table space while COPY is running. Default: false
session.array_sizeThe array size to be used for all read and write database operations. Default: 2000
session.insert_buffering.atomic_arraysSpecifies whether arrays should be inserted atomically. Insert buffering with non-atomic arrays does not report errors accurately.. Values: [auto, no, simulated, yes]. Default: auto
session.autocommit_modeSpecifies whether the connector commits transactions manually, or allows the database to commit transactions automatically at its discretion.. Values: [off, on]. Default: off
load_to_zos.batch_pipe_system_id *If the data is to be transferred to a Batch pipes system on z/OS, this option identifies its name.
before_after.beforeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once before any data is processed.
before_after.before_nodeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once on each node before any data is processed on that node.
before_afterBefore/After SQL properties. Default: false
table_action.generate_create_statement.create_table_bufferpoolSpecifies the buffer pool be use for the implicitly created table space and determines the page size of the table space. Do not specify BUFFERPOOL if the table space name is specified by using the IN table-space-name clause or the IN ACCELERATOR clause is specified. If you do not specify the BUFFERPOOL clause, Db2 selects the buffer pool as described in Implicitly defined table spaces (Db2 for z/OS only).
load_to_zosDetermines whether the target table is on Db2 for z/OS.. Default: false
load_control.bulkload_with_lob_xmlIndicates whether there are any LOB or XML columns in the target Db2 table.. Default: false
load_to_zos.encoding.ccsid *Specifies the coded character set identifier (CCSID) for SBCS data, Mixed data, and DBCS data in the input file.
load_control.cpu_parallelismSpecifies the number of processes or threads that the load utility spawns for processing records.. Default: 0
load_to_zos.image_copy_function.change_limit_percent1Specifies the percentage limit of changed pages in the table space at which an incremental image-copy is to be taken
load_to_zos.image_copy_function.change_limit_percent2Specifies the percentage limit of changed pages in the table space at which an incremental image-copy is to be taken
load_to_zos.encoding.character_setThe IANA character set name for the encoding. If not specified, ibm-1047-s390 will be used for EBCDIC, ASCII for ASCII, and UTF-16BE for UNICODE.
sql.user_defined_sql.file.character_setIANA character set name
load_control.check_pending_cascadeSpecifies whether the check pending state of the loaded table is immediately cascaded to all descendants.. Values: [deferred, immediate]. Default: deferred
load_control.partitioned_db_config.check_truncationIf selected, data records are checked for truncation at input and output (CHECK_TRUNCATION).. Default: false
load_control.cleanup_on_failClean-up on failures during stage execution.. Default: false
logging.log_column_values.delimiterSpecifies the delimiter to use between columns. Values: [comma, newline, space, tab]. Default: space
table_action.generate_create_statement.create_table_compressSpecifies whether data compression applies to the rows. On Db2 for z/OS: if the IN table-space-name clause or the IN ACCELERATOR clause is specified, COMPRESS YES or COMPRESS NO must not be specified.. Values: [database_default, no, yes]. Default: database_default
load_to_zos.shr_levelLevel of application's concurrent access to the table space or partition. Parameter's value corresponds to SHRLEVEL option of LOAD command.. Values: [change, none, reference]. Default: none
load_control.copy_loaded_dataSpecifies the method that is used for making a copy of the loaded data.. Values: [no_copy, use_tivoli_to_make_a_copy, use_device_or_directory, use_shared_library]. Default: no_copy
table_action.generate_create_statement.create_statement *A statement to be executed when creating the target database table
load_to_zos.dsn_prefixThis option identifies a prefix to be used when creating MVS dataset names. If omitted, the transfer user name is used.
load_control.data_buffer_sizeSpecifies the number of pages (size 4KB) that are used as buffered space for transferring data within the load utility.. Default: 0
load_to_zos.data_file_attributes.discard_data_set.data_classSpecifies the SMS data class (DATACLAS). The name must be a valid SMS data class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.error_data_set.data_classSpecifies the SMS data class (DATACLAS). The name must be a valid SMS data class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.input_data_files.data_classSpecifies the SMS data class (DATACLAS). The name must be a valid SMS data class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.map_data_set.data_classSpecifies the SMS data class (DATACLAS). The name must be a valid SMS data class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work1_data_set.data_classSpecifies the SMS data class (DATACLAS). The name must be a valid SMS data class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work2_data_set.data_classSpecifies the SMS data class (DATACLAS). The name must be a valid SMS data class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_backup_file.data_classSpecifies the SMS data class (DATACLAS). The name must be a valid SMS data class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_file.data_classSpecifies the SMS data class (DATACLAS). The name must be a valid SMS data class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_backup.data_classSpecifies the SMS data class (DATACLAS). The name must be a valid SMS data class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_file.data_classSpecifies the SMS data class (DATACLAS). The name must be a valid SMS data class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.discard_data_set.dataset_nameDataset name used by the Load utility. If empty (default), then a dataset name is generated using the DSN prefix property.
load_to_zos.data_file_attributes.error_data_set.dataset_nameDataset name used by the Load utility. If empty (default), then a dataset name is generated using the DSN prefix property.
load_to_zos.data_file_attributes.input_data_files.dataset_nameDataset name used by the Load utility. If empty (default), then a dataset name is generated using the DSN prefix property.
load_to_zos.data_file_attributes.map_data_set.dataset_nameDataset name used by the Load utility. If empty (default), then a dataset name is generated using the DSN prefix property.
load_to_zos.data_file_attributes.work1_data_set.dataset_nameDataset name used by the Load utility. If empty (default), then a dataset name is generated using the DSN prefix property.
load_to_zos.data_file_attributes.work2_data_set.dataset_nameDataset name used by the Load utility. If empty (default), then a dataset name is generated using the DSN prefix property.
load_to_zos.image_copy_function.image_copy_backup_file.dataset_nameDataset name used by the Load utility. If empty (default), then a dataset name is generated using the DSN prefix property.
load_to_zos.image_copy_function.image_copy_file.dataset_nameDataset name used by the Load utility. If empty (default), then a dataset name is generated using the DSN prefix property.
load_to_zos.image_copy_function.recovery_backup.dataset_nameDataset name used by the Load utility. If empty (default), then a dataset name is generated using the DSN prefix property.
load_to_zos.image_copy_function.recovery_file.dataset_nameDataset name used by the Load utility. If empty (default), then a dataset name is generated using the DSN prefix property.
sql.delete_statement *Statement to be executed when deleting rows from the database
load_to_zos.device_typeThis option specifies the device type to be used for various datasets that the LOAD utility may need. If omitted, the default is SYSDA.. Default: SYSDA
sql.direct_insertIf set to Yes, the connector insert directly into the target table. In this mode, when running with multiple processing nodes it is possible to have partially committed data if one or more of the processing nodes encounters an error. If set No, the connector inserts into the temporary work table (TWT) first and then from TWT into the target. In this mode the data will either be completely committed or completely rolled back guarantying consistency.. Default: true
load_control.data_file_path *Specifies the location where the command file and data file (used by the load operation) should be created.
load_to_zos.transfer.data_file_pathSpecifies the location where the data files will be created before transfer to z/OS.
session.use_external_tables.log_directorySpecifies the directory for the log and bad files. If it is left blank, the connector will use the value of the environment variable TMPDIR. If TMPDIR is not defined, it will default to /tmp on Unix and to system temporary directory on Windows.
load_control.directory_for_named_pipeSpecifies the location where the named pipe used by the load operation should be created. This property applies to Unix systems only.. Default: /tmp
session.use_external_tables.directory_for_named_pipeSpecifies the location where the named pipe used by the load operation should be created. This property applies to Unix systems only.. Default: /tmp
load_control.disk_parallelismSpecifies the number of processes or threads that the load utility spawns for writing data.. Default: 0
table_action.generate_create_statement.create_table_distribute_bySpecifies the database partitioning or the way the data is distributed across multiple database partitions (Db2 LUW only).. Values: [hash, none, random]. Default: none
table_action.generate_drop_statement.drop_statement *A statement to be executed when dropping the database table
session.temporary_work_table.drop_tableIf set to Yes, the connector will drop the temporary work table.. Default: true
session.schema_reconciliation.drop_unmatched_fieldsDrop fields that don't exist in the input schema. Default: true
load_control.file_type_modifiers.dump_fileSpecifies a fully qualified file path to use with the 'dumpfile' file type modifier. If no path is specified, the 'dumpfile' modifier is not used
enable_quoted_i_dsSpecifies whether or not to enclose database object names in quotes when generating DDL and DML. Default: false
load_to_zos.encodingSpecifies the input dataset encoding. Values: [ascii, ccsid, ebcdic, unicode]. Default: ebcdic
load_control.exception_tableName of tables where rows that violate constraints will be stored.
sql.user_defined_sql.fail_on_errorAbort the statement sequence when an error occurs. Default: true
before_after.after.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.after_node.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.before.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.before_node.fail_on_errorAbort the job if there is an error executing a command. Default: true
table_action.generate_create_statement.fail_on_errorAbort the job if there is an error executing a command. Default: true
table_action.generate_drop_statement.fail_on_errorAbort the job if there is an error executing a command. Default: true
table_action.generate_truncate_statement.fail_on_errorAbort the job if there is an error executing a command. Default: true
session.fail_on_row_error_pxFail the job if a write operation to the target is unsuccessful. Default: true
session.schema_reconciliation.fail_on_size_mismatchFail if the sizes of numeric and string fields are not compatible when validating the design schema against the database. Default: true
session.schema_reconciliation.fail_on_type_mismatchFail if the types of fields are not compatible when validating the design schema against the database. Default: true
sql.user_defined_sql.file *File on the conductor node that contains SQL statements to be executed for each input row
load_control.file_typeSpecifies the format of the data in the data file.. Values: [asc, del]. Default: asc
load_control.files_onlyIndicates whether input files should be created without executing the load operation.. Default: false
load_to_zos.files_onlyLOAD will not actually be executed, only MVS datasets are created.. Default: false
generate_sqlSpecifies whether to generate SQL statements at run time.. Default: false
table_action.generate_create_statementSpecifies whether to generate a create table statement at runtime. Default: true
table_action.generate_drop_statementSpecifies whether to generate a drop table statement at runtime. Default: true
table_action.generate_truncate_statementSpecifies whether to generate a truncate table statement at runtime. Default: true
load_to_zos.encoding.graphic_character_setThe IANA character set name for the graphic encoding. If not specified, UTF-16BE is the default.
load_to_zos.transfer.uss_file_directory *The presence of this option indicates that HFS files is to be used and gives the directory name where the files will be created. The value should be a fully qualified HFS directory name.
load_control.partitioned_db_config.port_range.max_value *Higher port number of the port range
load_to_zos.image_copy_function.image_copy_backup_fileSpecifies whether or not to create a backup of the image-copy file. Default: false
load_to_zos.image_copy_functionSpecifies whether to run an Image-copy function after Load. Values: [concurrent, full, incremental, no]. Default: no
table_action.generate_create_statement.create_table_inIdentifies the database and/or table space in which the table is created: IN database-name.table-space-name. On Db2 LUW if required add CYCLE / NO CYCLE indication.
table_action.generate_create_statement.create_table_index_inSpecifies the table space in which indexes or long column values are to be stored (Db2 LUW only).
load_control.indexing_modeSpecifies whether indexes are rebuilt or extended incrementally.. Values: [automatic_selection, do_not_update_table_indexes, extend_existing_indexes, rebuild_table_indexes]. Default: automatic_selection
session.insert_bufferingSpecifies whether to enable the insert buffering optimization in partitioned database environments.. Values: [default, ignore_duplicates, off, on]. Default: default
sql.insert_statement *Statement to be executed when inserting rows into the database
load_to_zos.transfer.retry_connection.retry_interval *Enter the time in seconds to wait between retries to establish a connection.. Default: 10
load_control.partitioned_db_config.isolate_part_errorsSpecifies the reaction of the load operation to errors that occur on individual partitions (ISOLATE_PART_ERRS).. Values: [load_errors_only, no_isolation, setup_and_load_errors, setup_errors_only]. Default: load_errors_only
session.isolation_level *Specifies the isolation level that is used for all database transactions.. Values: [cursor_stability, read_stability, read_uncommitted, repeatable_read]. Default: cursor_stability
load_to_zos.resumeSelect Yes to add records to the end of the table if the table space is not empty or No to load data into an empty table space. The value corresponds to the RESUME option value in the LOAD command.. Default: true
sql.key_columnsA comma-separated list of key column names.
table_action.generate_create_statement.create_table_distribute_by.hash_key_columns *A comma-separated list of key column names.
load_control.copy_loaded_data.copy_load_library_name *Specifies the name of the library that is used to generate the copy.
limit_parallelismBy default the connector runs one player process per database partition. If you want to force the connector to run fewer player processes, set this property to Yes. Default: false
load_control.load_methodDetermines the load method to use.. Values: [named_pipes, sequential_files]. Default: named_pipes
load_to_zos.load_methodLoad method to be used for loading input data into Db2 for z/OS.. Values: [batch_pipes, mvs_datasets, uss_pipes]. Default: mvs_datasets
load_control.load_modeSpecifies the mode where the load operates.. Values: [insert, replace, restart, terminate]. Default: insert
load_control.load_timeoutSpecifies the time in seconds to attempt opening a socket for the load operation before timing out.. Default: 300
load_to_zos.load_with_loggingIndicates whether logging is to occur during the load process.. Default: false
load_control.copy_loaded_data.copy_to_device_or_directory *A comma-separated list of devices or directories where the copy is generated.
load_control.lob_path_listA list of fully qualified paths or devices to identify the location of the individual LOB files to be loaded.
lock_wait_modeSpecifies the lock wait strategy that is used when a lock cannot be obtained immediately.. Values: [return_an_sqlcode_and_sqlstate, use_the_lock_timeout_database_configuration_parameter, user_specified, wait_indefinitely]. Default: use_the_lock_timeout_database_configuration_parameter
lock_wait_mode.lock_wait_mode_time *Time to wait for a lock
load_control.lock_with_forceIf selected, the load operation forces off other applications that hold conflicting locks.. Default: false
logging.log_column_valuesSpecifies whether to log column values for the first row that fails to be written. Default: false
logging.log_column_values.log_keys_onlySpecifies whether to log key columns or all columns for failing statements. Default: false
load_control.partitioned_db_config.port_range.min_value *Lower port number of the port range
load_to_zos.data_file_attributes.discard_data_set.management_classSpecifies the SMS management class (MGMTCLAS). The name must be a valid SMS management class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.error_data_set.management_classSpecifies the SMS management class (MGMTCLAS). The name must be a valid SMS management class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.input_data_files.management_classSpecifies the SMS management class (MGMTCLAS). The name must be a valid SMS management class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.map_data_set.management_classSpecifies the SMS management class (MGMTCLAS). The name must be a valid SMS management class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work1_data_set.management_classSpecifies the SMS management class (MGMTCLAS). The name must be a valid SMS management class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work2_data_set.management_classSpecifies the SMS management class (MGMTCLAS). The name must be a valid SMS management class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_backup_file.management_classSpecifies the SMS management class (MGMTCLAS). The name must be a valid SMS management class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_file.management_classSpecifies the SMS management class (MGMTCLAS). The name must be a valid SMS management class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_backup.management_classSpecifies the SMS management class (MGMTCLAS). The name must be a valid SMS management class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_file.management_classSpecifies the SMS management class (MGMTCLAS). The name must be a valid SMS management class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_control.partitioned_db_config.max_num_part_agentsSpecifies the maximum number of partitioning agents to be used in a load session (MAX_NUM_PART_AGENTS).. Default: 25
session.use_external_tables.max_errorsThe number of rejected records at which the system stops processing and immediately rolls back the load. The default is 1 (that is, a single rejected record results in a rollback). Default: 1
load_control.message_file *Specifies the file where Db2 writes diagnostic messages.. Default: loadMsgs.out
load_control.allow_access_mode.table_spaceSpecifies the table space that is used for building a shadow copy of the index if the indexes are being rebuilt.
load_control.non_recoverable_txThe load transaction is marked as non-recoverable.. Default: false
load_to_zos.data_file_attributes.discard_data_set.file_disposition.normal_terminationThis option specifies the normal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.data_file_attributes.error_data_set.file_disposition.normal_terminationThis option specifies the normal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.data_file_attributes.input_data_files.file_disposition.normal_terminationThis option specifies the normal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: keep
load_to_zos.data_file_attributes.map_data_set.file_disposition.normal_terminationThis option specifies the normal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.data_file_attributes.work1_data_set.file_disposition.normal_terminationThis option specifies the normal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: delete
load_to_zos.data_file_attributes.work2_data_set.file_disposition.normal_terminationThis option specifies the normal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: delete
load_to_zos.image_copy_function.image_copy_backup_file.file_disposition.normal_terminationThis option specifies the normal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.image_copy_function.image_copy_file.file_disposition.normal_terminationThis option specifies the normal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.image_copy_function.recovery_backup.file_disposition.normal_terminationThis option specifies the normal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.image_copy_function.recovery_file.file_disposition.normal_terminationThis option specifies the normal termination disposition of the dataset(s) used by the LOAD utility: Keep (KEEP), Delete (DELETE), Catalog (CATLG), Uncatalog (UNCATLG).. Values: [catalog, delete, keep, uncatalog]. Default: catalog
load_to_zos.data_file_attributes.discard_data_set.number_of_buffersSpecifies the number of buffers to use (BUFNO). If empty (default), then the property is not used.
load_to_zos.data_file_attributes.error_data_set.number_of_buffersSpecifies the number of buffers to use (BUFNO). If empty (default), then the property is not used.
load_to_zos.data_file_attributes.input_data_files.number_of_buffersSpecifies the number of buffers to use (BUFNO). If empty (default), then the property is not used.
load_to_zos.data_file_attributes.map_data_set.number_of_buffersSpecifies the number of buffers to use (BUFNO). If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work1_data_set.number_of_buffersSpecifies the number of buffers to use (BUFNO). If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work2_data_set.number_of_buffersSpecifies the number of buffers to use (BUFNO). If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_backup_file.number_of_buffersSpecifies the number of buffers to use (BUFNO). If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_file.number_of_buffersSpecifies the number of buffers to use (BUFNO). If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_backup.number_of_buffersSpecifies the number of buffers to use (BUFNO). If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_file.number_of_buffersSpecifies the number of buffers to use (BUFNO). If empty (default), then the property is not used.
load_to_zos.transfer.retry_connection.retry_count *Enter the number of times to try to establish a connection after a failure on the initial attempt.. Default: 3
table_action.generate_create_statement.create_table_organize_bySpecifies how the data is organized in the data pages of the table: row-organized table / column-organized table (Db2 LUW only).. Values: [column, database_default, row]. Default: database_default
session.use_external_tables.other_optionsAdditional options to be passed to the external table statement
table_action.generate_create_statement.create_table_other_optionsOther options/clauses to the CREATE TABLE statement eg. Partitioning-clause.
load_control.partitioned_db_config.output_db_part_numsList of database partition numbers (OUTPUT_DBPARTNUMS). The database partition numbers represent the database partitions on which the load operation is to be performed. Items in the list must be separated by commas.
pad_characterSpecifies the pad character that is used in the WHERE clause for string columns that are smaller than the column size.
load_control.partitioned_db_config.run_stat_db_partnumSpecifies the database partition where statistics are collected (RUN_STAT_DBPARTNUM).. Default: -1
load_to_zos.partition_numberIf this option is present, then only the indicated partition will be loaded. If this option is omitted, all partitions of the Db2 for z/OS database table will be loaded.
load_control.partitioned_db_configIf selected, data is loaded into a partitioned table.. Default: false
load_control.partitioned_db_config.dist_fileIf this option is specified, the load utility generates a database partition distribution file with the given name (DISTFILE).
load_control.partitioned_db_config.partitioning_db_part_numsList of database partition numbers that are used in the distribution process (PARTITIONING_DBPARTNUMS). Items in the list must be separated by commas.
load_to_zos.transfer.passwordDetermines the password for transferring.
table_action.table_action_firstSelect Yes to perform table action first. Select No to run Before SQL statements first.. Default: true
load_control.partitioned_db_config.port_rangeSpecifies the range of TCP ports that are used to create sockets for internal communications (PORT_RANGE).
prefix_for_expression_columns *Specifies the prefix for columns that contain the result of expressions.. Default: EXPR
load_to_zos.data_file_attributes.discard_data_set.primary_allocationSpecifies the z/OS disk space primary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.data_file_attributes.error_data_set.primary_allocationSpecifies the z/OS disk space primary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.data_file_attributes.input_data_files.primary_allocationSpecifies the z/OS disk space primary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.data_file_attributes.map_data_set.primary_allocationSpecifies the z/OS disk space primary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.data_file_attributes.work1_data_set.primary_allocationSpecifies the z/OS disk space primary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.data_file_attributes.work2_data_set.primary_allocationSpecifies the z/OS disk space primary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.image_copy_function.image_copy_backup_file.primary_allocationSpecifies the z/OS disk space primary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.image_copy_function.image_copy_file.primary_allocationSpecifies the z/OS disk space primary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.image_copy_function.recovery_backup.primary_allocationSpecifies the z/OS disk space primary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.image_copy_function.recovery_file.primary_allocationSpecifies the z/OS disk space primary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
before_after.after_node.read_from_file_after_sql_nodeSelect Yes to read the SQL statement from the file that is specified in the After SQL (node) statement property.. Default: false
before_after.after.read_from_file_after_sqlSelect Yes to read the SQL statement from the file that is specified in the After SQL statement property.. Default: false
before_after.before_node.read_from_file_before_sql_nodeSelect Yes to read the SQL statement from the file that is specified in the Before SQL (node) statement property.. Default: false
before_after.before.read_from_file_before_sqlSelect Yes to read the SQL statement from the file that is specified in the Before SQL statement property.. Default: false
table_action.generate_create_statement.read_create_statement_from_fileSelect YES to read the CREATE statement from the file that is specified in the CREATE statement property.. Default: false
table_action.generate_drop_statement.read_drop_statement_from_fileSelect YES to read the DROP statement from the file that is specified in the DROP statement property.. Default: false
table_action.generate_truncate_statement.read_truncate_statement_from_fileSelect YES to read the TRUNCATE statement from the file that is specified in the TRUNCATE statement property.. Default: false
transaction.record_countNumber of records per transaction. The value 0 means all available records. Default: 2000
load_to_zos.image_copy_function.recovery_backupSpecifies whether or not to create a backup of the image-copy recovery file. Default: false
load_to_zos.image_copy_function.recovery_fileSpecifies whether or not to create the image-copy recovery file. Default: false
load_control.remove_intermediate_data_fileSelect Yes to remove the intermediate data file after completing the load operation.. Default: true
re_optimizationSpecifies the type of reoptimization that is done by Db2.. Values: [always, none, once]. Default: none
load_to_zos.image_copy_function.report_onlyRun utility to produce report only. Default: false
load_control.restart_phaseSpecifies which Db2 load phase is to be restarted. The original input data needs to be reproduced for restarting the Load phase. Build and Delete phases ignore the input data.. Values: [build, delete, load]. Default: load
load_to_zos.transfer.retry_connectionSelect Yes to try to establish a connection again when the initial attempt to connect is unsuccessful.. Default: true
load_control.i_row_countThe number of physical records to be loaded. Allows a user to load only the first rowcnt rows in a file.. Default: 0
load_to_zos.row_count_estimateThis option specifies an estimated count of the total number of rows to be loaded into all partitions combined. This estimate is used to calculate the amount of disk space which will be needed on z/OS for various datasets. If not present, a row estimate of 1000 is used.. Default: 1000
load_control.save_countSpecifies the number of records to load before establishing a consistency point.. Default: 0
load_to_zos.image_copy_function.scopeScope of the image-copy. Values: [full, single_partition]. Default: full
load_to_zos.data_file_attributes.discard_data_set.secondary_allocationSpecifies the z/OS disk space secondary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.data_file_attributes.error_data_set.secondary_allocationSpecifies the z/OS disk space secondary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.data_file_attributes.input_data_files.secondary_allocationSpecifies the z/OS disk space secondary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.data_file_attributes.map_data_set.secondary_allocationSpecifies the z/OS disk space secondary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.data_file_attributes.work1_data_set.secondary_allocationSpecifies the z/OS disk space secondary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.data_file_attributes.work2_data_set.secondary_allocationSpecifies the z/OS disk space secondary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.image_copy_function.image_copy_backup_file.secondary_allocationSpecifies the z/OS disk space secondary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.image_copy_function.image_copy_file.secondary_allocationSpecifies the z/OS disk space secondary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.image_copy_function.recovery_backup.secondary_allocationSpecifies the z/OS disk space secondary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.image_copy_function.recovery_file.secondary_allocationSpecifies the z/OS disk space secondary allocation amount. The range of values is from 1 to 1677215. If empty (default), then its value equals computed (from the row estimate) space requirements in cylinders or tracks, respectively to Space type setting. For Map dataset, the resulting value is doubled.
load_to_zos.set_copy_pendingSpecifies whether or not the table space is set to the copy-pending status. (Applicable only when Load with logging = No). Default: false
load_control.sort_buffer_sizeSpecifies the number of pages (size 4KB) of memory that are used for sorting index keys during a load operation.. Default: 0
load_to_zos.data_file_attributes.discard_data_set.space_typeSpecifies the z/OS disk space allocation unit type (SPACE(?,?)). Valid values are Cylinders (CYL) and Tracks (TRK). The default value is Cylinders.. Values: [cylinders, tracks]. Default: cylinders
load_to_zos.data_file_attributes.error_data_set.space_typeSpecifies the z/OS disk space allocation unit type (SPACE(?,?)). Valid values are Cylinders (CYL) and Tracks (TRK). The default value is Cylinders.. Values: [cylinders, tracks]. Default: cylinders
load_to_zos.data_file_attributes.input_data_files.space_typeSpecifies the z/OS disk space allocation unit type (SPACE(?,?)). Valid values are Cylinders (CYL) and Tracks (TRK). The default value is Cylinders.. Values: [cylinders, tracks]. Default: cylinders
load_to_zos.data_file_attributes.map_data_set.space_typeSpecifies the z/OS disk space allocation unit type (SPACE(?,?)). Valid values are Cylinders (CYL) and Tracks (TRK). The default value is Cylinders.. Values: [cylinders, tracks]. Default: cylinders
load_to_zos.data_file_attributes.work1_data_set.space_typeSpecifies the z/OS disk space allocation unit type (SPACE(?,?)). Valid values are Cylinders (CYL) and Tracks (TRK). The default value is Cylinders.. Values: [cylinders, tracks]. Default: cylinders
load_to_zos.data_file_attributes.work2_data_set.space_typeSpecifies the z/OS disk space allocation unit type (SPACE(?,?)). Valid values are Cylinders (CYL) and Tracks (TRK). The default value is Cylinders.. Values: [cylinders, tracks]. Default: cylinders
load_to_zos.image_copy_function.image_copy_backup_file.space_typeSpecifies the z/OS disk space allocation unit type (SPACE(?,?)). Valid values are Cylinders (CYL) and Tracks (TRK). The default value is Cylinders.. Values: [cylinders, tracks]. Default: cylinders
load_to_zos.image_copy_function.image_copy_file.space_typeSpecifies the z/OS disk space allocation unit type (SPACE(?,?)). Valid values are Cylinders (CYL) and Tracks (TRK). The default value is Cylinders.. Values: [cylinders, tracks]. Default: cylinders
load_to_zos.image_copy_function.recovery_backup.space_typeSpecifies the z/OS disk space allocation unit type (SPACE(?,?)). Valid values are Cylinders (CYL) and Tracks (TRK). The default value is Cylinders.. Values: [cylinders, tracks]. Default: cylinders
load_to_zos.image_copy_function.recovery_file.space_typeSpecifies the z/OS disk space allocation unit type (SPACE(?,?)). Valid values are Cylinders (CYL) and Tracks (TRK). The default value is Cylinders.. Values: [cylinders, tracks]. Default: cylinders
sql.user_defined_sql.statements *SQL statements to be executed for each input row
load_control.statisticsInstructs load to collect statistics during the load according to the profile defined for this table (STATISTICS).. Default: false
load_to_zos.statisticsThis option requests that statistics be displayed at the end of the load.. Values: [all, index, none, table]. Default: none
session.use_external_tables.statisticsInstructs load to collect statistics during the load according to the profile defined for this table (STATISTICS).. Default: false
session.use_external_tables.statistics.run_stats_on_columnsGenerates statistics on the columns. If no column specified, statistics are collected on all columns by default.
load_to_zos.data_file_attributes.discard_data_set.file_disposition.statusThis option specifies the disposition status of the dataset(s) used by the LOAD utility (DISP(?,?,?)). Specify the disposition status of the input data set used by load utility. This property is disabled when batch pipe system ID has any value. The default value is Replace. The valid values for this property are: Replace - Deletes an existing data set and creates a new data set (NEW). New - Indicates that the file does not currently exist (NEW). Old - Overwrites an existing data set or fails if the data set does not exist (OLD). Share - Identical to Old except that several jobs can read from the data set at the same time (SHR). Append - Appends to the end of an existing data set or creates a new data set if it does not already exist (MOD).. Values: [append, new, old, replace, share]. Default: replace
load_to_zos.data_file_attributes.error_data_set.file_disposition.statusThis option specifies the disposition status of the dataset(s) used by the LOAD utility (DISP(?,?,?)). Specify the disposition status of the input data set used by load utility. This property is disabled when batch pipe system ID has any value. The default value is Replace. The valid values for this property are: Replace - Deletes an existing data set and creates a new data set (NEW). New - Indicates that the file does not currently exist (NEW). Old - Overwrites an existing data set or fails if the data set does not exist (OLD). Share - Identical to Old except that several jobs can read from the data set at the same time (SHR). Append - Appends to the end of an existing data set or creates a new data set if it does not already exist (MOD).. Values: [append, new, old, replace, share]. Default: replace
load_to_zos.data_file_attributes.input_data_files.file_disposition.statusThis option specifies the disposition status of the dataset(s) used by the LOAD utility (DISP(?,?,?)). Specify the disposition status of the input data set used by load utility. This property is disabled when batch pipe system ID has any value. The default value is Replace. The valid values for this property are: Replace - Deletes an existing data set and creates a new data set (NEW). New - Indicates that the file does not currently exist (NEW). Old - Overwrites an existing data set or fails if the data set does not exist (OLD). Share - Identical to Old except that several jobs can read from the data set at the same time (SHR). Append - Appends to the end of an existing data set or creates a new data set if it does not already exist (MOD).. Values: [new, old, replace]. Default: replace
load_to_zos.data_file_attributes.map_data_set.file_disposition.statusThis option specifies the disposition status of the dataset(s) used by the LOAD utility (DISP(?,?,?)). Specify the disposition status of the input data set used by load utility. This property is disabled when batch pipe system ID has any value. The default value is Replace. The valid values for this property are: Replace - Deletes an existing data set and creates a new data set (NEW). New - Indicates that the file does not currently exist (NEW). Old - Overwrites an existing data set or fails if the data set does not exist (OLD). Share - Identical to Old except that several jobs can read from the data set at the same time (SHR). Append - Appends to the end of an existing data set or creates a new data set if it does not already exist (MOD).. Values: [append, new, old, replace, share]. Default: replace
load_to_zos.data_file_attributes.work1_data_set.file_disposition.statusThis option specifies the disposition status of the dataset(s) used by the LOAD utility (DISP(?,?,?)). Specify the disposition status of the input data set used by load utility. This property is disabled when batch pipe system ID has any value. The default value is Replace. The valid values for this property are: Replace - Deletes an existing data set and creates a new data set (NEW). New - Indicates that the file does not currently exist (NEW). Old - Overwrites an existing data set or fails if the data set does not exist (OLD). Share - Identical to Old except that several jobs can read from the data set at the same time (SHR). Append - Appends to the end of an existing data set or creates a new data set if it does not already exist (MOD).. Values: [append, new, old, replace, share]. Default: replace
load_to_zos.data_file_attributes.work2_data_set.file_disposition.statusThis option specifies the disposition status of the dataset(s) used by the LOAD utility (DISP(?,?,?)). Specify the disposition status of the input data set used by load utility. This property is disabled when batch pipe system ID has any value. The default value is Replace. The valid values for this property are: Replace - Deletes an existing data set and creates a new data set (NEW). New - Indicates that the file does not currently exist (NEW). Old - Overwrites an existing data set or fails if the data set does not exist (OLD). Share - Identical to Old except that several jobs can read from the data set at the same time (SHR). Append - Appends to the end of an existing data set or creates a new data set if it does not already exist (MOD).. Values: [append, new, old, replace, share]. Default: replace
load_to_zos.image_copy_function.image_copy_backup_file.file_disposition.statusThis option specifies the disposition status of the dataset(s) used by the LOAD utility (DISP(?,?,?)). Specify the disposition status of the input data set used by load utility. This property is disabled when batch pipe system ID has any value. The default value is Replace. The valid values for this property are: Replace - Deletes an existing data set and creates a new data set (NEW). New - Indicates that the file does not currently exist (NEW). Old - Overwrites an existing data set or fails if the data set does not exist (OLD). Share - Identical to Old except that several jobs can read from the data set at the same time (SHR). Append - Appends to the end of an existing data set or creates a new data set if it does not already exist (MOD).. Values: [append, new, old, replace, share]. Default: replace
load_to_zos.image_copy_function.image_copy_file.file_disposition.statusThis option specifies the disposition status of the dataset(s) used by the LOAD utility (DISP(?,?,?)). Specify the disposition status of the input data set used by load utility. This property is disabled when batch pipe system ID has any value. The default value is Replace. The valid values for this property are: Replace - Deletes an existing data set and creates a new data set (NEW). New - Indicates that the file does not currently exist (NEW). Old - Overwrites an existing data set or fails if the data set does not exist (OLD). Share - Identical to Old except that several jobs can read from the data set at the same time (SHR). Append - Appends to the end of an existing data set or creates a new data set if it does not already exist (MOD).. Values: [append, new, old, replace, share]. Default: replace
load_to_zos.image_copy_function.recovery_backup.file_disposition.statusThis option specifies the disposition status of the dataset(s) used by the LOAD utility (DISP(?,?,?)). Specify the disposition status of the input data set used by load utility. This property is disabled when batch pipe system ID has any value. The default value is Replace. The valid values for this property are: Replace - Deletes an existing data set and creates a new data set (NEW). New - Indicates that the file does not currently exist (NEW). Old - Overwrites an existing data set or fails if the data set does not exist (OLD). Share - Identical to Old except that several jobs can read from the data set at the same time (SHR). Append - Appends to the end of an existing data set or creates a new data set if it does not already exist (MOD).. Values: [append, new, old, replace, share]. Default: replace
load_to_zos.image_copy_function.recovery_file.file_disposition.statusThis option specifies the disposition status of the dataset(s) used by the LOAD utility (DISP(?,?,?)). Specify the disposition status of the input data set used by load utility. This property is disabled when batch pipe system ID has any value. The default value is Replace. The valid values for this property are: Replace - Deletes an existing data set and creates a new data set (NEW). New - Indicates that the file does not currently exist (NEW). Old - Overwrites an existing data set or fails if the data set does not exist (OLD). Share - Identical to Old except that several jobs can read from the data set at the same time (SHR). Append - Appends to the end of an existing data set or creates a new data set if it does not already exist (MOD).. Values: [append, new, old, replace, share]. Default: replace
load_control.partitioned_db_config.status_intervalSpecifies the number of megabytes of data to load before generating a progress message (STATUS_INTERVAL).. Default: 100
load_to_zos.data_file_attributes.discard_data_set.storage_classSpecifies the SMS storage class (STORCLAS). The name must be a valid SMS storage class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.error_data_set.storage_classSpecifies the SMS storage class (STORCLAS). The name must be a valid SMS storage class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.input_data_files.storage_classSpecifies the SMS storage class (STORCLAS). The name must be a valid SMS storage class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.map_data_set.storage_classSpecifies the SMS storage class (STORCLAS). The name must be a valid SMS storage class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work1_data_set.storage_classSpecifies the SMS storage class (STORCLAS). The name must be a valid SMS storage class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work2_data_set.storage_classSpecifies the SMS storage class (STORCLAS). The name must be a valid SMS storage class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_backup_file.storage_classSpecifies the SMS storage class (STORCLAS). The name must be a valid SMS storage class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_file.storage_classSpecifies the SMS storage class (STORCLAS). The name must be a valid SMS storage class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_backup.storage_classSpecifies the SMS storage class (STORCLAS). The name must be a valid SMS storage class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_file.storage_classSpecifies the SMS storage class (STORCLAS). The name must be a valid SMS storage class and must not exceed 8 characters in length. If empty (default), then the property is not used.
load_to_zos.image_copy_function.system_pagesSpecifies whether the copy utility puts system pages at the beginning of the image-copy file. Default: true
table_action *Select the action to perform on the database table. Values: [append, create, replace, truncate]. Default: append
session.temporary_work_table.table_name *The name of the existing temporary work table.
table_name *The table name to be used in generated SQL. The table name must be schema qualified in order to preview data.
table_action.generate_create_statement.create_table_on_zosDetermines whether the target table is on Db2 LUW or Db2 for z/OS.. Default: false
load_control.directory_for_tmp_filesSpecifies the path name that is used by the Db2 server to store temporary files.
session.temporary_work_tableIf set to Automatic, the connector will automatically create the temporary work table using an internally generated name.. Values: [automatic, existing]. Default: automatic
transaction.time_intervalSpecify the amount of time to pass before a commit is issued. If you set a small value you force frequent commits and therefore if your program terminates unexpectedly, your table can still contain partial results. However, you may pay a performance penalty because of the high frequency of commits. If you set a large value, Db2 must log a correspondingly large amount of rollback information which may also slow your job.. Default: 0
limit_parallelism.player_process_limit *The total number of player processes across all processing nodes.
load_control.partitioned_db_config.traceSpecifies the number of records to trace in a dump of the data conversion process and the output of the hashing values (TRACE).. Default: 0
load_to_zos.transfer.transfer_cmdUser-entered command sent just before data transfer begins. FTP example: quote site vcount=7 datakeepalive=60
load_to_zos.transfer.transfer_to *Name of the target machine to which the data is sent.
load_to_zos.transfer.transfer_typeDetermines the method of transferring data to the z/OS machine.. Values: [ftp, lftp, sftp]. Default: ftp
sql.user_defined_sql.suppress_warningsDo not abort the job when a Db2 Warning is encountered, report as an information messsage.. Default: false
table_action.generate_truncate_statement.truncate_statement *A statement to be executed when truncating the database table
session.temporary_work_table.truncate_tableIf set to Yes, the temporary work table is truncated before any data is written to it.. Default: false
table_action.generate_create_statement.create_table_compress.create_table_compress_luwSpecifies whether adaptive compression or classic row compression is used (Db2 LUW only).. Values: [adaptive, static]. Default: adaptive
load_to_zos.uss_pipe_directory *The presence of this option indicates that USS piping is to be used and gives the directory name where the pipes will be created. The value should be a fully qualified USS directory name.
sql.use_unique_key_column.unique_key_column *Unique key column name used in UPDATE statement
load_to_zos.data_file_attributes.discard_data_set.unitSpecifies the device number, device type (generic), or group name for the data set (UNIT). If empty (default), then the property is not used.
load_to_zos.data_file_attributes.error_data_set.unitSpecifies the device number, device type (generic), or group name for the data set (UNIT). If empty (default), then the property is not used.
load_to_zos.data_file_attributes.input_data_files.unitSpecifies the device number, device type (generic), or group name for the data set (UNIT). If empty (default), then the property is not used.
load_to_zos.data_file_attributes.map_data_set.unitSpecifies the device number, device type (generic), or group name for the data set (UNIT). If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work1_data_set.unitSpecifies the device number, device type (generic), or group name for the data set (UNIT). If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work2_data_set.unitSpecifies the device number, device type (generic), or group name for the data set (UNIT). If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_backup_file.unitSpecifies the device number, device type (generic), or group name for the data set (UNIT). If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_file.unitSpecifies the device number, device type (generic), or group name for the data set (UNIT). If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_backup.unitSpecifies the device number, device type (generic), or group name for the data set (UNIT). If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_file.unitSpecifies the device number, device type (generic), or group name for the data set (UNIT). If empty (default), then the property is not used.
sql.update_columnsComma separated list of columns to be updated with UPDATE statement. If no column specified, all columns are updated.
sql.update_statement *Statement to be executed when updating rows in the database
session.use_external_tablesIndicates whether external tables are used.. Default: false
sql.use_unique_key_columnUse Unique key column in UPDATE statement. Default: false
load_to_zos.transfer.userDetermines the user name for transferring.
sql.user_defined_sql *Source of the user-defined SQL statements. Values: [file, statements]. Default: statements
load_to_zos.utility_idThis is a unique identifier within Db2 for the execution of the LOAD utility. The default value is Db2ZLOAD.. Default: DB2ZLOAD
table_action.generate_create_statement.create_table_value_compressionThis determines the row format that is to be used. Each data type has a different byte count depending on the row format that is used (Db2 LUW only).. Default: false
load_to_zos.data_file_attributes.discard_data_set.volumesSpecifies a list of volume serial numbers for this allocation (VOLUMES(?)). The value or list of values may be entered by the user with or without enclosing parentheses. In the TEMPLATE statement, however, the parentheses are required. So, if the user does not enter them, the syntax generator should add them. The value for this property is a comma-separated list of string values. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.error_data_set.volumesSpecifies a list of volume serial numbers for this allocation (VOLUMES(?)). The value or list of values may be entered by the user with or without enclosing parentheses. In the TEMPLATE statement, however, the parentheses are required. So, if the user does not enter them, the syntax generator should add them. The value for this property is a comma-separated list of string values. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.input_data_files.volumesSpecifies a list of volume serial numbers for this allocation (VOLUMES(?)). The value or list of values may be entered by the user with or without enclosing parentheses. In the TEMPLATE statement, however, the parentheses are required. So, if the user does not enter them, the syntax generator should add them. The value for this property is a comma-separated list of string values. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.map_data_set.volumesSpecifies a list of volume serial numbers for this allocation (VOLUMES(?)). The value or list of values may be entered by the user with or without enclosing parentheses. In the TEMPLATE statement, however, the parentheses are required. So, if the user does not enter them, the syntax generator should add them. The value for this property is a comma-separated list of string values. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work1_data_set.volumesSpecifies a list of volume serial numbers for this allocation (VOLUMES(?)). The value or list of values may be entered by the user with or without enclosing parentheses. In the TEMPLATE statement, however, the parentheses are required. So, if the user does not enter them, the syntax generator should add them. The value for this property is a comma-separated list of string values. If empty (default), then the property is not used.
load_to_zos.data_file_attributes.work2_data_set.volumesSpecifies a list of volume serial numbers for this allocation (VOLUMES(?)). The value or list of values may be entered by the user with or without enclosing parentheses. In the TEMPLATE statement, however, the parentheses are required. So, if the user does not enter them, the syntax generator should add them. The value for this property is a comma-separated list of string values. If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_backup_file.volumesSpecifies a list of volume serial numbers for this allocation (VOLUMES(?)). The value or list of values may be entered by the user with or without enclosing parentheses. In the TEMPLATE statement, however, the parentheses are required. So, if the user does not enter them, the syntax generator should add them. The value for this property is a comma-separated list of string values. If empty (default), then the property is not used.
load_to_zos.image_copy_function.image_copy_file.volumesSpecifies a list of volume serial numbers for this allocation (VOLUMES(?)). The value or list of values may be entered by the user with or without enclosing parentheses. In the TEMPLATE statement, however, the parentheses are required. So, if the user does not enter them, the syntax generator should add them. The value for this property is a comma-separated list of string values. If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_backup.volumesSpecifies a list of volume serial numbers for this allocation (VOLUMES(?)). The value or list of values may be entered by the user with or without enclosing parentheses. In the TEMPLATE statement, however, the parentheses are required. So, if the user does not enter them, the syntax generator should add them. The value for this property is a comma-separated list of string values. If empty (default), then the property is not used.
load_to_zos.image_copy_function.recovery_file.volumesSpecifies a list of volume serial numbers for this allocation (VOLUMES(?)). The value or list of values may be entered by the user with or without enclosing parentheses. In the TEMPLATE statement, however, the parentheses are required. So, if the user does not enter them, the syntax generator should add them. The value for this property is a comma-separated list of string values. If empty (default), then the property is not used.
load_control.warning_countSpecifies the number of warnings that are allowed before the load operation stops.. Default: 0
load_control.without_promptingSelect Yes to add the WITHOUT PROMPTING parameter to the LOAD command in the generated command file.. Default: false
write_mode *The mode to be used when writing to a database table. Values: [bulk_load, delete, delete_then_insert, insert, insert_new_rows_only, insert_then_update, update, update_then_insert, user-defined_sql]. Default: insert




IBM Db2 Big SQL


Description: IBM Db2 Big SQL is an enterprise-grade, hybrid ANSI-compliant SQL-on-Hadoop engine that delivers massively parallel processing (MPP) and advanced data query.
Data source type ID: 2bdd9544-f13a-47b6-b6c3-f5964a08066a
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Db2 Warehouse


Description: IBM Db2 Warehouse is an analytics data warehouse that offers a high level of control over your data and applications. You can use this connector to connect to a database in IBM Db2 Warehouse instances on the cloud or in an on-premises environment.
Data source type ID: cfdcb449-1204-44ba-baa6-9a8a878e6aa7
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
api_key *An application programming interface key that identifies the calling application or user
auth_method. Values: [apikey, username_password]
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
portThe port of the database. Default: 50001
sslThe port is configured to accept SSL connections. Default: true
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Db2 for i


Description: IBM Db2 for i is an integrated Relational Database Management System that leverages the high performance, virtualization, and energy efficiency features of IBM Power Systems.
Data source type ID: 335cbfe7-e495-474e-8ad7-78ad63c05091
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
host *The hostname or IP address of the database
database *The unique name of the Db2 location you want to access
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Db2 for z/OS


Description: IBM Db2 for z/OS is an enterprise data server for IBM Z. It manages core business data across an enterprise and supports key business applications.
Data source type ID: c8d3eab2-25f6-4a90-8e10-0b4226693c45
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
collectionThe ID of the collection of packages to use
host *The hostname or IP address of the database
database *The unique name of the Db2 location you want to access
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [none, random]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Db2 on Cloud


Description: IBM Db2 on Cloud is an SQL database that is fully managed for you by IBM Cloud.
Data source type ID: 506039fb-802f-4ef2-a2bf-c1682e9c8aa2
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
portThe port of the database. Default: 50001
sslThe port is configured to accept SSL connections. Default: true
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Informix


Description: IBM Informix is a database that contains relational, object-relational, or dimensional data. You can use the Informix connector to access data from an on-prem Informix database server or from IBM Informix on Cloud.
Data source type ID: 029e5d1c-ba73-4b09-b742-14c3a39b6cf9
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
db_localeThe value of DB_LOCALE property.
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
server *The name of the database server to connect to
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [none, random]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM MQ


Description: IBM MQ is messaging and queuing middleware that works with a broad range of computing platforms, applications, web services, and communications protocols for security-rich message delivery. MQ was formerly named IBM WebSphere MQ.
Data source type ID: 21364ca9-5b2d-323e-bd4d-59ba961f75fb
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
client_channel_definition.channel_nameName of the channel.
client_channel_definition.connection_nameName of the connection. The format of this value must match the selected transport type.
passwordPassword for the username that connects to the MQ server.
queue_manager_nameName of the queue manager to access. The value must match the queue manager in the client connection channel definition.
client_channel_definition.transport_typeTransport protocol to use.. Values: [tcp, udp]. Default: tcp
usernameName of the user that connects to the MQ server.

Interaction properties (when used as a source)

Name Type Description
access_modeAccess mode to use when opening source queue. Values: [as_in_queue_definition, exclusive, exclusive_if_granted, shared]. Default: as_in_queue_definition
other_queue_settings.alternate_user_idAlternate user identifier to specify when opening the queue.
work_queue.append_node_numberIf running in parallel, there may be multiple work queues each with node number appended. This property spedifies whether to append the node number to the work queue name.. Default: true
header_fields_filter.appl_identity_dataList of acceptable application identity data values for source messages
header_fields_filter.appl_origin_dataList of application origin data values for source messages
header_fields_filter.backout_countList of acceptable backout count values and ranges for source messages
transaction.end_of_dayEnable blocking transaction processing. Default: false
header_fields_filter.coded_char_set_idList of acceptable coded character set identifier values and ranges for source messages
message_options.message_conversion.coded_char_set_idCoded character set identifier to which to convert character data in source messages. When omitted or set to the default value 0, the 'DEFAULT' coded character set identifier for the current queue manager connection is assumed. The values -1 and -2 are used to specify special 'INHERIT' and 'EMBEDDED' coded character set identifier values. Default: 0
pub_sub.content_filterContent filter value to specify when registering and/or deregistering the subscriber. If neither registration nor deregistration is enabled, this value is ignored
error_queue.context_modeContext mode to use when opening error queue. The value should be chosen based on whether messages that may end up on the error queue should preserve identity context fields (Set identity), origin context fields (Set all) or none of the context fields (None). Values: [none, set_all, set_identity]. Default: none
work_queue.context_modeContext mode to use for opening work queue. It defines whether to preserve none, identity or all of the context message fields in the messages moved from the source queue to the work queue.. Values: [none, set_all, set_identity]. Default: set_all
pub_sub.deregistration.deregistration_correl_idCorrelation identifier for deregistration
pub_sub.registration.registration_correl_idCorrelation identifier for registration
header_fields_filter.feedback.custom_valueList of acceptable feedback and reason code custom values and ranges for source report messages
header_fields_filter.format.custom_valueList of acceptable format custom values for source messages
header_fields_filter.msg_type.custom_valueList of acceptable message type custom values and ranges for source messages
header_fields_filter.put_appl_type.custom_valueList of acceptable put application type custom values for source messages
pub_sub.deregistrationSwitch that controls whether to deregister with the broker when the job ends. Default: false
pub_sub.pub_sub_dynamic_reply_to_queueSwitch that controls whether to open the reply queue as dynamic queue. Default: false
message_options.pass_by_referenceThe option to enable for passing payload data by reference. When this option is not enabled, the payload data is passed inline. Default: false
header_fields_filter.encodingList of acceptable encoding values and ranges for source messages
message_options.message_conversion.encodingEncoding to which to convert numeric data in source messages. When omitted or set to default value -1, the 'NATIVE' numeric encoding for the current queue manager connection is assumed. Default: -1
transaction.end_of_wave.end_of_dataSpecify whether to insert EOW marker for the final set of records when their number is smaller than the value specified for the transaction record count. Note that if the specified transaction record count value is 0 (representing all available records), there is only one transaction wave which consists of all the records, and so the End of data value should be set to Yes in order for EOW marker to be inserted for that transaction wave. Default: true
end_of_data_message_typeMessage type that marks the end message receiving. When the connector receives a message of this message type, it stops reading additional messages from the queue (except other messages from the same group when message group assembly is required)
transaction.end_of_waveProvide settings for the end of wave handling. None means EOW markers are never inserted, Before means EOW markers are inserted before completing the transaction, After means EOW markers are inserted after completing the transaction. Values: [after, before, none]. Default: none
error_queueSwitch that controls whether to use error queue. Default: false
header_fields_filter.expiryList of acceptable expiry interval values and ranges (in tenths of seconds) for source messages
message_options.extract_keySwitch that controls whether to extract a binary value from the message payload and provide it on output through the column that has WSMQ.EXTRACTEDKEY data element specified.. Default: false
header_fields_filterSwitch that controls whether to filter source messages based on the provided filtering criteria. Default: false
pub_sub.deregistration.subscriberGeneral deregistration options for the subscriber. Values: [correlation_id_as_identity, deregister_all, leave_only, variable_user_id]
pub_sub.registration.subscriber_generalGeneral registration options for the subscriber. Note: Option [Anonymous] is ignored in MQRFH2 service mode. Values: [anonymous, correlation_id_as_identity, duplicates_ok, local, new_publications_only]
other_queue_settings.alternate_security_id.hexSwitch that controls whether the provided alternate security identifier value should be treated as an array of hex-digit pairs rather than a plain text value. Since the only currently supported Alternate security ID values are 40-byte long values that contain the Windows SID of the user, this property value should always be set to Yes. Default: true
header_fields_filter.accounting_token.hexSwitch that controls whether the provided accounting token value should be treated as an array of hex-digit pairs rather than a plain text value. Default: false
header_fields_filter.correl_id.hexSwitch that controls whether the provided correlation identifier value should be treated as an array of hex-digit pairs rather than a plain text value. Default: false
header_fields_filter.group_id.hexSwitch that controls whether the provided group identifier value should be treated as an array of hex-digit pairs rather than a plain text value. Default: false
header_fields_filter.msg_id.hexSwitch that controls whether the provided message identifier value should be treated as an array of hex-digit pairs rather than a plain text value. Default: false
pub_sub.registration.subscriber_identityIdentity-related registration options for the subscriber. Values: [add_name, join_exclusive, join_shared, no_alteration, variable_user_id]
message_options.extract_key.key_length *Length in bytes of the binary value to extract. The value -1 specifies that the binary key value should contain all bytes from the specified key offset to the end of the message payload.. Default: 0
message_options.extract_key.key_offset *Offset in bytes in the message payload from which to extract binary key value.. Default: 0
header_fields_filter.msg_flags.must_match_allSwitch that controls whether source messages must have ALL of the specified message flag values or ANY of the specified message flag values in order to be accepted. Default: false
header_fields_filter.report.must_match_allSwitch that controls whether source messages must have ALL of the specified report values or ANY of the specified report values in order to be accepted. Default: false
work_queue.monitor_queue_depth.max_queue_depth *Maximum queue depth that the connector allows for the queue
transaction.message_controlledSettings for the module and function that the connector should invoke for each input message to determine whether the transaction should be committed after each message. Default: false
message_options.message_conversionSwitch that controls whether to perform conversion of numeric and character data in source messages. Default: false
message_optionsVarious options that control ordering, structure and access mode for the message. Default: false
message_options.message_order_and_assemblyValue that specifies how to retrieve message segments and group messages from the source queue. Values: [assemble_groups, assemble_logical_messages, individual_ordered, individual_unordered]. Default: individual_ordered
message_options.message_paddingSwitch that controls whether to pad message payload column for source messages. When selected, message payload text column is padded with space character, and message payload binary column is padded with NULL byte value. Default: false
message_quantityNumber of messages to retrieve from the source queue. Note that this is the number of queue messages, not the number of rows. When message group assembly is required, each group counts as one message. Default: -1
message_read_modeMode for reading source messages. Messages can be read within or outside of transaction, and they can be received destructively or kept on the source queue. Values: [delete, delete_under_transaction, keep, move_to_work_queue]. Default: delete_under_transaction
header_fields_filter.msg_seq_numberList of acceptable message sequence number values and ranges for source messages
message_options.message_truncationSwitch that controls whether to allow truncation of the source message payload so that it fits in the message payload column. When selected, message payload for text message column is truncated to the column length in characters, and message payload for binary payload column is truncated to the column size in bytes. Default: true
transaction.end_of_day.method_name *Name of the method that determines whether a message represents a blocking transaction
transaction.message_controlled.method_name *Name of the method
work_queue.monitor_queue_depth.min_queue_depth *Minimum queue depth that the connector tries to maintain on the queue.
transaction.end_of_day.module_name *Fully-qualified name of the module (shared library) that implements the method for identifying blocking transactions
transaction.message_controlled.module_name *Fully-qualified name of the module (shared library)
work_queue.monitor_queue_depthSwitch that controls whether to monitor the work queue depth. Default: false
work_queue.name *Name of the work queue.
header_fields_filter.offsetList of acceptable offset values and ranges for source messages
header_fields_filter.original_lengthList of acceptable original length values and ranges for source messages (used for segments of report messages)
other_queue_settingsAdditional settings for the source queue from which to receive messages. Default: false
header_fields_filter.msg_payload_sizeList of acceptable message payload size values and ranges for source messages (format headers are not counted towards payload size)
refresh.periodAmount of time after which the cursor should be periodically rewound. When omitted, the cursor is rewound each time the end of the queue is reached. Default: -1
header_fields_filter.persistenceList of acceptable persistence values and ranges for source messages. Values: [as_in_queue_definition, not_persistent, persistent]
pub_sub.registration.subscriber_persistencePersistence registration options for the subscriber. Values: [non_persistent, persistent, persistent_as_publish, persistent_as_queue]. Default: persistent_as_publish
header_fields_filter.priorityList of acceptable priority values and ranges for source messages
end_of_data_message_type.process_end_of_data_messageSwitch that controls whether the end of data message should also be processed and provided on output.. Default: true
pub_subSwitch that controls whether the connector is in Publish/Subscribe mode of operation. Default: false
header_fields_filter.put_appl_nameList of acceptable put application name values for source messages
header_fields_filter.put_dateList of acceptable put date values and ranges for source messages (in YYYYMMDD format)
header_fields_filter.put_timeList of acceptable put time values and ranges for source messages (in HHMMSSTH format)
error_queue.queue_manager_nameName of the queue manager that hosts the error queue. When the value is not provided, the connector assumes that the error queue resides on the currently connected queue manager.
error_queue.name *Name of the error queue.
pub_sub.pub_sub_dynamic_reply_to_queue.name *Name of the dynamic reply queue. Use asterisk to specify incomplete name. Part of the name right of the asterisk (including the asterisk) will be automatically generated by the queue manager. Only one asterisk character may be specified. If specified, it must be the last character in the value and its position must be between 1 and 33 (inclusive). Default: *
queue_nameName of the source queue from which to receive messages. In publish/subscribe mode this is used as the subscriber queue. Note: if dynamic queue options are specified, this value is the name of the model queue to use as template for creating the dynamic queue
transaction.record_countNumber of records per transaction. The value 0 means all available records. Default: 0
refreshSwitch that controls whether to periodically rewind the cursor on the source queue. Default: false
pub_sub.registrationSwitch that controls whether to register with the broker when the job starts. Default: false
message_options.remove_mqrfh2headerSwitch that controls whether to remove MQRFH2 header.When selected, only the message body is passed.. Default: false
pub_sub.pub_sub_reply_to_queueName of the queue to which the broker should send replies for the command messages. Note: If dynamic reply queue usage is specified, the value specified here is used as the model queue name
header_fields_filter.reply_to_qList of acceptable reply to queue values for source messages
header_fields_filter.reply_to_q_mgrList of acceptable reply to queue manager values for source messages
pub_sub.service_typeRules and formatting header version to use for command messages. Values: [mqrfh, mqrfh2]. Default: mqrfh
pub_sub.stream_nameStream name value to specify when registering and/or deregistering the subscriber. If neither registration nor deregistration is enabled, this value is ignored. Default: SYSTEM.BROKER.DEFAULT.STREAM
pub_sub.sub_identitySubscription identity value to specify when registering and/or deregistering the subscriber. If neither registration nor deregistration is enabled, this value is ignored
pub_sub.sub_nameSubscription name value to specify when registering and/or deregistering the subscriber. If neither registration nor deregistration is enabled, this value is ignored
pub_sub.sub_pointSubscription point value to specify when registering and/or deregistering the subscriber. If neither registration nor deregistration is enabled, this value is ignored
header_fields_filter.feedback.system_valueAcceptable feedback and reason code system values for source report messages. Values: [confirm_on_arrival, confirm_on_delivery, expiration, message_too_big_for_queue_mqrc, message_too_big_for_queue_manager_mqrc, negative_action_notification, none, not_authorized_mqrc, persistent_not_allowed_mqrc, positive_action_notification, put_inhibited_mqrc, queue_full_mqrc, queue_space_not_available_mqrc, quit]
header_fields_filter.format.system_valueAcceptable format system values for source messages. Values: [mqadmin, mqchcom, mqcics, mqcmd1, mqcmd2, mqdead, mqevent, mqhdist, mqhmde, mqhref, mqhrf, mqhrf2, mqhwih, mqims, mqimsvs, mqnone, mqpcf, mqstr, mqtrig, mqxmit]
header_fields_filter.msg_type.system_valueAcceptable message type system values for source messages. Values: [datagram, reply, report, request]
header_fields_filter.put_appl_type.system_valueAcceptable put application type system values for source messages. Values: [aix,_unix, broker, channelinitiator, cics, cicsbridge, cicsvse, dos, dqm, guardian,_nsk, ims, imsbridge, java, mvs,_os390,_zos, nocontext, notesagent, os2, os400, qmgr, unknown, user, vms, vos, windows, windowsnt, xcf]
transaction.time_intervalNumber of seconds per transaction. The value 0 means unlimited time. Default: 0
transaction.end_of_day.timeoutMaximum amount of time to wait for a blocking transaction to complete. The value -1 specifies unlimited time. The operation fails if the transaction does not complete within the specified time.. Default: -1
pub_sub.deregistration.deregistration_topicTopic(s) for deregistration
pub_sub.registration.registration_topic *Topic(s) for registration
error_queue.tranmission_queue_nameName of the transmission queue to use when the error queue is a remote queue. If the value is not specified, the default transmission queue is used.
message_options.treat_eol_as_row_terminatorSwitch that controls whether end-of-line character in the message payload should be treated as row terminator. When selected, each source message may result in multiple rows of data. Default: false
header_fields_filter.accounting_token.use_wildcardSwitch that controls whether the initial asterisk in the provided accounting token value should be treated as a wildcard rather than plain text. Default: false
header_fields_filter.correl_id.use_wildcardSwitch that controls whether the initial asterisk in the provided correlation identifier value should be treated as a wildcard rather than plain text. Default: false
header_fields_filter.group_id.use_wildcardSwitch that controls whether the initial asterisk in the provided group identifier value should be treated as a wildcard rather than plain text. Default: false
header_fields_filter.msg_id.use_wildcardSwitch that controls whether the initial asterisk in the provided message identifier value should be treated as a wildcard rather than plain text. Default: false
header_fields_filter.user_identifierList of acceptable user identifier values for source messages
other_queue_settings.alternate_security_id.valueThe value for the alternate security identifier.
header_fields_filter.accounting_token.valueAcceptable accounting token value for source messages.
header_fields_filter.correl_id.valueAcceptable correlation identifier value for source messages.
header_fields_filter.group_id.valueAcceptable group identifier value for source messages.
header_fields_filter.msg_id.valueAcceptable message identifier value for source messages.
header_fields_filter.msg_flags.valueAcceptable message flag values for source messages. Values: [last_message_in_group, last_segment, message_in_group, segment, segmentation_allowed]
header_fields_filter.report.valueAcceptable report values for source messages. Values: [confirm_on_arrival, confirm_on_arrival_with_data, confirm_on_arrival_with_full_data, confirm_on_delivery, confirm_on_delivery_with_data, confirm_on_delivery_with_full_data, discard_message, exception, exception_with_data, exception_with_full_data, expiration, expiration_with_data, expiration_with_full_data, negative_action_notification, pass_correlation_id, pass_message_id, positive_action_notification]
wait_timeMaximum amount of time (in seconds) to wait for a new message to arrive on the source queue. Default: -1

Interaction properties (when used as a target)

Name Type Description
other_queue_settings.alternate_user_idAlternate user identifier to specify when opening the queue.
header_fields_setter.appl_identity_dataApplication identity data value for destination messages
header_fields_setter.appl_origin_dataApplication origin data to set for the destination messages
other_queue_settings.cluster_queue.binding_modeBinding mode to use when selecting physical queue instance from the cluster. Options are to resolve the instance when opening the shared cluster queue or each time a message is sent to the shared cluster queue. Additionally, the default binding mechanism for the shared cluster queue may be used. Values: [as_in_queue_definition, not_fixed, on_open]. Default: as_in_queue_definition
other_queue_settings.cluster_queueSwitch that controls whether to access destination queue as a shared cluster queue. Default: false
header_fields_setter.coded_char_set_idCoded character set identifier value to set for destination messages. The values 0, -1 and -2 are used to specify the special 'DEFAULT', 'INHERIT' and 'EMBEDDED' coded character set identifier values. Default: 0
context_modeContext mode to use when opening destination queue. The value should be chosen based on whether destination messages will include identity context fields (Set identity), origin context fields (Set all) or none of the context fields (None). Values: [none, set_all, set_identity]. Default: none
error_queue.context_modeContext mode to use when opening error queue. The value should be chosen based on whether messages that may end up on the error queue should preserve identity context fields (Set identity), origin context fields (Set all) or none of the context fields (None). Values: [none, set_all, set_identity]. Default: none
pub_sub.deregistration.deregistration_correl_idCorrelation identifier for deregistration
pub_sub.registration.registration_correl_idCorrelation identifier for registration
pub_sub.publish.publication_format.custom_valueFormat custom value to set for the publication payload
header_fields_setter.feedback.custom_valueFeedback and reason code custom value to set for destination messages
header_fields_setter.format.custom_valueFormat custom value to set for destination messages
header_fields_setter.msg_type.custom_valueMessage type custom value to set for destination messages
header_fields_setter.put_appl_type.custom_valuePut application type custom value for destination messages
pub_sub.deregistrationSwitch that controls whether to deregister with the broker when the job ends. Default: false
other_queue_settings.dynamic_queueSwitch that controls whether to open the queue as dynamic queue. To open the queue as a dynamic queue, use a model queue name where a regular queue name was specified. Default: false
pub_sub.pub_sub_dynamic_reply_to_queueSwitch that controls whether to open the reply queue as dynamic queue. Default: false
header_fields_setter.encodingEncoding value to set for destination messages. The value -1 is used to specify the special 'NATIVE' encoding value. Default: -1
error_queueSwitch that controls whether to use error queue. Default: false
header_fields_setter.expiryExpiry interval value (in tenths of seconds) to set for destination messages. Default: -1
pub_sub.deregistration.publisherGeneral deregistration options for the publisher. Values: [correlation_id_as_identity, deregister_all]
pub_sub.registration.publisherGeneral registration options for the publisher. Values: [anonymous, correlation_id_as_identity, local]
header_fields_setter.versionHeader version value to set for destination messages. Values: [1, 2]. Default: 2
other_queue_settings.alternate_security_id.hexSwitch that controls whether the provided alternate security identifier value should be treated as an array of hex-digit pairs rather than a plain text value. Since the only currently supported Alternate security ID values are 40-byte long values that contain the Windows SID of the user, this property value should always be set to Yes. Default: true
header_fields_setter.accounting_token.hexSwitch that controls whether the provided accounting token value should be treated as an array of hex-digit pairs rather than a plain text value. Default: false
header_fields_setter.correl_id.hexSwitch that controls whether the provided correlation identifier value should be treated as an array of hex-digit pairs rather than a plain text value. Default: false
header_fields_setter.group_id.hexSwitch that controls whether the provided group identifier value should be treated as an array of hex-digit pairs rather than a plain text value. Default: false
header_fields_setter.msg_id.hexSwitch that controls whether the provided message identifier value should be treated as an array of hex-digit pairs rather than a plain text value. Default: false
pub_sub.publish.message_content_descriptorSwitch that controls whether to include message content descriptor service folder () in the publication messages. Default: false
header_fields_setter.msg_flagsMessage flags to set for destination messages. Note: if message segmentation was selected, the connector will automatically set the offset values on the generated message segments. The flags for Segmentation Allowed, Segment and Last Segment in this property will be ignored. Values: [last_message_in_group, last_segment, message_in_group, segment, segmentation_allowed]
message_optionsVarious options that control ordering, structure and access mode for the message. Default: false
header_fields_setter.msg_seq_numberMessage sequence number to set for destination messages. Default: 1
pub_sub.publish.msg_seq_numberSwitch that controls whether to update message sequence number in the published messages. Default: false
pub_sub.publish.message_content_descriptor.message_service_domain *Service domain for the publication messages. Values: [idoc, mrm, none, xml, xmlns]. Default: mrm
pub_sub.publish.message_content_descriptor.message_setName of the message set for the publication messages
pub_sub.publish.message_content_descriptor.message_typeName of the message type for the publication messages
message_write_modeMode for writing destination messages. Messages can be written within or outside of transaction, and the connector can be configured to send only messages that have non-zero length payload. Values: [create, create_under_transaction, create_on_content, create_on_content_under_transaction]. Default: create_under_transaction
header_fields_setter.offsetOffset value to set for destination messages. Note: if message segmentation was selected, the connector will automatically set the offset values on the generated message segments. The value of this property will be ignored. Default: 0
other_queue_settingsAdditional settings for the source queue from which to receive messages. Default: false
header_fields_setter.persistencePersistence value to set for destination messages. Values: [as_in_queue_definition, not_persistent, persistent]. Default: as_in_queue_definition
pub_sub.publish.message_content_descriptor.mrm_physical_formatName of the MRM physical format in the specified message set used for the publication messages
header_fields_setter.priorityPriority value to set for destination messages. Default: -1
pub_sub.publish.publicationPublication options to specify when publishing messages. Note: Option [No Registration] is ignored in MQRFH2 service mode. Values: [correlation_id_as_identity, no_registration, retain_publication]
pub_subSwitch that controls whether the connector is in Publish/Subscribe mode of operation. Default: false
header_fields_setter.put_appl_namePut application name value for destination messages
header_fields_setter.put_datePut date value to set for destination messages (in YYYYMMDD format)
header_fields_setter.put_timePut time value to set for destination messages (in HHMMSSTH format)
error_queue.queue_manager_nameName of the queue manager that hosts the error queue. When the value is not provided, the connector assumes that the error queue resides on the currently connected queue manager.
other_queue_settings.cluster_queue.queue_manager_nameName of the cluster queue manager. If the value is not specified, the queue manager is selected dynamically from the cluster.
error_queue.name *Name of the error queue.
other_queue_settings.dynamic_queue.name *Name of the dynamic queue. Use asterisk to specify incomplete name (stem). Part of the name to the right of the asterisk (including the asterisk) will be automatically generated by the queue manager. Only one asterisk character may be specified. If specified, it must be the last character in the value and its position must be between 1 and 33 (inclusive). Default: *
pub_sub.pub_sub_dynamic_reply_to_queue.name *Name of the dynamic reply queue. Use asterisk to specify incomplete name. Part of the name right of the asterisk (including the asterisk) will be automatically generated by the queue manager. Only one asterisk character may be specified. If specified, it must be the last character in the value and its position must be between 1 and 33 (inclusive). Default: *
queue_nameName of the source queue from which to receive messages. In publish/subscribe mode this is used as the subscriber queue. Note: if dynamic queue options are specified, this value is the name of the model queue to use as template for creating the dynamic queue
transaction.record_countNumber of records per transaction. The value 0 means all available records. Default: 0
pub_sub.registrationSwitch that controls whether to register with the broker when the job starts. Default: false
pub_sub.publish.registrationRegistration options to specify when publishing messages. Note: Option [Anonymous] is ignored in MQRFH2 service mode. The remaining registration options are used in MQRFH2 service mode as publication options. Values: [anonymous, correlation_id_as_identity, local]
pub_sub.pub_sub_reply_to_queueName of the queue to which the broker should send replies for the command messages. Note: If dynamic reply queue usage is specified, the value specified here is used as the model queue name
other_queue_settings.dynamic_queue.close_optionsClose options to use when closing dynamic reply queue in request/reply mode of operation. Values: [delete, none, purge_and_delete]. Default: none
header_fields_setter.reply_to_qReply to queue value to set for destination messages. In request/reply mode, if dynamic reply queue is used, this value is the name of the model queue to use for the dynamic reply queue
header_fields_setter.reply_to_q_mgrReply to queue manager value to set for destination messages
header_fields_setter.reportReport values to set for destination messages. Values: [confirm_on_arrival, confirm_on_arrival_with_data, confirm_on_arrival_with_full_data, confirm_on_delivery, confirm_on_delivery_with_data, confirm_on_delivery_with_full_data, discard_message, exception, exception_with_data, exception_with_full_data, expiration, expiration_with_data, expiration_with_full_data, negative_action_notification, pass_correlation_identifier, pass_message_identifier, positive_action_notification]
message_options.row_buffer_countNumber of rows that the connector buffers before sending a message with a payload comprised of a concatenation of the buffered rows. Message header fields and message format headers (if any) from the first buffered row are used for the composite destination message. Note that if a Record count value (under Transaction settings) is specified, that value must be a multiple of the Row buffer count value.. Default: 1
message_options.create_segmented_message.segment_size *Size of each segment in bytes. The last segment to be created may have smaller size than the specified value. Default: 1024
message_options.create_segmented_messageSwitch that controls whether to separate data for destination message into segments and to send those separate segments rather than a single message to the destination queue. Note that if an error occurs while sending some of the segments, the whole input message will be sent to the error queue (if defined) or to the reject link (if defined), rather than the individual segments.. Default: false
pub_sub.service_typeRules and formatting header version to use for command messages. Values: [mqrfh, mqrfh2]. Default: mqrfh
header_fields_setterSwitch that controls whether to override specified message header fields for destination messages. Default: false
message_options.set_message_id_column_valueSwitch that controls whether the message ID should be set to the value of the column with WSMQ.MSGID data element.. Default: false
pub_sub.publish.msg_seq_number.start_valueThe initial message sequence number used for the first published message and regularly incremented for the subsequent published messages. Default: 1
pub_sub.publish.publication_format.system_valueFormat system value to set for the publication payload. Values: [mqadmin, mqchcom, mqcics, mqcmd1, mqcmd2, mqdead, mqevent, mqhdist, mqhmde, mqhref, mqhrf, mqhrf2, mqhwih, mqims, mqimsvs, mqnone, mqpcf, mqstr, mqtrig, mqxmit]. Default: mqstr
header_fields_setter.feedback.system_valueFeedback and MQRC (reason code) system value to set for destination messages. Values: [confirm_on_arrival, confirm_on_delivery, expiration, message_too_big_for_queue_mqrc, message_too_big_for_queue_manager_mqrc, negative_action_notification, none, not_authorized_mqrc, persistent_not_allowed_mqrc, positive_action_notification, put_inhibited_mqrc, queue_full_mqrc, queue_space_not_available_mqrc, quit]. Default: none
header_fields_setter.format.system_valueFormat system value to set for destination messages. Values: [mqadmin, mqchcom, mqcics, mqcmd1, mqcmd2, mqdead, mqevent, mqhdist, mqhmde, mqhref, mqhrf, mqhrf2, mqhwih, mqims, mqimsvs, mqnone, mqpcf, mqstr, mqtrig, mqxmit]. Default: mqstr
header_fields_setter.msg_type.system_valueMessage type system value to set for destination messages. Values: [datagram, reply, report, request]. Default: datagram
header_fields_setter.put_appl_type.system_valuePut application type system value for destination messages. Values: [aix,_unix, broker, channelinitiator, cics, cicsbridge, cicsvse, dos, dqm, guardian,_nsk, ims, imsbridge, java, mvs,_os390,_zos, nocontext, notesagent, os2, os400, qmgr, unknown, user, vms, vos, windows, windowsnt, xcf]. Default: nocontext
pub_sub.publish.timestampSwitch that controls whether to include timestamps in the published messages. Default: false
pub_sub.deregistration.deregistration_topicTopic(s) for deregistration
pub_sub.publish.publish_topicTopic of the publication message
pub_sub.registration.registration_topic *Topic(s) for registration
error_queue.tranmission_queue_nameName of the transmission queue to use when the error queue is a remote queue. If the value is not specified, the default transmission queue is used.
other_queue_settings.transmission_queue_nameName of the transmission queue to use when the destination queue is a remote queue. If the value is not specified, the default transmission queue is used.
header_fields_setter.user_identifierUser identifier value to set for destination messages
other_queue_settings.alternate_security_id.valueThe value for the alternate security identifier.
header_fields_setter.accounting_token.valueAccounting token value for destination messages
header_fields_setter.correl_id.valueCorrelation identifier value for destination messages
header_fields_setter.group_id.valueGroup identifier value for destination messages
header_fields_setter.msg_id.valueMessage identifier value for destination messages




IBM Match 360


Description: IBM Match 360 with Watson is a master data management service that helps you resolve data across different sources and gain a clearer view of each customer, supplier, or partner.
Data source type ID: 99265578-2e54-4b6b-baea-3058fc2ecc96
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
api_key *API key from the user account
crnCloud Resource Name. To find the CRN, go to the [Resource list] at https://cloud.ibm.com/resources. Expand [Services and software]. Select the [IBM Match 360 with Watson] instance, and click the Location column.

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
model_type_name *The IBM Match 360 data model type, such as records or relationships. Values: [records, relationships]. Default: records
record_sub_type_nameThe IBM Match 360 record subtype, such as customers, patients, or households. This value only applies when the data category is records
model_sub_type_name *The IBM Match 360 data model subtype. This value can be a record type (such as such as person or organization) or a relationship type (such as spouse or doctor-patient)
row_limitThe maximum number of rows to return

Interaction properties (when used as a target)

Name Type Description
model_type_name *The IBM Match 360 data model type, such as records or relationships. Values: [records, relationships]. Default: records
record_sub_type_nameThe IBM Match 360 record subtype, such as customers, patients, or households. This value only applies when the data category is records
model_sub_type_name *The IBM Match 360 data model subtype. This value can be a record type (such as such as person or organization) or a relationship type (such as spouse or doctor-patient)




IBM Netezza Performance Server


Description: IBM Netezza Performance Server is a platform for high performance data warehousing and analytics.
Data source type ID: c2a82a72-0711-4376-a468-4e9951cabf22
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [none, random]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




IBM Netezza Performance Server (optimized)


Description: Connect to a Netezza Performance Server database for the DataStage Netezza Performance Server (optimized) connector. The IBM Netezza Performance Server (optimized) connector has more features for DataStage.
Data source type ID: 63e2d853-e650-3b59-91a5-95e7bf725b9b
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
database *Specify the name of the database to connect to.
use_separate_connection_for_twt.database *The name of the database for ETs and TWT.
hostname *A database hostname to connect to.
password *Specify the password to use to connect to the database.
use_separate_connection_for_twt.passwordPassword for authentication purposes.
port *Port that database process is listening on.. Default: 5480
use_separate_connection_for_twtUse a separate connection for creating/dropping/accessing External tables (ETs) and the Temporary Work Table.. Default: false
use_separate_connection_for_twt.usernameName of the user for authentication purposes.
username *Specify the user name to use to connect to the database.

Interaction properties (when used as a source)

Name Type Description
before_after_sql.after_sqlOne or more statements to be executed after the connector finished processing all input rows. Multiple statements are separated by semi-colon. Executed once from the conductor node.
before_after_sql.after_sql_nodeOne or more statements to be executed after the connector finished processing all input rows. Multiple statements are separated by semi-colon. Executed once in each processing node. (Parallel canvas only)
session.array_sizeEnter a number that represents the number of records to process in read and write operations on the database. Enter 0 to process all records in a single array. Enter 1 to process one record at a time.. Default: 2000
before_after_sql.after_sql.fail_on_error.atomic_modeIf the atomic mode is yes, execute all statements in one transaction. Otherwise, execute each statement as separate transaction.. Default: true
before_after_sql.after_sql_node.fail_on_error.atomic_modeIf the atomic mode is yes, execute all statements in one transaction. Otherwise, execute each statement as separate transaction.. Default: true
before_after_sql.before_sql.fail_on_error.atomic_modeIf the atomic mode is yes, execute all statements in one transaction. Otherwise, execute each statement as separate transaction.. Default: true
before_after_sql.before_sql_node.fail_on_error.atomic_modeIf the atomic mode is yes, execute all statements in one transaction. Otherwise, execute each statement as separate transaction.. Default: true
before_after_sql.before_sqlOne or more statements to be executed before the connector starts processing any input rows. Multiple statements are separated by semi-colon. Executed once from the conductor node.
before_after_sql.before_sql_nodeOne or more statements to be executed before the connector starts processing any input rows Multiple statements are separated by semi-colon. Executed once in each processing node. (Parallel canvas only)
before_after_sqlSetting it to Yes, enables child properties for specifying Before and After SQL statements.. Default: false
session.unload_options.directory_for_named_pipeSpecifies the directory for the named pipe on Unix. It is ignored on Windows. If it is left blank, the connector will use the value of the environment variable TMPDIR. If TMPDIR is not defined, it will default to /tmp.
enable_case_sensitive_i_dsIf set to Yes, table and column names will be assumed to be case sensitive.. Default: true
sql.enable_partitioned_readsIf set to Yes, the connector will allow reading in parallel. The level of parallelism will be determined by the number of nodes in the APT configuration file. If set to No, the connector will force sequential execution.. Default: false
before_after_sql.after_sql.fail_on_errorIf set to Yes, the job will be aborted if the statement fails. If set to No, the statement errors will be ignored.. Default: true
before_after_sql.after_sql_node.fail_on_errorIf set to Yes, the job will be aborted if the statement fails. If set to No, the statement errors will be ignored.. Default: true
before_after_sql.before_sql.fail_on_errorIf set to Yes, the job will be aborted if the statement fails. If set to No, the statement errors will be ignored.. Default: true
before_after_sql.before_sql_node.fail_on_errorIf set to Yes, the job will be aborted if the statement fails. If set to No, the statement errors will be ignored.. Default: true
generate_sqlIndicates whether the connector should generate a SELECT statement or use the provided SQL statement.. Default: true
limit_rows.limitEnter the maximum number of rows that will be returned by the connector.. Default: 1000
limit_rowsSelect Yes to limit the number of rows that are returned by the connector.. Default: false
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
transaction.mark_end_of_waveIf set to Yes, the connector emits an end-of-wave marker after the specified number of rows (Record count) is read from the Netezza server. If set to No, the connector will not emit end-of-wave markers.. Default: false
before_after_sql.after_sql.fail_on_error.log_level_for_after_sqlThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
before_after_sql.after_sql_node.fail_on_error.log_level_for_after_sql_nodeThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
before_after_sql.before_sql.fail_on_error.log_level_for_before_sqlThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
before_after_sql.before_sql_node.fail_on_error.log_level_for_before_sql_nodeThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
session.schema_reconciliation.mismatch_reporting_actionThe type of message that will be logged if one or more columns unmatched or mismatched.. Values: [info, none, warning]. Default: warning
session.schema_reconciliation.mismatch_reporting_action_sourceThe type of message that will be logged if one or more columns unmatched or mismatched.. Values: [info, none, warning]. Default: warning
before_after_sql.after_sql.read_from_fileIf set to Yes, the After SQL property specifies a file name/path containing the SQL statement. Otherwise it specifies the actual SQL statement.. Default: false
before_after_sql.after_sql_node.read_from_fileIf set to Yes, the After SQL (Node) property specifies a file name/path containing the SQL statement. Otherwise it specifies the actual SQL statement.. Default: false
before_after_sql.before_sql.read_from_fileIf set to Yes, the Before SQL property specifies a file name/path containing the SQL statement. Otherwise it specifies the actual SQL statement.. Default: false
before_after_sql.before_sql_node.read_from_fileIf set to Yes, the Before SQL (Node) property specifies a file name/path containing the SQL statement. Otherwise it specifies the actual SQL statement.. Default: false
sql.select_statement.read_user_defined_sql_from_fileIf set to Yes, the User-defined SQL property specifies a file name/path containing the SQL statement.. Default: false
transaction.record_countIndicates the number of records (rows) in a single wave.. Default: 2000
sql.select_statement *Enter a SELECT statement. The statement is used to read rows from the database.
table_name *The name of the target table. This table name will be used in the generated SQL statement(s). Never enter the name with quotes.
session.schema_reconciliation.type_mismatch_actionAction to take upon detecting a type mismatch. The value Fail will cause the job to abort.. Values: [drop, fail, keep]. Default: drop
session.schema_reconciliation.type_mismatch_action_sourceAction to take upon detecting a type mismatch. The value Fail will cause the job to abort.. Values: [drop, fail, keep]. Default: drop
session.schema_reconciliation.unmatched_link_column_action_request_input_linkAction to take when an input link column does not match any columns in the table.. Values: [drop, fail]. Default: drop
session.schema_reconciliation.unmatched_link_column_action_sourceAction to take when an input link column does not match any columns in the table.. Values: [drop, fail]. Default: drop
session.schema_reconciliation.unmatched_table_or_query_column_action_requestAction to take when a table or a query column does not match any output link columns. Values: [fail, ignore]. Default: ignore
session.schema_reconciliation.unmatched_table_or_query_column_action_sourceAction to take when a table or a query column does not match any output link columns. Values: [fail, ignore]. Default: ignore

Interaction properties (when used as a target)

Name Type Description
sql.action_column *The name of the char (1) column that identifies one of the action the row should participate in.
before_after_sql.after_sqlOne or more statements to be executed after the connector finished processing all input rows. Multiple statements are separated by semi-colon. Executed once from the conductor node.
before_after_sql.after_sql_nodeOne or more statements to be executed after the connector finished processing all input rows. Multiple statements are separated by semi-colon. Executed once in each processing node. (Parallel canvas only)
before_after_sql.after_sql.fail_on_error.atomic_modeIf the atomic mode is yes, execute all statements in one transaction. Otherwise, execute each statement as separate transaction.. Default: true
before_after_sql.after_sql_node.fail_on_error.atomic_modeIf the atomic mode is yes, execute all statements in one transaction. Otherwise, execute each statement as separate transaction.. Default: true
before_after_sql.before_sql.fail_on_error.atomic_modeIf the atomic mode is yes, execute all statements in one transaction. Otherwise, execute each statement as separate transaction.. Default: true
before_after_sql.before_sql_node.fail_on_error.atomic_modeIf the atomic mode is yes, execute all statements in one transaction. Otherwise, execute each statement as separate transaction.. Default: true
sql.atomic_modeWhen set to Yes, all write mode statements will be executed in one transaction. If set to No, each statement will be executed in a separate transaction.. Default: true
before_after_sql.before_sqlOne or more statements to be executed before the connector starts processing any input rows. Multiple statements are separated by semi-colon. Executed once from the conductor node.
before_after_sql.before_sql_nodeOne or more statements to be executed before the connector starts processing any input rows Multiple statements are separated by semi-colon. Executed once in each processing node. (Parallel canvas only)
before_after_sqlSetting it to Yes, enables child properties for specifying Before and After SQL statements.. Default: false
sql.check_duplicate_rowsWhen set to Yes, the connector detects duplicate rows. When set to No (Default), the connector does not detect duplicate rows.. Default: false
sql.enable_record_ordering.order_key.column_nameThe name of the column representing the ordering key
session.temporary_work_table.create_statementSpecifies a user-defined CREATE TABLE statement.
table_action.generate_create_statement.create_statement *Specifies a user-defined CREATE TABLE statement.
sql.direct_insertIf set to Yes, the connector insert directly into the target table. In this mode, when running with multiple processing nodes it is possible to have partially committed data if one or more of the processing nodes encounters an error. If set No, the connector inserts into the temporary work table (TWT) first and then from TWT into the target. In this mode the data will either be completely committed or completely rolled back guarantying consistency.. Default: false
session.load_options.directory_for_log_filesSpecifies the directory for the nzlog and nzbad files. If it is left blank, the connector will use the value of the environment variable TMPDIR. If TMPDIR is not defined, it will default to /tmp on Unix and to system temporary directory on Windows.
session.load_options.directory_for_named_pipeSpecifies the directory for the named pipe on Unix. It is ignored on Windows. If it is left blank, the connector will use the value of the environment variable TMPDIR. If TMPDIR is not defined, it will default to /tmp.
table_action.generate_create_statement.distribution_keyIf set to Automatic, the Netezza server will choose the key. There is no guaranty which columns will be used and the feature can vary between Netezza software releases. If set to Random, rows will be sent to processing nodes in a random fashion (key-less). If set to User-defined, the columns listed in Key columns below will be used as a distribution key. This property only applies to a generated CREATE TABLE statement (run-time or design-time) and has no effect if the statement is entered manually. Values: [automatic, random, user-defined]. Default: random
table_action.generate_drop_statement.drop_statement *Specifies a user-defined DROP TABLE statement.
session.temporary_work_table.drop_tableIf set to Yes, the connector will drop the temporary work table.. Default: true
sql.check_duplicate_rows.duplicate_row_actionWhen set to Filter (Default), the connector filters out duplicate rows. When set to Fail, the job fails if any duplicate rows are found without making changes to the target table.. Values: [fail, filter]. Default: filter
enable_case_sensitive_i_dsIf set to Yes, table and column names will be assumed to be case sensitive.. Default: false
sql.enable_record_orderingSetting it to Yes, enables record ordering and enables child properties for specifying order columns. Default: false
session.temporary_work_table.enable_merge_joinSets the ENABLE_MERGEJOIN configuration parameter. If set to Yes, the connector will enable Netezza query planner's use of merge-join plan types. If set to No, the planner will choose one of other algorithms, including hash-join as a highly optimized one. If set to Database default, the property will not be set before query execution, using the current database setting.. Values: [database_default, no, yes]. Default: database_default
before_after_sql.after_sql.fail_on_errorIf set to Yes, the job will be aborted if the statement fails. If set to No, the statement errors will be ignored.. Default: true
before_after_sql.after_sql_node.fail_on_errorIf set to Yes, the job will be aborted if the statement fails. If set to No, the statement errors will be ignored.. Default: true
before_after_sql.before_sql.fail_on_errorIf set to Yes, the job will be aborted if the statement fails. If set to No, the statement errors will be ignored.. Default: true
before_after_sql.before_sql_node.fail_on_errorIf set to Yes, the job will be aborted if the statement fails. If set to No, the statement errors will be ignored.. Default: true
table_action.generate_create_statement.fail_on_errorAbort the job if there is an error executing a command. Default: true
table_action.generate_drop_statement.fail_on_errorAbort the job if there is an error executing a command. Default: false
table_action.generate_truncate_statement.fail_on_errorAbort the job if there is an error executing a command. Default: true
table_action.generate_create_statementIf set to Yes, the CREATE TABLE statement will be generated by the connector at runtime based on the columns in the input link and the table name provided in the Table name property. Default: true
table_action.generate_drop_statementIf set the Yes, the DROP TABLE statement will be generated at runtime. If set to No, the user needs to provide a custom statement in the Drop statement property.. Default: true
session.load_options.generate_statisticsIf set to Yes, the connector will generate statistics. If set to No, no statistics will be generated.. Default: false
session.load_options.generate_statistics.generate_statistics_modeChoose whether to generate statistics for a table or the whole database. Values: [database, table]. Default: table
session.load_options.generate_statistics.generate_statistics_columnsSelect the columns to generate the statistics for. If no columns are selected (default state), all columns of the target table will be included.
table_action.generate_truncate_statementIf set the Yes, the TRUNCATE TABLE statement will be generated at runtime.. Default: true
sql.key_columns *A comma-separated list of key column names.
table_action.generate_create_statement.distribution_key.key_columns *Comma-separated list of column names that should comprise the distribution key.
session.load_options.max_reject_countThe number of rejected records that are logged before the job aborts.. Default: 1
before_after_sql.after_sql.fail_on_error.log_level_for_after_sqlThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
before_after_sql.after_sql_node.fail_on_error.log_level_for_after_sql_nodeThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
before_after_sql.before_sql.fail_on_error.log_level_for_before_sqlThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
before_after_sql.before_sql_node.fail_on_error.log_level_for_before_sql_nodeThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
table_action.generate_create_statement.fail_on_error.log_level_for_create_statementThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
table_action.generate_drop_statement.fail_on_error.log_level_for_drop_statementThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
table_action.generate_truncate_statement.fail_on_error.log_level_for_truncate_statementThe type of message that will be logged if a SQL statement fails.. Values: [info, none, warning]. Default: warning
session.schema_reconciliation.mismatch_reporting_actionThe type of message that will be logged if one or more columns unmatched or mismatched.. Values: [info, none, warning]. Default: warning
session.load_options.other_optionsAdditional options to be passed to the external table create statement.
table_action.generate_create_statement.read_create_statement_from_fileIf set to Yes, the Create statement property specifies a file name/path containing the SQL statement.. Default: false
table_action.generate_drop_statement.read_drop_statement_from_fileIf set to Yes, the Drop statement property specifies a file name/path containing the SQL statement. Otherwise it specifies the actual SQL statement.. Default: false
before_after_sql.after_sql.read_from_fileIf set to Yes, the After SQL property specifies a file name/path containing the SQL statement. Otherwise it specifies the actual SQL statement.. Default: false
before_after_sql.after_sql_node.read_from_fileIf set to Yes, the After SQL (Node) property specifies a file name/path containing the SQL statement. Otherwise it specifies the actual SQL statement.. Default: false
before_after_sql.before_sql.read_from_fileIf set to Yes, the Before SQL property specifies a file name/path containing the SQL statement. Otherwise it specifies the actual SQL statement.. Default: false
before_after_sql.before_sql_node.read_from_fileIf set to Yes, the Before SQL (Node) property specifies a file name/path containing the SQL statement. Otherwise it specifies the actual SQL statement.. Default: false
table_action.generate_truncate_statement.read_truncate_statement_from_fileIf set to Yes, the Truncate statement property specifies a file name/path containing the SQL statement. Otherwise it specifies the actual SQL statement.. Default: false
sql.user_defined_sql.read_user_defined_sql_from_fileIf set to Yes, the User-defined SQL property specifies a file name/path containing the SQL statement.. Default: false
table_action *Select the action to perform before writing data to the table.. Values: [append, create, replace, truncate]. Default: append
session.temporary_work_table.table_name *The name of the temporary work table.
table_name *The name of the target table. This table name will be used in the generated SQL statement(s). Never enter the name with quotes.
session.temporary_work_tableIf set to Automatic, the connector will automatically create the temporary work table using an internally generated name. If set to User-defined, the connector will automatically create the temporary work table using the specified table name or create table statement.. Values: [automatic, existing, user-defined]. Default: automatic
truncate_column_namesIf Set to Yes, the names of the input link columns will be truncated.. Default: false
truncate_column_names.truncate_length *The maximum length in characters of column names after truncation.. Default: 128
table_action.generate_truncate_statement.truncate_statement *Specifies a user-defined TRUNCATE TABLE statement.
session.temporary_work_table.truncate_tableIf set to Yes, the temporary work table is truncated before any data is written to it.. Default: false
session.schema_reconciliation.type_mismatch_actionAction to take upon detecting a type mismatch. The value Fail will cause the job to abort.. Values: [drop, fail, keep]. Default: drop
sql.use_unique_key_column.unique_key_column *The name of the unique key column.
session.schema_reconciliation.unmatched_link_column_actionAction to take when an input link column does not match any columns in the table.. Values: [drop, fail, keep]. Default: drop
session.schema_reconciliation.unmatched_table_column_actionAction to take when a table column does not match any input link columns. If an existing temporary work table is provided, its columns are checked against the input link columns. Otherwise target table columns are checked. Values: [fail, ignore_all, ignore_nullable]. Default: ignore_nullable
sql.update_columnsA comma-separated list of column names whose value will be updated by the statement(s).
sql.use_unique_key_columnWhen set to Yes, the connector will generate an update statement that uses the unique key specified below. When set to No, the connector will generate a simpler update statement that does not use a unique column.. Default: false
sql.user_defined_sql *These SQL statements in general should update the target table using the data in the temporary work table.
write_modeType of the SQL statement (or statements) to be executed.. Values: [action_column, delete, delete_then_insert, insert, update, update_then_insert, user-defined_sql]. Default: insert




IBM Planning Analytics


Description: IBM Planning Analytics is an enterprise performance management database that stores data in in-memory multidimensional OLAP cubes. IBM Planning Analytics was formerly named TM1.
Data source type ID: c8f3d379-78b2-4bad-969d-2e928277377e
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
auth_typeThe type of authentication to be used to access the TM1 server API. Values: [bearer, cam_credentials, basic]
gateway_urlThe URL of the gateway for Planning Analytics service
namespace *The namespace to use for connecting to the TM1 sever API
password *The password associated with the username for accessing the data source
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
service_root *The URL used to access the TM1 server API implementing the OData protocol
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
use_creation_orderWhether to read cube dimensions using creation order. Default: false
cube_name *The cube to be processed
mdx_statementThe MDX statement to select a view
row_limitThe maximum number of rows to return
view_name *The view to be processed
view_group *The group that the view belongs to

Interaction properties (when used as a target)

Name Type Description
cube_name *The cube to be processed
mdx_statementThe MDX statement to select a view
view_nameThe view to be processed
view_groupThe group that the view belongs to
write_to_consolidationIf writing to consolidation




IBM Product Master


Description: Use the IBM Product Master connector to read or write product information to or from IBM Product Master.
Data source type ID: fa2785b6-f327-4515-9c9d-af1b2afd5d71
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
app_urlProduct Master application cluster URL on Cloud Pak for Data
company *Company name for IBM Product Master
password *Password for IBM Product Master
usernameUsername for IBM Product Master

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
catalog_name *Name of the Catalog
category_nameIdentifier of the category associated with product
hierarchy_nameName of the primary or secondary hierarchy
row_limitThe maximum number of rows to return

Interaction properties (when used as a target)

Name Type Description
catalog_name *Name of the Catalog
category_nameIdentifier of the category associated with product
hierarchy_nameName of the primary or secondary hierarchy




IBM Watson Query


Description: IBM Watson Query, formerly IBM Data Virtualization, virtualizes data across multiple data sources and turns all this data into one logical data view. A Watson Query connection is created automatically when a virtual object is published to a catalog or assigned to a project.
Data source type ID: 49686982-255f-423a-a5de-d825bfc0abe3
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
api_key *An application programming interface key that identifies the calling application or user
auth_method. Values: [apikey, username_password]
database *The name of the database
host *The hostname or IP address of the database
instance_id *36 characters instance ID of the Watson Query database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
inherit_access_tokenUse your Cloud Pak for Data credentials to authenticate to the data source. Default: false
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description




Looker


Description: Looker is a business intelligence software and big data analytics platform that helps you explore, analyze, and share real-time business analytics.
Data source type ID: 69857d6b-2be8-4a59-8a70-723405f09708
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
client_id *The client ID for authorizing access to Looker
client_secret *The password associated with the client ID for authorizing access to Looker
host *The hostname or IP address of the Looker server
portThe port of the Looker server. Default: 19999

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
file_formatThe format of the file. Values: [csv, delimited, excel, json]. Default: csv
file_name *The name of the file to read
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
sheet_nameThe name of the Excel worksheet to read from

Interaction properties (when used as a target)

Name Type Description




MariaDB


Description: MariaDB is an open-source relational database. You can use the MariaDB connector to connect to either a MariaDB server or to a Microsoft Azure Database for MariaDB service in the cloud.
Data source type ID: f3ee04c2-7c3b-4534-b300-eb6ef701646d
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [none, random]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [delete, delete_insert, insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Microsoft Azure Blob Storage


Description: Microsoft Azure Blob Storage is used for storing large amounts of unstructured data in the cloud.
Data source type ID: 9a22e0af-8d19-4c4e-9aea-1d733e81315b
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
connection_string *Connection string from the storage account's Access keys page on the Microsoft Azure portal
containerThe name of the container that contains the files to access

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
containerThe name of the container that contains the files to read
container_sourceSpecify the container
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
delimited_syntax.field_formats.date_formatSpecify a string that defines the format for fields that have the Date data type.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
delimited_syntax.field_formats.decimal_formatSpecify a string that defines the format for fields that have the Decimal or Numeric data type.
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
delimited_syntax.record_def.record_def_sourceEnter a delimited string that specifies the names and data types and length of each fields. Use the format name:data_type[length], and separate each field with the delimiter specified as the >B<Field delimiter>/B< property. If the record definition is in a delimited string file or Osh schema file, specify the full path of the file.
display_value_labelsDisplay the value labels
delimited_syntax.encodingSpecify the encoding of the files to read or write, for example, UTF-8.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
delimited_syntax.escapeSpecify the character to use to escape field and row delimiters. If an escape character exists in the data, the escape character is also escaped. Because escape characters require additional processing, do not specify a value for this property if you do not need to include escape characters in the data.
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
delimited_syntax.field_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default: ,
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_name_source *Specify the file name to read from Azure
_file_formatSpecify the format of the files to read or write.. Values: [comma-separated_value_csv, delimited]. Default: delimited
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_name *The name of the file to read
filename_columnSpecify the name of the column to write the source file name to.
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
delimited_syntax.headerSelect Yes if the first row of the file contains field headers and is not part of the data. If you select Yes, when the connector writes data, the field names will be the first row of the output. If runtime column propagation is enabled, metadata can be obtained from the first row of the file.. Default: false
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
recurseSpecify whether to read files that are in child folders of the prefix that is specified for the File name property.. Default: true
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
delimited_syntax.null_valueSpecify the character or string that represents null values in the data. For a source stage, input data that has the value that you specify is set to null on the output link. For a target stage, in the output file that is written to the file system, null values are represented by the value that is specified for this property. To specify that an empty string represents a null value, specify "" (two double quotation marks).
null_valueThe value that represents null (a missing value) in the file, for example, NULL
delimited_syntax.quotes. Values: [double, none, single]. Default: none
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
_read_mode *Select the Read mode. Values: [list_containers/fileshares, list_files, read_multiple_files, read_single_file]. Default: read_single_file
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
delimited_syntax.record_defSelect whether the record definition is provided to the connector from the source file, a delimited string, a file that contains a delimited string, or a schema file. When runtime column propagation is enabled, this metadata provides the column definitions. If a schema file is provided, the schema file overrides the values of formatting properties in the stage and the column definitions that are specified on the Columns page of the output link.. Values: [delimited_string, delimited_string_in_a_file, file_header, none, schema_file]. Default: none
delimited_syntax.record_limitSpecify the maximum number of records to read from the file per node. If a value is not specified for this property, the entire file is read.
reject_modeSpecify what the connector does when a record that contains invalid data is found in the source file. Select Continue to read the rest of the file, Fail to stop the job with an error message, or Reject to send the rejected data to a reject link.. Values: [continue, fail, reject]. Default: continue
delimited_syntax.row_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default:
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
delimited_syntax.field_formats.time_formatSpecify a string that defines the format for fields that have the Time data type.
time_formatThe format of time values, for example, HH:mm:ss[.f]
delimited_syntax.field_formats.timestamp_formatSpecify a string that defines the format for fields that have the Timestamp data type.
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
wave_handling.append_uidUse this property to choose if a unique identifier is to be appended to the file name. When the value of this property is set to yes, then the file name gets appended with the unique identifier, and a new file would be written for every wave of data that is streamed into the stage. When the value of this property is set to No, then the file would be overwritten on every wave.. Default: false
blob_type *Type of blob to write. Values: [append, block, page]. Default: block
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
_containerSpecify the container
containerThe name of the container that contains the files to write to
parallel_write.temp_containerSpecify the temporary container that can be used to create temporary files while performing parallel write. When no value is specified, the container specified in "Container" option will be used.
_create_containerSelect this property if you want to create container if it doesn't exist.. Default: false
create_containerCreate the container that contains the files to write to. Default: false
parallel_write.create_temp_containerUse this option if you want to create the container to store temporary files.. Default: false
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
delimited_syntax.field_formats.date_formatSpecify a string that defines the format for fields that have the Date data type.
decimal_formatThe format of decimal values, for example, #,###.##
delimited_syntax.field_formats.decimal_formatSpecify a string that defines the format for fields that have the Decimal or Numeric data type.
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
parallel_writeUse this option to perform a parallel write in Blob Storage.. Default: false
delimited_syntax.encodingSpecify the encoding of the files to read or write, for example, UTF-8.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
delimited_syntax.escapeSpecify the character to use to escape field and row delimiters. If an escape character exists in the data, the escape character is also escaped. Because escape characters require additional processing, do not specify a value for this property if you do not need to include escape characters in the data.
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
delimited_syntax.field_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default: ,
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
_file_name *Specify the file name
_file_formatSpecify the format of the files to read or write.. Values: [comma-separated_value_csv, delimited]. Default: delimited
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_name *The name of the file to write to or delete
wave_handling.file_size_thresholdSpecify the threshold for the file size in megabytes. Processing nodes will start a new file each time the size exceeds the value specified in the threshold.. Default: 1
first_line_headerIndicates whether the row where reading starts is the header. Default: false
delimited_syntax.headerSelect Yes if the first row of the file contains field headers and is not part of the data. If you select Yes, when the connector writes data, the field names will be the first row of the output. If runtime column propagation is enabled, metadata can be obtained from the first row of the file.. Default: false
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
file_existsSpecify what the connector does when it tries to write a file that already exists. Select Overwrite file to overwrite a file if it already exists, Do not overwrite file to not overwrite the file and stop the job, or Fail to stop the job with an error message.. Values: [do_not_overwrite_file, fail, overwrite_file]. Default: overwrite_file
delimited_syntax.encoding.output_bomSpecify whether to include a byte order mark in the file when the file encoding is a Unicode encoding such as UTF-8, UTF-16, or UTF-32.. Default: false
delimited_syntax.header.include_typesSelect Yes to append the data type to each field name that the connector writes in the first row of the output.. Default: false
include_typesInclude data types in the first line of the file. Default: false
names_as_labelsSet column labels to the value of the column name
delimited_syntax.null_valueSpecify the character or string that represents null values in the data. For a source stage, input data that has the value that you specify is set to null on the output link. For a target stage, in the output file that is written to the file system, null values are represented by the value that is specified for this property. To specify that an empty string represents a null value, specify "" (two double quotation marks).
null_valueThe value that represents null (a missing value) in the file, for example, NULL
partitionedWrite the file as multiple partitions. Default: false
delimited_syntax.quotes. Values: [double, none, single]. Default: none
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
delimited_syntax.row_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default:
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
delimited_syntax.field_formats.time_formatSpecify a string that defines the format for fields that have the Time data type.
time_formatThe format of time values, for example, HH:mm:ss[.f]
delimited_syntax.field_formats.timestamp_formatSpecify a string that defines the format for fields that have the Timestamp data type.
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
sheet_nameThe name of the Excel worksheet to write to
_write_mode *Select the Write mode. Values: [delete, write]. Default: write
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




Microsoft Azure Cosmos DB


Description: Microsoft Azure Cosmos DB is a fully managed NoSQL database service.
Data source type ID: 0c431748-2572-11ea-978f-2e728ce88125
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
host *The Azure Cosmos DB database that stores the read-write keys
master_key *The Azure Cosmos DB primary read-write key
portThe Azure Cosmos DB database port number. Default: 443

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
collection *The collection to connect to
database *The database to connect to
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data

Interaction properties (when used as a target)

Name Type Description
collection *The collection to connect to
create_collectionCreate the collection to connect to
database *The database to connect to
input_formatThe format of the source data. Values: [json, relational]. Default: relational
offer_throughputThe throughput allocated for bulk operations out of the collection's total throughput
write_modeWhether to write to, or delete, the target. Values: [delete, write]. Default: write




Microsoft Azure Data Lake Store


Description: Microsoft Azure Data Lake Store (ADLS) is a scalable data storage and analytics service. The Microsoft Azure Data Lake Store connector supports access to both Gen1 and Gen2 Azure Data Lake Storage repositories.
Data source type ID: 6863060d-97c4-4653-abbe-958bde533f8c
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
client_id *The client ID for authorizing access to Microsoft Azure Data Lake Store
client_secret *The authentication key associated with the client ID for authorizing access to Microsoft Azure Data Lake Store
proxy_host *The server proxy host
proxy_port *The server proxy port
proxy_protocolThe proxy server protocol. Values: [http, https]
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
proxyUse server proxy. Default: false
tenant_id *The Azure Active Directory tenant ID
url *The WebHDFS URL for accessing HDFS

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
display_value_labelsDisplay the value labels
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_name *The name of the file to read
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
null_valueThe value that represents null (a missing value) in the file, for example, NULL
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
decimal_formatThe format of decimal values, for example, #,###.##
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_name *The name of the file to write to or delete
first_line_headerIndicates whether the row where reading starts is the header. Default: false
include_typesInclude data types in the first line of the file. Default: false
names_as_labelsSet column labels to the value of the column name
null_valueThe value that represents null (a missing value) in the file, for example, NULL
partitionedWrite the file as multiple partitions. Default: false
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
time_formatThe format of time values, for example, HH:mm:ss[.f]
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
sheet_nameThe name of the Excel worksheet to write to
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




Microsoft Azure File Storage


Description: Microsoft Azure File Storage is a distributed, cloud-based file system.
Data source type ID: 2a7b4fa1-c770-4807-8871-a3c5def5aa2d
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
connection_string *Connection string from the storage account's Access keys page on the Microsoft Azure portal
containerThe name of the container that contains the files to access

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
rangeThe range of cells to retrieve from the Excel worksheet, for example, C1:F10
containerThe name of the container that contains the files to read
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
delimited_syntax.field_formats.date_formatSpecify a string that defines the format for fields that have the Date data type.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
decimal_formatThe format of decimal values, for example, #,###.##
delimited_syntax.field_formats.decimal_formatSpecify a string that defines the format for fields that have the Decimal or Numeric data type.
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
delimited_syntax.record_def.record_def_sourceEnter a delimited string that specifies the names and data types and length of each fields. Use the format name:data_type[length], and separate each field with the delimiter specified as the >B<Field delimiter>/B< property. If the record definition is in a delimited string file or Osh schema file, specify the full path of the file.
display_value_labelsDisplay the value labels
delimited_syntax.encodingSpecify the encoding of the files to read or write, for example, UTF-8.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
delimited_syntax.escapeSpecify the character to use to escape field and row delimiters. If an escape character exists in the data, the escape character is also escaped. Because escape characters require additional processing, do not specify a value for this property if you do not need to include escape characters in the data.
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
exclude_missing_valuesSet values that have been defined as missing values to null
delimited_syntax.field_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default: ,
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
xml_path_fieldsThe path that identifies the specified elements to retrieve from the root path of a XML document, for example, ../publisher
file_name_source *Specify the file name to read from Azure
file_share_source *Specify the File share
_file_formatSpecify the format of the files to read or write.. Values: [comma-separated_value_csv, delimited]. Default: delimited
file_formatThe format of the file. Values: [avro, csv, delimited, excel, json, orc, parquet, sas, sav, shp, xml]. Default: csv
file_name *The name of the file to read
filename_columnSpecify the name of the column to write the source file name to.
first_lineIndicates at which row start reading. Default: 0
first_line_headerIndicates whether the row where reading starts is the header. Default: false
delimited_syntax.headerSelect Yes if the first row of the file contains field headers and is not part of the data. If you select Yes, when the connector writes data, the field names will be the first row of the output. If runtime column propagation is enabled, metadata can be obtained from the first row of the file.. Default: false
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
infer_timestamp_as_dateInfer columns containing date and time data as date. Default: true
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_null_as_empty_stringTreat empty values in string type columns as empty strings instead of null. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
json_pathThe path that identifies the elements to retrieve from a JSON document, for example, $.book.publisher
labels_as_namesSet column names to the value of the column label
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
delimited_syntax.null_valueSpecify the character or string that represents null values in the data. For a source stage, input data that has the value that you specify is set to null on the output link. For a target stage, in the output file that is written to the file system, null values are represented by the value that is specified for this property. To specify that an empty string represents a null value, specify "" (two double quotation marks).
null_valueThe value that represents null (a missing value) in the file, for example, NULL
delimited_syntax.quotes. Values: [double, none, single]. Default: none
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
_read_mode *Select the Read mode. Values: [list_containers/fileshares, list_files, read_multiple_files, read_single_file]. Default: read_single_file
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
delimited_syntax.record_defSelect whether the record definition is provided to the connector from the source file, a delimited string, a file that contains a delimited string, or a schema file. When runtime column propagation is enabled, this metadata provides the column definitions. If a schema file is provided, the schema file overrides the values of formatting properties in the stage and the column definitions that are specified on the Columns page of the output link.. Values: [delimited_string, delimited_string_in_a_file, file_header, none, schema_file]. Default: none
delimited_syntax.record_limitSpecify the maximum number of records to read from the file per node. If a value is not specified for this property, the entire file is read.
reject_modeSpecify what the connector does when a record that contains invalid data is found in the source file. Select Continue to read the rest of the file, Fail to stop the job with an error message, or Reject to send the rejected data to a reject link.. Values: [continue, fail, reject]. Default: continue
delimited_syntax.row_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default:
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
xml_schemaThe schema that specified metadata information of elements, for example, data type, values, min, max
delimited_syntax.field_formats.time_formatSpecify a string that defines the format for fields that have the Time data type.
time_formatThe format of time values, for example, HH:mm:ss[.f]
delimited_syntax.field_formats.timestamp_formatSpecify a string that defines the format for fields that have the Timestamp data type.
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2
use_field_formatsFormat data using specified field formats
use_variable_formatsFormat data using specified variable formats.
sheet_nameThe name of the Excel worksheet to read from
xml_pathThe path that identifies the root elements to retrieve from a XML document, for example, /book/publisher

Interaction properties (when used as a target)

Name Type Description
wave_handling.append_uidUse this property to choose if a unique identifier is to be appended to the file name. When the value of this property is set to yes, then the file name gets appended with the unique identifier, and a new file would be written for every wave of data that is streamed into the stage. When the value of this property is set to No, then the file would be overwritten on every wave.. Default: false
codec_avroThe compression codec to use when writing. Values: [bzip2, deflate, null, snappy]
codec_csvThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_delimitedThe compression codec to use when writing. Values: [gzip, uncompressed]
codec_orcThe compression codec to use when writing. Values: [lz4, lzo, none, snappy, zlib]
codec_parquetThe compression codec to use when writing. Values: [gzip, uncompressed, snappy]
containerThe name of the container that contains the files to write to
create_file_shareSelect this property if you want to create file share if it doesn't exist.. Default: false
create_containerCreate the container that contains the files to write to. Default: false
date_formatThe format of date values, for example, yyyy-[M]M-[d]d
delimited_syntax.field_formats.date_formatSpecify a string that defines the format for fields that have the Date data type.
decimal_formatThe format of decimal values, for example, #,###.##
delimited_syntax.field_formats.decimal_formatSpecify a string that defines the format for fields that have the Decimal or Numeric data type.
decimal_format_grouping_separatorThe character used to group digits of similar significance. This property and decimal separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
decimal_format_decimal_separatorThe character used to separate the integer part from the fractional part of a number. This property and decimal grouping separator property must be unique. If you encounter error related to them not being unique when only one of them was provided then please provide the missing one explicitly.
delimited_syntax.encodingSpecify the encoding of the files to read or write, for example, UTF-8.
encodingThe appropriate character encoding for your data, for example, UTF-8. Default: utf-8
encryption_keyKey to decrypt sav file
delimited_syntax.escapeSpecify the character to use to escape field and row delimiters. If an escape character exists in the data, the escape character is also escaped. Because escape characters require additional processing, do not specify a value for this property if you do not need to include escape characters in the data.
escape_characterThe character that's used to escape other characters, for example, a backslash. Escaping is a string technique that identifies characters as being part of a string value.. Values: [, backslash, double_quote, none, single_quote]. Default: none
escape_character_value *The custom character that is used to escape other characters.
delimited_syntax.field_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default: ,
field_delimiterThe character that separates each value from the next value, for example, a comma. Values: [, colon, comma, tab]. Default: comma
field_delimiter_value *The custom character that separates each value from the next value
_file_name *Specify the file name
file_share *Specify the File share
_file_formatSpecify the format of the files to read or write.. Values: [comma-separated_value_csv, delimited]. Default: delimited
file_formatThe format of the file to write to. Values: [avro, csv, delimited, excel, json, orc, parquet, sav, xml]. Default: csv
file_name *The name of the file to write to or delete
wave_handling.file_size_thresholdSpecify the threshold for the file size in megabytes. Processing nodes will start a new file each time the size exceeds the value specified in the threshold.. Default: 1
first_line_headerIndicates whether the row where reading starts is the header. Default: false
delimited_syntax.headerSelect Yes if the first row of the file contains field headers and is not part of the data. If you select Yes, when the connector writes data, the field names will be the first row of the output. If runtime column propagation is enabled, metadata can be obtained from the first row of the file.. Default: false
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
file_existsSpecify what the connector does when it tries to write a file that already exists. Select Overwrite file to overwrite a file if it already exists, Do not overwrite file to not overwrite the file and stop the job, or Fail to stop the job with an error message.. Values: [do_not_overwrite_file, fail, overwrite_file]. Default: overwrite_file
delimited_syntax.encoding.output_bomSpecify whether to include a byte order mark in the file when the file encoding is a Unicode encoding such as UTF-8, UTF-16, or UTF-32.. Default: false
delimited_syntax.header.include_typesSelect Yes to append the data type to each field name that the connector writes in the first row of the output.. Default: false
include_typesInclude data types in the first line of the file. Default: false
names_as_labelsSet column labels to the value of the column name
delimited_syntax.null_valueSpecify the character or string that represents null values in the data. For a source stage, input data that has the value that you specify is set to null on the output link. For a target stage, in the output file that is written to the file system, null values are represented by the value that is specified for this property. To specify that an empty string represents a null value, specify "" (two double quotation marks).
null_valueThe value that represents null (a missing value) in the file, for example, NULL
partitionedWrite the file as multiple partitions. Default: false
delimited_syntax.quotes. Values: [double, none, single]. Default: none
quote_characterThe character that's used to enclose string values, for example, a double quotation mark. Values: [double_quote, none, single_quote]. Default: none
quote_numericsEnclose numeric values the same as strings using the quote character. Default: true
delimited_syntax.row_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code.. Default:
row_delimiterThe character or characters that separate one line from another, for example, CR/LF (Carriage Return/Line Feed). Values: [new_line, carriage_return, carriage_return_line_feed, line_feed]. Default: new_line
delimited_syntax.field_formats.time_formatSpecify a string that defines the format for fields that have the Time data type.
time_formatThe format of time values, for example, HH:mm:ss[.f]
delimited_syntax.field_formats.timestamp_formatSpecify a string that defines the format for fields that have the Timestamp data type.
timestamp_formatThe format of timestamp values, for example, yyyy-MM-dd H:m:s
sheet_nameThe name of the Excel worksheet to write to
_write_mode *Select the Write mode. Values: [delete, write]. Default: write
write_modeWhether to write to, or delete, the target. Values: [delete, write, write_raw]. Default: write




Microsoft Azure SQL Database


Description: Microsoft Azure SQL Database is a fully managed cloud database.
Data source type ID: e375c0ae-cba9-47fc-baf7-523bef88c09e
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: true
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Microsoft SQL Server


Description: Microsoft SQL Server is a relational database management system.
Data source type ID: 48695e79-6279-474a-b539-342625d3dfc2
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
domainThe name of the domain
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
instance_name *The name of the instance
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
use_active_directoryAllows the Microsoft SQL Server connection to authenticate using NTLM
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
query_timeoutSpecify the Query Timeout. If not specified the default value of 300 seconds or 5 minutes will be used.
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




MongoDB


Description: MongoDB is a distributed database that stores data in JSON-like documents. For the MongoDB connector on IBM Cloud, select IBM Cloud Databases for MongoDB.
Data source type ID: c6fb9293-51eb-4f2b-b20c-4dafa3136744
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
auth_databaseThe name of the database in which the user was created
column_discovery_sample_sizeThe number of rows sampled per collection to determine table schemas. The default is 1000.
database *The name of the database
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
schema_filterA comma-separated list of database:collection pairs for which the driver should fetch metadata. For more information look into DataDirect driver documentation.
special_char_behaviorSpecifies whether special characters in names that do not conform to SQL identifier syntax should be stripped (the default), included, or replaced with underscores. Values: [include, replace, strip]
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, static_statement, update, update_statement, update_statement_table_action]. Default: insert




MySQL


Description: MySQL is an open-source relational database management system.
Data source type ID: b2cc3dc2-aff7-4a80-8f80-5e8c5703e9d2
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
encodingThe character encoding for your data. If not specified, the default character set of the database server is used. If you change the value, enter a valid character encoding, for example, UTF-8
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [none, random]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




ODBC


Description: DataStage connector that uses the ODBC application programming interface (API) to a data source.
Data source type ID: 0ca92c3d-0e46-3b42-a573-77958d53c9be
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
additional_propsA semicolon-separated list of additional connection properties.
cluster_nodes *A comma-separated list of member nodes in your cluster.
dsn_type *The ODBC data source type.. Values: [Cassandra, Hive, GreenPlum, DB2, DB2zOS, DB2AS400, Informix, Netezza, Impala, MicrosoftSQLServer, MongoDB, MySQL, Oracle, PostgreSQL, SybaseASE, SybaseIQ]. Default: DB2
database *Database name.
hostname *The hostname of the database.
keyspaceThe name of the Keyspace
network_address *Server name or IP address followed by a comma and the port number.
password *The password used to connect to the database
port *Port
service_name *The Oracle service name that specifies the database used for the connection.
username *The username used to connect to the database

Interaction properties (when used as a source)

Name Type Description
before_after.after_nodeEnter the SQL statements or the fully-qualified name of the file that contains the SQL statements to run once one each node after all data is processed on that node.
before_after.afterEnter the SQL statements or the fully-qualified name of the file that contains the SQL statements to run once after all data is processed.
session.array_sizeThe array size to be used for all read and write database operations. Default: 2000
session.autocommit_modeSpecifies whether the connector commits transactions manually or allows the driver to commit automatically at its discretion. Values: [off, on]. Default: off
before_after.before_nodeEnter the SQL statements or the fully-qualified name of the file that contains the SQL statements to run once on each node before any data is processed on that node.
before_after.beforeEnter the SQL statements or the fully-qualified name of the file that contains the SQL statements to run once before any data is processed.
session.code_pageSpecify a code page to use for this data source. Values: [default, unicode, user-specified]. Default: default
session.code_page.code_page_name *An ICU code page name compatible with this data source
sql.enable_partitioning.partitioning_method.key_field *Specifies the key column that is used by the selected partitioned reads method. This column must be a numeric data type.
session.pass_lob_locator.column *Use to choose columns containing LOBs to be passed by locator (reference)
session.pass_lob_locatorEnables/disables the ability to specify LOB columns to be passed using locator (reference) information. LOB columns not specified will be passed inline. Default: false
sql.enable_partitioningEnable or disable partitioned reads by using the selected partitioning method.. Default: false
enable_quoted_i_dsSpecifies whether or not to enclose database object names in quotes when generating DDL and DML. Default: true
transaction.end_of_wave.end_of_dataSpecifies whether to insert an EOW marker for the last set of records when the number is less than the specified transaction record count value. Default: false
transaction.end_of_waveSpecify settings for the end of wave handling. None means EOW markers are never inserted, Before means EOW markers are inserted before committing the transaction, After means EOW markers are inserted after committing the transaction. Values: [after, before, none]. Default: none
before_after.after_node.fail_on_errorSelect Yes to stop the job if the After SQL (node) statements fails.. Default: true
before_after.after.fail_on_errorSelect Yes to stop the job if the After SQL statements fails.. Default: true
before_after.before_node.fail_on_errorSelect Yes to stop the job if the Before SQL (node) statements fails.. Default: true
before_after.before.fail_on_errorSelect Yes to stop the job if the Before SQL statements fails.. Default: true
session.schema_reconciliation.fail_on_size_mismatchFail if the sizes of numeric and string fields are not compatible when validating the design schema against the database. Default: true
session.schema_reconciliation.fail_on_type_mismatchFail if the types of fields are not compatible when validating the design schema against the database. Default: true
generate_sqlSpecifies whether to generate SQL statements at run time.. Default: false
session.isolation_levelThe isolation level used for all database transactions. Values: [default, read_committed, read_uncommitted, repeatable_read, serializable]. Default: read_uncommitted
limit_rows.limitEnter the maximum number of rows that will be returned by the connector.. Default: 1000
limit_rowsSelect Yes to limit the number of rows that are returned by the connector.. Default: false
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
sql.other_clauseThe other clause predicate of the SQL statement
sql.enable_partitioning.partitioning_methodThe method to use for partitioned reads.. Values: [minimum_and_maximum_range, modulus]. Default: minimum_and_maximum_range
before_after.after_node.read_from_file_after_sql_nodeSelect Yes to read the SQL statements from the file that is specified in the After SQL (node) statements property.. Default: false
before_after.after.read_from_file_after_sqlSelect Yes to read the SQL statements from the file that is specified in the After SQL statements property.. Default: false
before_after.before_node.read_from_file_before_sql_nodeSelect Yes to read the SQL statements from the file that is specified in the Before SQL (node) statements property.. Default: false
before_after.before.read_from_file_before_sqlSelect Yes to read the SQL statements from the file that is specified in the Before SQL statements property.. Default: false
sql.select_statement.read_statement_from_fileSelect Yes to read the SELECT statement from the file specified in the SELECT statement property.. Default: false
transaction.record_countNumber of records per transaction. The value 0 means all available records. Default: 2000
before_afterSelect Yes to run specified SQL statements before and after data is accessed in the database.. Default: false
sql.select_statement *Statement to be executed when reading rows from the database or absolute path to the file containing the SQL statements.
sql.enable_partitioning.partitioning_method.table_name *Specifies the table that is used by the selected partitioned reads method.
table_name *The table name to be used in generated SQL
sql.where_clauseThe where clause predicate of the SQL statement

Interaction properties (when used as a target)

Name Type Description
before_after.after_nodeEnter the SQL statements or the fully-qualified name of the file that contains the SQL statements to run once one each node after all data is processed on that node.
before_after.afterEnter the SQL statements or the fully-qualified name of the file that contains the SQL statements to run once after all data is processed.
session.array_sizeThe array size to be used for all read and write database operations. Default: 2000
session.autocommit_modeSpecifies whether the connector commits transactions manually or allows the driver to commit automatically at its discretion. Values: [off, on]. Default: off
before_after.before_nodeEnter the SQL statements or the fully-qualified name of the file that contains the SQL statements to run once on each node before any data is processed on that node.
before_after.beforeEnter the SQL statements or the fully-qualified name of the file that contains the SQL statements to run once before any data is processed.
sql.user_defined_sql.file.character_setIANA character set name
session.code_pageSpecify a code page to use for this data source. Values: [default, unicode, user-specified]. Default: default
session.code_page.code_page_name *An ICU code page name compatible with this data source
logging.log_column_values.delimiterSpecifies the delimiter to use between columns. Values: [comma, newline, space, tab]. Default: space
table_action.generate_create_statement.create_statement *A statement to be executed when creating the target database table
sql.delete_statement *Statement to be executed when deleting rows from the database
table_action.generate_drop_statement.drop_statement *A statement to be executed when dropping the target database table
session.schema_reconciliation.drop_unmatched_fieldsDrop fields that don't exist in the input schema. Default: true
enable_quoted_i_dsSpecifies whether or not to enclose database object names in quotes when generating DDL and DML. Default: false
sql.user_defined_sql.fail_on_errorAbort the SQL statements when an error occurs. Default: true
table_action.generate_create_statement.fail_on_errorAbort the job if there is an error executing a command. Default: true
table_action.generate_drop_statement.fail_on_errorAbort the job if there is an error executing a command. Default: false
table_action.generate_truncate_statement.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.after_node.fail_on_errorSelect Yes to stop the job if the After SQL (node) statements fails.. Default: true
before_after.after.fail_on_errorSelect Yes to stop the job if the After SQL statements fails.. Default: true
before_after.before_node.fail_on_errorSelect Yes to stop the job if the Before SQL (node) statements fails.. Default: true
before_after.before.fail_on_errorSelect Yes to stop the job if the Before SQL statements fails.. Default: true
session.fail_on_row_error_pxFail the job if a write operation to the target is unsuccessful. Default: true
session.schema_reconciliation.fail_on_size_mismatchFail if the sizes of numeric and string fields are not compatible when validating the design schema against the database. Default: true
session.schema_reconciliation.fail_on_type_mismatchFail if the types of fields are not compatible when validating the design schema against the database. Default: true
sql.user_defined_sql.file *File on the conductor node that contains SQL statements to be executed for each input row
table_action.generate_create_statementSpecifies whether to generate a CREATE TABLE statement at run time. Default: true
table_action.generate_drop_statementSpecifies whether to generate a DROP TABLE statement at run time. Default: true
generate_sqlSpecifies whether to generate SQL statements at run time.. Default: false
table_action.generate_truncate_statementSpecifies whether to generate a TRUNCATE TABLE statement at run time. Default: true
sql.insert_statement *Statement to be executed when inserting rows into the database
session.isolation_levelThe isolation level used for all database transactions. Values: [default, read_committed, read_uncommitted, repeatable_read, serializable]. Default: read_uncommitted
logging.log_column_valuesSpecifies whether to log column values for the first row that fails to be written. Default: false
logging.log_column_values.log_keys_onlySpecifies whether to log key columns or all columns for failing statements. Default: false
before_after.after_node.read_from_file_after_sql_nodeSelect Yes to read the SQL statements from the file that is specified in the After SQL (node) statements property.. Default: false
before_after.after.read_from_file_after_sqlSelect Yes to read the SQL statements from the file that is specified in the After SQL statements property.. Default: false
before_after.before_node.read_from_file_before_sql_nodeSelect Yes to read the SQL statements from the file that is specified in the Before SQL (node) statements property.. Default: false
before_after.before.read_from_file_before_sqlSelect Yes to read the SQL statements from the file that is specified in the Before SQL statements property.. Default: false
transaction.record_countNumber of records per transaction. The value 0 means all available records. Default: 2000
before_afterSelect Yes to run specified SQL statements before and after data is accessed in the database.. Default: false
sql.user_defined_sql.statements *SQL statements to be executed for each input row
table_action *Select the action to perform on the database table. Values: [append, create, replace, truncate]. Default: append
table_name *The table name to be used in generated SQL
table_action.generate_truncate_statement.truncate_statement *A statement to be executed when truncating the database table
sql.update_statement *Statement to be executed when updating rows in the database
sql.user_defined_sql *Source of the user-defined SQL statements. Values: [file, statements]. Default: statements
write_mode *The mode to be used when writing to a database table. Values: [delete, delete_then_insert, insert, insert_new_rows_only, insert_then_update, update, update_then_insert, user-defined_sql]. Default: insert




OData


Description: The OData (Open Data) protocol is a REST-based data access protocol. The OData connector reads data from a data source that uses the OData protocol. To connect to SAP OData, use the SAP OData connection.
Data source type ID: 27c3e1b0-b7d2-4e32-9511-1b8aaa197de0
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
api_key *The api key to use for connecting to the service root
auth_typeThe type of authentication to be used to access the service root. Values: [api_key, none, basic]
password *The password associated with the username for accessing the data source
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
service_root *The URL used to access the service root of a site implementing the OData protocol.
timeout_secondsTimeout value for HTTP calls in seconds
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
entity_set_name *The entity set to be processed
row_limitThe maximum number of rows to return
row_startThe first row of data to read

Interaction properties (when used as a target)

Name Type Description
entity_set_name *The entity set to be processed
write_modeThe mode to be used when writing to the entity. Values: [insert, update]




Oracle


Description: Oracle is a multi-model database management system.
Data source type ID: 971223d3-093e-4957-8af9-a83181ee9dd9
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
connection_mode. Values: [sid, service_name]
sid *The unique name of the database instance. If you provide a SID, do not provide a service name
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
metadata_discoveryDetermines what types of metadata can be discovered, 'No Remarks' option will be set as default. Values: [no_remarks, no_remarks_or_synonyms, no_synonyms, remarks_and_synonyms]
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
service_name *The name of the service. If you provide a service name, do not provide a SID
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Oracle (optimized)


Description: Connect to an Oracle database for the DataStage Oracle (optimized) connector. The Oracle (optimized) connector has more features for DataStage.
Data source type ID: dd22f798-8c9b-41fa-841e-d66cbdf50722
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
oracle_db_host *Oracle host name
password *Specify the password to use to connect to the database.
oracle_db_port *Oracle port
oracle_service_name *Oracle service name
username *Specify the user name to use to connect to the database.

Interaction properties (when used as a source)

Name Type Description
after_sql_nodeEnter the SQL statement to run once one each node after all data is processed on that node.
after_sqlEnter the SQL statement to run once after all data is processed.
array_sizeEnter a number that represents the number of records to process in read and write operations on the database.. Default: 2000
before_sql_nodeEnter the SQL statement to run once on each node before any data is processed on that node.
before_sqlEnter the SQL statement to run once before any data is processed.
read_strategy_column_name *Enter the name of the column that the specified partitioned reads method uses. The column must be of NUMBER(p) type, where p is between 1 and 38, and it must be an existing column in the specified partitioned reads table.
pass_lob_locator.column *Select the LOB columns to pass by reference (locator). Columns that are not selected are passed as actual values (inline).
disconnectEnter the condition under which the connection to the database shall be closed. Values: [0, 1]. Default: 0
pass_lob_locatorSelect Yes to use references (locators) for LOB columns instead of their actual values.. Default: false
enable_partitioned_readsSelect Yes to read data in parallel from multiple processing nodes.. Default: false
enable_quoted_idsSelect Yes to enclose database object names in quotation marks when SQL statements are generated. Quotation marks preserve the case of object names.. Default: true
treat_fetch_truncate_as_errorSelect Yes to stop the job if a truncation occurs when fetching the data. Select No to only log a warning and resume the job.. Default: true
after_sql_node.fail_on_errorSelect Yes to stop the job if the After SQL (node) statement fails.. Default: true
after_sql.fail_on_errorSelect Yes to stop the job if the After SQL statement fails.. Default: true
before_sql_node.fail_on_errorSelect Yes to stop the job if the Before SQL (node) statement fails.. Default: true
before_sql.fail_on_errorSelect Yes to stop the job if the Before SQL statement fails.. Default: true
generate_sqlSelect Yes to automatically generate SQL statements at runtime.. Default: false
inactivity_period *Enter the period of inactivity in seconds after which the connection should be closed. Default: 300
retry_interval *Enter the interval in seconds to wait between attempts to establish a connection. Default: 10
isolation_levelIsolation level. Values: [0, 2, 1]. Default: 0
limitEnter the maximum number of rows that will be returned by the connector.
limit_rowsSelect Yes to limit the number of rows that are returned by the connector.. Default: false
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
application_failover_controlSelect Yes to configure the connector to participate in the Oracle transparent application failover (TAF) process and to report failover progress in the log.. Default: false
end_of_waveSelect Yes to generate end-of-wave record after each wave of records where the number of records in each wave is specified in the Record count property. When the Record count property is set to 0, the end-of-wave records are never generated.. Values: [0, 2]. Default: 0
number_of_retriesEnter a number that represents how many retries the connector will allow for completion of transparent application failover (TAF) after it has been initiated.. Default: 10
retry_count *Enter the number of attempts to establish a connection. Default: 3
other_clauseThe other clause predicate of the SQL statement
pl_sql_statement *Enter a PL/SQL anonymous block. The block must begin with the keyword BEGIN or DECLARE and end with the keyword END.
partition_name *Enter the name of the partition to access.
read_strategy_partition_nameEnter the name of the partition (or subpartition) to use as input for the specified partitioned reads method. This value should typically be set to match the name of the partition (or subpartition) from which the data is fetched.
partitioned_reads_strategySelect the method to use to read data in parallel from multiple processing nodes.. Values: [3, 2, 4, 5, 0, 1]. Default: 0
prefetch_memory_sizeEnter the size of the buffer (in KB) to use for prefetching rows.. Default: 0
prefetch_row_countEnter the number of rows to prefetch when the query runs.. Default: 1
preserve_trailing_blanksSelect Yes to preserve trailing blanks in input text values. Select No for the connector to trim trailing spaces from input text values.. Default: true
treat_warnings_as_errorsSelect Yes to stop the job when the first warning message occurs.. Default: false
read_modeSelect the mode to use to read from the database.. Values: [1, 0]. Default: 0
reconnectSelect Yes to retry to establish connection to the database when the initial connection is unsuccessful or when the active connection is dropped.. Default: false
record_countEnter the number of records to process in each transaction. The record count must be a multiple of the value specified for the Array size property. To process all available records in one transaction, enter 0.. Default: 2000
replay_before_sql_nodeSelect Yes to run the Before SQL (node) statement on each node in a parallel job after a successful transparent application failover (TAF).. Default: false
replay_before_sqlSelect Yes to run the Before SQL statement after a successful transparent application failover (TAF).. Default: false
resume_writeSelect Yes to resubmit the current transaction and to resume sending records to the database after the failover has completed.. Default: false
before_afterSelect Yes to run specified SQL statements before and after data is accessed in the database.. Default: false
select_statement *Enter a SELECT statement. The statement is used to read rows from the database.
subpartition_name *Enter the name of the subpartition to access.
table_name *Enter the name of the Oracle database table or view to access.
read_strategy_table_nameEnter the name of the table to use as input for the specified partition read method. This value should typically be set to match the name of the source table from which the data is fetched.
table_scopeSelect the part of the table to access.. Values: [0, 1, 2]. Default: 0
wait_timeEnter a number, in seconds, that represents the time to wait between transparent application failover (TAF) retries.. Default: 10
transfer_bfile_contTransfer BFILE contents. Default: false
where_clauseThe where clause predicate of the SQL statement

Interaction properties (when used as a target)

Name Type Description
generate_create_statement.fail_on_errorSelect Yes to stop the job if the CREATE TABLE statement fails.. Default: true
generate_drop_statement.fail_on_errorSelect Yes to stop the job if the DROP TABLE statement fails.. Default: true
generate_truncate_statement.fail_on_errorSelect Yes to stop the job if the TRUNCATE TABLE statement fails.. Default: true
after_sql_nodeEnter the SQL statement to run once one each node after all data is processed on that node.
after_sqlEnter the SQL statement to run once after all data is processed.
enable_parallel_load_sessionsSelect Yes to allow multiple concurrent load sessions on the target table, partition or subpartition to which the data is loaded.. Default: true
array_sizeEnter a number that represents the number of records to process in read and write operations on the database.. Default: 2000
before_sql_nodeEnter the SQL statement to run once on each node before any data is processed on that node.
before_sqlEnter the SQL statement to run once before any data is processed.
before_afterBefore and after SQL. Default: false
buffer_size_in_kilobytesEnter the size, in KB, to use for the direct path load buffer. Default: 1024
cache_sizeEnter the size, in elements, of the Oracle date cache.. Default: 1000
delimiterSpecifies the delimiter to use between columns. Values: [3, 1, 0, 2]. Default: 0
cont_fileControl file name
create_statement *Enter the CREATE TABLE statement to run to create the target database table.
data_fileData file name
degree_of_parallelismEnter a number that represents the degree of parallelism to use in the parallel clause. Leave this property blank for Oracle database to automatically calculate the optimal parallelism degree.
delete_statement *Enter a DELETE statement. The statement is used to delete rows from the database.
directory_cont_fileControl file name
disable_when_fullSelect Yes to disable the use of the Oracle date cache if it becomes full during the bulk load.. Default: false
before_load.disable_constraintsSelect Yes to disable all constraints on the table before the bulk load starts.. Default: false
disable_redo_logSelect Yes to disable the generation of Oracle redo and invalidation redo logs. Select No to use the default attributes of the table, partition or subpartition segment to which the data is loaded.. Default: false
before_load.disable_triggersSelect Yes to disable all triggers on the table before the bulk load starts.. Default: false
disconnectEnter the condition under which the connection to the database shall be closed. Values: [0, 1]. Default: 0
drop_statement *Enter the DROP TABLE statement to run to drop the target database table.
drop_unmatched_fieldsSelect Yes to ignore the input schema columns that could not be mapped to any columns in the target database. Select No to stop the job if any unused columns are detected on the input schema.. Default: false
enable_constraintsSelect Yes to enable constraints on the table after the bulk load ends.. Default: false
enable_quoted_idsSelect Yes to enclose database object names in quotation marks when SQL statements are generated. Quotation marks preserve the case of object names.. Default: false
enable_triggersSelect Yes to enable triggers on the table after the bulk load ends.. Default: false
exceptions_table_nameEnter the name of the table to use to store the row identifiers for rows that failed the constraint checks. If the table exists, it will be truncated before the bulk load. If the table does not exist, it will be created.
fail_if_no_rows_deletedSelect Yes to stop the job if the input record does not result in any table deletes and the record is not sent to the reject link. Select No to resume the job.. Default: false
fail_if_no_rows_updatedSelect Yes to stop the job if the input record does not result in any table updates and the record is not sent to the reject link. Select No to resume the job.. Default: false
after_sql_node.fail_on_errorSelect Yes to stop the job if the After SQL (node) statement fails.. Default: true
after_sql.fail_on_errorSelect Yes to stop the job if the After SQL statement fails.. Default: true
before_sql_node.fail_on_errorSelect Yes to stop the job if the Before SQL (node) statement fails.. Default: true
before_sql.fail_on_errorSelect Yes to stop the job if the Before SQL statement fails.. Default: true
fail_on_rebuild_indexSelect Yes to stop the job if the rebuilding of the indexes fails.. Default: false
fail_on_row_error_pxFail the job if a write operation to the target is unsuccessful. Default: true
generate_sqlSelect Yes to automatically generate SQL statements at runtime.. Default: false
generate_create_statementSelect Yes to automatically generate the CREATE TABLE statement at runtime.. Default: false
generate_drop_statementSelect Yes to automatically generate the DROP TABLE statement at runtime.. Default: false
generate_truncate_statementSelect Yes to automatically generate the TRUNCATE TABLE statement at runtime.. Default: false
inactivity_period *Enter the period of inactivity in seconds after which the connection should be closed. Default: 300
skip_indexesSelect the maintenance option to use for indexes during the bulk load.. Values: [0, 2, 1]. Default: 0
insert_statement *Enter an INSERT statement. The statement is used to insert rows into the database.
retry_interval *Enter the interval in seconds to wait between attempts to establish a connection. Default: 10
isolation_levelSelect the isolation level to use for each transaction.. Values: [0, 2, 1]. Default: 0
load_optLoad options
log_column_valuesSpecifies whether to log column values for the first row that fails to be written. Default: false
log_keys_onlySpecifies whether to log key columns or all columns for failing statements. Default: false
logging_clauseSelect the logging clause to include in the ALTER INDEX statement when rebuilding indexes.. Values: [0, 2, 1]. Default: 0
application_failover_controlSelect Yes to configure the connector to participate in the Oracle transparent application failover (TAF) process and to report failover progress in the log.. Default: false
manual_modeManual mode
number_of_retriesEnter a number that represents how many retries the connector will allow for completion of transparent application failover (TAF) after it has been initiated.. Default: 10
retry_count *Enter the number of attempts to establish a connection. Default: 3
pl_sql_statement *Enter a PL/SQL anonymous block. The block must begin with the keyword BEGIN or DECLARE and end with the keyword END.
parallel_clauseSelect the parallel clause to include in the ALTER INDEX statement when rebuilding indexes.. Values: [0, 1, 2, 3]. Default: 0
partition_name *Enter the name of the partition to access.
after_loadSelect Yes to perform selected operations on the table after the bulk load ends.. Default: false
before_loadSelect Yes to perform selected operations on the table before the bulk load starts.. Default: false
table_action_firstSelect Yes to perform table action first. Select No to run Before SQL statements first.. Default: true
preserve_trailing_blanksSelect Yes to preserve trailing blanks in input text values. Select No for the connector to trim trailing spaces from input text values.. Default: true
treat_warnings_as_errorsSelect Yes to stop the job when the first warning message occurs.. Default: false
rebuild_indexesSelect Yes to rebuild indexes on the table after the bulk load ends.. Default: false
reconnectSelect Yes to retry to establish connection to the database when the initial connection is unsuccessful or when the active connection is dropped.. Default: false
record_countEnter the number of records to process in each transaction. The record count must be a multiple of the value specified for the Array size property. To process all available records in one transaction, enter 0.. Default: 2000
replay_before_sql_nodeSelect Yes to run the Before SQL (node) statement on each node in a parallel job after a successful transparent application failover (TAF).. Default: false
replay_before_sqlSelect Yes to run the Before SQL statement after a successful transparent application failover (TAF).. Default: false
resume_writeSelect Yes to resubmit the current transaction and to resume sending records to the database after the failover has completed.. Default: false
subpartition_name *Enter the name of the subpartition to access.
table_action *Select the action to perform before writing data to the table.. Values: [0, 1, 2, 3]. Default: 0
table_name *Enter the name of the Oracle database table or view to access.
table_scopeSelect the part of the table to access.. Values: [0, 1, 2]. Default: 0
wait_timeEnter a number, in seconds, that represents the time to wait between transparent application failover (TAF) retries.. Default: 10
truncate_statement *Enter the TRUNCATE TABLE statement to run to truncate the target database table.
update_statement *Enter an UPDATE statement. The statement is used to update rows into the database.
use_date_cacheSelect Yes to use the Oracle date cache. Using the date cache may improve performance when many identical date values are loaded into date columns in the target table.. Default: false
write_modeSelect the mode to use to write to the database.. Values: [6, 2, 5, 0, 9, 3, 8, 1, 4]. Default: 0




PostgreSQL


Description: PostgreSQL is an open source and customizable object-relational database. For the PostgreSQL connector on IBM Cloud, select IBM Cloud Databases for PostgreSQL.
Data source type ID: e1c23729-99d8-4407-b3df-336e33ffdc82
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
query_timeoutSets the default query timeout in seconds for all statements created by a connection. If not specified the default value of 300 seconds will be used.. Default: 300
retry_limitSpecify the maximum number of retry connection attempts to be made by the connector with an increasing delay between each retry. If no value is provided, two attempts will be made by default if necessary.
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Presto


Description: Presto is a fast and reliable SQL engine for data analytics and the data lakehouse.
Data source type ID: 11849f0a-54cc-448d-bb8c-d79206636e3d
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
host *The hostname or IP address of the database
passwordThe password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
catalog_nameThe name of the catalog that contains the schema to read from. It is required when a fully qualified table name has not been provided.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from. It is required when a fully qualified table name has not been provided.
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description




SAP ASE


Description: SAP ASE is a relational model database server. SAP ASE was formerly named Sybase.
Data source type ID: 6976a3fc-b2ad-4db6-818c-ea049cac309d
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
ssl_certificate_hostHostname in the SubjectAlternativeName or Common Name (CN) part of the SSL certificate
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, static_statement, update, update_statement, update_statement_table_action]. Default: insert




SAP IQ


Description: SAP IQ is a column-based, petabyte scale, relational database software system used for business intelligence, data warehousing, and data marts. SAP IQ was formerly named Sybase IQ.
Data source type ID: 49079262-fac2-4762-99d1-452c1caf6b49
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, static_statement, update, update_statement, update_statement_table_action]. Default: insert




SAP OData


Description: Use the SAP OData connector to extract data from a SAP system through its exposed OData services.
Data source type ID: 79a0a133-cbb6-48d0-a3b0-0956a9655401
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
api_key *The api key to use for connecting to the service root
auth_typeThe type of authentication to be used to access the service root. Values: [api_key, none, basic]
password *The password associated with the username for accessing the data source
sap_gateway_url *The URL used to access the SAP gateway catalog
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
timeout_secondsTimeout value for HTTP calls in seconds
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
entity_set_name *The entity set to be processed
row_limitThe maximum number of rows to return
row_startThe first row of data to read
service_name *The name of the service containing the entity set to be processed
service_versionThe version of the service containing the entity set to be processed

Interaction properties (when used as a target)

Name Type Description
csrf_protectionA flag indicating if this service has Cross-Site Request Forgery protection enabled. Default: true
entity_set_name *The entity set to be processed
service_name *The name of the service containing the entity set to be processed
service_versionThe version of the service containing the entity set to be processed
write_modeThe mode to be used when writing to the entity. Values: [insert, update]




Salesforce.com


Description: Salesforce.com offers cloud-based customer relationship management (CRM) software. The Salesforce.com connector supports the standard SQL query language to select, insert, update, and delete data from Salesforce products and other supported products that use the Salesforce API.
Data source type ID: 06847b16-07b4-4415-a924-c63d11a17aa1
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
password *The password associated with the username for accessing the data source
server_nameThe name of the server to log in. Default: login.salesforce.com
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Salesforce.com (optimized)


Description: Connect to Salesforce.com for the DataStage Salesforce.com (optimized) connector. The Salesforce.com (optimized) connector has more features for DataStage.
Data source type ID: 3a00dbd2-2540-4976-afc2-5fc59f68ed35
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
authentication_typeSelect the Authentication Type. Values: [oauth_jwt, oauth_username_and_password, username_and_password]. Default: username_and_password
consumer_keyConsumer Key for OAuth
consumer_secret_keyConsumer Secret Key for OAuth 2.0 Username-Password Flow
password *Salesforce.com password
server_certificateSecurity Certificate for OAuth 2.0 JWT Bearer Flow
token_expiry_timeToken Expiry Time for OAuth Authentication
url *SOAP endpoint URL from your SOAP project
username *Salesforce.com user name

Interaction properties (when used as a source)

Name Type Description
access_methodSelect the access method. Values: [bulk_mode, real_time_mode]. Default: real_time_mode
salesforce_object_name *The business object to retrieve or update.
batch_sizeData batch size, default is 200. Default: 200
end_time *Enter end time for delta extraction: yyyy-mm-dd hh:mm:ss. Default: CurrentTime
delta_extract_id *Delta extraction ID
start_time *Enter start time for delta extraction: yyyy-mm-dd hh:mm:ss. Default: LastExtractTime
pk_chunkingSpecify true to enable PK Chunking for bulk query operation. Default: true
enable_flat_fileEnable or disable the ability to specify a column that contains a file path for load or extract large object. Default: false
flat_file_column_name *Specify the name of the column that contains a file full path. UNC and relative paths are not supported.
flat_file_content_name *Specify the field name in the object that contains the name of the file to be downloaded
flat_file_folder_location *Specify the path to the folder location for downloading the files that are extracted from Salesforce. UNC and relative paths are not supported.
flat_file_overwriteWhen set as true it will overwrite the file on the disk. When set as false, a new file will be created with the date and time of the file creation added at the beginning of the file name. Default: true
job_id *Enter the job ID or the name of the file for the job ID
file_path_job_id *Specify the absolute file path for saving Salesforce bulk mode job ID
sf_job_id_in_fileSet to yes if job ID to be specified or saved in a file. Default: false
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
reference_soqlSOQL query generated by the importer program for reference
read_modeSalesforce read operation. Values: [get_deleted_delta, get_the_bulk_load_status, get_updated_delta, query, query_all]. Default: query
soql_string *SOQL query statement to be sent to the Salesforce
sleepNumber of seconds between job and batch status recheck. Default: 60
tenacityMaximum number of seconds to recheck the job and batch status. Default: 1800

Interaction properties (when used as a target)

Name Type Description
access_methodSelect the access method. Values: [bulk_mode, real_time_mode]. Default: real_time_mode
salesforce_object_name *The business object to retrieve or update.
batch_sizeData batch size, default is 200. Default: 200
hard_delete_propertySpecify true to empty the recycle bin after delete operation. Default: false
enable_flat_fileEnable or disable the ability to specify a column that contains a file path for load or extract large object. Default: false
flat_file_column_name *Specify the name of the column that contains a file full path. UNC and relative paths are not supported.
file_path_job_id *Specify the absolute file path for saving Salesforce bulk mode job ID
sf_job_id_in_fileSet to yes if job ID to be specified or saved in a file. Default: false
keep_temp_fileSpecify yes to keep the temporary files. Default: false
backend_load_methodSelect the Salesforce.com concurrency mode. Values: [parallel, sequential]. Default: parallel
write_modeSalesforce write operation. Values: [create, delete, update, upsert]. Default: upsert




SingleStoreDB


Description: SingleStoreDB is a cloud-based SQL data storage and analytics service.
Data source type ID: 54221732-8b20-4857-8be1-233c5e98d0ad
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
database *The name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
port *The port of the database
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [none, random]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Snowflake


Description: Snowflake is a cloud-based data storage and analytics service.
Data source type ID: 2fc1372f-b58c-4d45-b0c4-dfb32a1c78a5
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
account_name *The full name of your account (provided by Snowflake)
auth_method. Values: [key_pair, username_password]. Default: username_password
database *The name of the database
key_passphraseThe key passphrase needed for private key decryption
authenticator_urlAuthenticate through native Okta. To enable native SSO through Okta, set this property to the Okta URL endpoint for your Okta account. Leave blank to use internal Snowflake authenticator.
password *The password associated with the username for accessing the data source
private_key *The private key
roleThe default access control role to use in the Snowflake session
username *The username for accessing the data source
warehouse *The virtual warehouse

Interaction properties (when used as a source)

Name Type Description
_before_after._after_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node after all of the data is processed on that node.
_before_after._after_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once after all of the data is processed.
_auto_commit_modeConfigure the stage to run in auto-commit mode. In auto-commit mode, the transaction is committed automatically after each statement is executed. When the connector writes records to the data source, the transaction is committed after each row is written to the data source. When the stage is configured to run multiple statements on each row, the transaction is committed after each statement is executed on the row. Values: [disable, enable]. Default: enable
_before_after._before_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node before any data is processed on that node.
_before_after._before_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once before any data is processed.
_begin_sqlEnter the SQL statement to run one time before any records are processed in the transaction
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
_session._character_set_for_non_unicode_columnsSelect the character set option for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode. If you select the Default option, the character set encoding of the engine host system locale is used. If you select the Custom option, you must provide the character set name to be used.. Values: [_custom, _default]. Default: _default
_session._character_set_for_non_unicode_columns._character_set_name *Specify the name of the character set encoding for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode.
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
_session._default_length_for_columnsEnter the default length for the Char, NChar, Binary, VarChar, NVarChar, and VarBinary link columns for which the Length attribute is not set.. Default: 200
_session._default_length_for_long_columnsEnter the default length for the LongVarChar, LongNVarChar and LongVarBinary link columns for which the Length attribute is not set.. Default: 20000
_enable_partitioned_readsSelect Yes to run the statement on each processing node. The [[node-number]], [[node-number-base-one]] and [[node-count]] placeholders in the statement are replaced on each processing node with the actual zero-based node index, one-based node index and total number of nodes, respectively. Default: false
_enable_quoted_idsSelect Yes to enclose the specified table name and column names on the links in quoting strings when SQL statements are generated. The connector queries the driver to determine the quoting string. If it fails to obtain this information from the driver, the connector uses the backtick (`) character as the quoting string. The default is No.. Default: false
_end_sqlEnter the SQL statement to run one time in the transaction after all the records were processed in the transaction and before the transaction completes successfully
_session._fetch_sizeSpecify the number of rows that the driver must try to fetch from the data source when the connector requests a single row. Fetching rows in addition to the row requested by the connector can improve performance because the driver can complete the subsequent requests for more rows from the connector locally without a need to access the data source. The default value is 0, which indicates that the driver optimizes the fetch operation based on its internal logic.. Default: 0
_generate_sqlSelect Yes to automatically generate the SQL statements at run time.. Default: true
_session._generate_all_columns_as_unicodeAlways generate columns as NChar, NVarChar and LongNVarChar columns instead of Char, VarChar and LongVarChar columns.. Default: false
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
_isolation_levelSpecify how the connector manages statements in transactions. As soon as the connector establishes a connection and issues the first transactional statement, the connector implicitly starts a transaction that uses the specified isolation level. Values: [default, read_committed]. Default: default
_session._keep_conductor_connection_aliveSelect Yes to keep the connection alive in the conductor process while the player processes are processing records. Select No to close the connection in the conductor process before player processes start processing records, and to connect again if necessary after the player processes complete processing the records.. Default: true
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
_end_of_waveSelect Yes to generate an end-of-wave record after each wave of records, where the number of records in each wave is specified in the Record count property. When the Record count property is set to 0, the end-of-wave records are not generated. Values: [_no, _yes]. Default: _no
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
_record_countSpecify the number of rows that the stage reads from or writes to the data source in a single transaction. When this property is set to 0, the transaction is committed only once on each processing node of the stage after the stage processes all the rows on that node. When rows arrive on the input link of the stage in waves, the Record count value applies to each wave separately. Default: 2000
_session._report_schema_mismatchSelect Yes to perform early comparison of the column definitions on the link with the column definitions in the data source and to issue warning messages for any detected discrepancies which can result in data corruption. Depending on the environment and the usage scenario the early detection of discrepancies may not be possible in which case the error messages are reported only when the actual data corruption is detected. . Default: false
_limit_rows._limitEnter the maximum number of rows that will be returned by the connector
row_limitThe maximum number of rows to return
_before_afterSelect Yes to run SQL statements before and after data is accessed in the database.. Default: false
_begin_end_sqlSelect Yes to run SQL statements every time when a transaction begins and every time before a transaction ends. Default: false
_run_end_sql_if_no_records_processedSelect Yes to run the End SQL statement irrespective of the number of records processed in the transaction. Select No to run the End SQL statement only if one or more records were processed in the transaction. Default: false
sampling_percentagePercentage for each row or block to be included in the sample
sampling_seedSeed to be used for getting a repeatable sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none, row]. Default: none
schema_nameThe name of the schema that contains the table to read from
_select_statement *Enter a SELECT statement or the fully qualified name of the file that contains the SELECT statement. The statement is used to read rows from the database.
select_statement *The SQL SELECT statement for retrieving data from the table
_before_after._after_sql_node._fail_on_errorSelect Yes to stop the job if the After SQL (node) statement fails.. Default: true
_before_after._after_sql._fail_on_errorSelect Yes to stop the job if the After SQL statement fails.. Default: true
_before_after._before_sql_node._fail_on_errorSelect Yes to stop the job if the Before SQL (node) statement fails.. Default: true
_before_after._before_sql._fail_on_errorSelect Yes to stop the job if the Before SQL statement fails.. Default: true
_table_name *Enter the fully qualified name of the table that you want to access in the data source.
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
_load_from_file._s3._access_key *Specify the Amazon Web Services access key
_before_after._after_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node after all of the data is processed on that node.
_before_after._after_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once after all of the data is processed.
_auto_commit_modeConfigure the stage to run in auto-commit mode. In auto-commit mode, the transaction is committed automatically after each statement is executed. When the connector writes records to the data source, the transaction is committed after each row is written to the data source. When the stage is configured to run multiple statements on each row, the transaction is committed after each statement is executed on the row. Values: [disable, enable]. Default: enable
_load_from_file._azure._storage_area_name *Specify the name of the Azure Storage account name
_session._batch_sizeEnter the number of records to include in the batch of records for each statement execution. The value 0 indicates that all input records are passed to the statements in a single batch.. Default: 2000
_before_after._before_sql_nodeEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once on each node before any data is processed on that node.
_before_after._before_sqlEnter the SQL statement or the fully qualified name of the file that contains the SQL statement to run once before any data is processed.
_begin_sqlEnter the SQL statement to run one time before any records are processed in the transaction
_load_from_file._file_format._binary_as_textSelect to enable binary as text. Default: false
_load_from_file._s3._bucket_name *Specify the S3 bucket name
_session._character_set_for_non_unicode_columnsSelect the character set option for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode. If you select the Default option, the character set encoding of the engine host system locale is used. If you select the Custom option, you must provide the character set name to be used.. Values: [_custom, _default]. Default: _default
_session._character_set_for_non_unicode_columns._character_set_name *Specify the name of the character set encoding for the values of Char, VarChar and LongVarChar link columns for which the Extended attribute is not set to Unicode.
_load_from_file._file_format._compressionSpecify the compression used for the files. Values: [auto, brotli, bz2, deflate, gzip, none, raw_deflate, zstd]. Default: none
_load_from_file._create_staging_areaSpecify whether to create staging area or use an existing one. Default: true
create_statementThe Create DDL statement for recreating the target table
_table_action._generate_create_statement._create_statement *Enter the CREATE TABLE statement to run to create the target database table
_custom_statementsCustom statements to be run for each input row
_load_from_file._file_format._date_formatSpecify the Date Format
_session._default_length_for_columnsEnter the default length for the Char, NChar, Binary, VarChar, NVarChar, and VarBinary link columns for which the Length attribute is not set.. Default: 200
_session._default_length_for_long_columnsEnter the default length for the LongVarChar, LongNVarChar and LongVarBinary link columns for which the Length attribute is not set.. Default: 20000
_load_from_file._delete_staging_areaSelect No is the staging area has to be retained, by default it is deleted. Default: false
_delete_statement *Enter a DELETE statement or the fully qualified name of the file that contains a DELETE statement. The statement is used to delete rows from the database.
_table_action._generate_drop_statement._drop_statement *Enter the DROP TABLE statement to run to drop the target database table.
_session._drop_unmatched_fieldsSelect Yes to drop any fields from the input link for which there are no matching parameters in the statements configured for the stage. Select No to issue error message when an unmatched field is present on the link.. Default: false
_enable_quoted_idsSelect Yes to enclose the specified table name and column names on the links in quoting strings when SQL statements are generated. The connector queries the driver to determine the quoting string. If it fails to obtain this information from the driver, the connector uses the backtick (`) character as the quoting string. The default is No.. Default: false
_load_from_file._file_format._encodingSpecify the Encoding
_load_from_file._staging_area_format._encodingSpecify the Encoding. Default: UTF-8
_load_from_file._azure._encryptionSpecify the Encryption method (either NONE or AZURE_CSE). If no value is provided then it would be considered as NONE. Values: [azure_cse, none]. Default: none
_load_from_file._s3._encryptionSpecify Encryption. Values: [aws_sse_s3, none]. Default: none
_end_sqlEnter the SQL statement to run one time in the transaction after all the records were processed in the transaction and before the transaction completes successfully
_load_from_file._staging_area_format._escape_characterSpecify the character to use to escape field and row delimiters. If an escape character exists in the data, the escape character is also escaped. Because escape characters require additional processing, do not specify a value for this property if you do not need to include escape characters in the data
_load_from_file._file_format._field_delimiterSpecify the Field Delimiter
_load_from_file._staging_area_format._field_delimiterSpecify the Field Delimiter. Default: ,
_load_from_file._gcs._file_name *Specify the URL path of the file (bucket/folder/filename)
_load_from_file._azure._file_format_name *Specify the name of the predefined file format to be used for this load
_load_from_file._gcs._file_format *Specify the name of the predefined file format to be used for this load. The specified file format should exist in the database
_load_from_file._file_formatSpecify file format options for when using external staging location. Values: [avro, csv, orc, parquet]. Default: csv
_load_from_file._azure._file_name *Specify the fully qualified filename in the container/folder/filename format
_generate_sqlSelect Yes to automatically generate the SQL statements at run time.. Default: true
_table_action._generate_create_statementSelect Yes to automatically generate the CREATE TABLE statement at run time. Depending on the input link column data types, the driver, and the data source in question, the connector may not be able to determine the corresponding native data types and produce a valid statement. Default: true
_table_action._generate_drop_statementSelect Yes to automatically generate the DROP TABLE statement at run time.. Default: true
_table_action._generate_truncate_statementSelect Yes to automatically generate the TRUNCATE TABLE statement at run time.. Default: true
_java._heap_sizeSpecify the maximum Java Virtual Machine heap size in megabytes.. Default: 256
_insert_statement *Enter an INSERT statement or the fully qualified name of the file that contains an INSERT statement. The statement is used to insert rows into the database.
_isolation_levelSpecify how the connector manages statements in transactions. As soon as the connector establishes a connection and issues the first transactional statement, the connector implicitly starts a transaction that uses the specified isolation level. Values: [default, read_committed]. Default: default
_session._keep_conductor_connection_aliveSelect Yes to keep the connection alive in the conductor process while the player processes are processing records. Select No to close the connection in the conductor process before player processes start processing records, and to connect again if necessary after the player processes complete processing the records.. Default: true
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
_load_from_file._azure._master_keySpecify the Master Key , required when Encryption = AZURE_CSE
_load_from_file._copy_options._on_errorSpecify to continue or abort the load if an error occurs. Values: [abort_statement, continue, skip_file]. Default: abort_statement
_load_from_file._copy_options._other_copy_optionsSpecify other copy options to be used while loading from S3
_load_from_file._staging_area_format._other_file_format_optionsSpecify format options apart from the ones listed above which will be used either in creating the staging area or in the execution of COPY command. If Create staging area=Yes, then these format options will be used when creating the staging area and if Create staging area = NO, then these format options will be used directly in the execution of COPY command. The options should be specified in the option=value format, for example EMPTY_FIELD_AS_NULL=TRUE, if more than one option is to be specified then provide a space between the two options. For list of File format options available please refer Snowflake documentation
_load_from_file._file_format._other_format_optionsSpecify other format options apart from the ones listed above in option=value format
_table_action._table_action_firstSelect Yes to perform the table action first. Select No to run the Before SQL statements first.. Default: true
_load_from_file._purge_copied_filesSpecify whether to purge the files which are copied into the table from the staging area. For external staging area, PURGE=TRUE would delete the file from the bucket after loading the file. Default: true
_load_from_file._staging_area_format._quotesFields optionally enclosed by - Character used to enclose strings. Value can be NONE, single quote character ('), or double quote character ("). Values: [double, none, single]. Default: none
_load_from_file._staging_area_format._record_delimiterSpecify a string or one of the following values: <NL>, <CR>, <LF>, <TAB>. The string can include Unicode escape strings in the form \uNNNN where NNNN is the Unicode character code. Default:
_record_countSpecify the number of rows that the stage reads from or writes to the data source in a single transaction. When this property is set to 0, the transaction is committed only once on each processing node of the stage after the stage processes all the rows on that node. When rows arrive on the input link of the stage in waves, the Record count value applies to each wave separately. Default: 2000
_load_from_file._file_format._record_delimiterSpecify the Record Delimiter
_session._report_schema_mismatchSelect Yes to perform early comparison of the column definitions on the link with the column definitions in the data source and to issue warning messages for any detected discrepancies which can result in data corruption. Depending on the environment and the usage scenario the early detection of discrepancies may not be possible in which case the error messages are reported only when the actual data corruption is detected. . Default: false
_run_end_sql_if_no_records_processedSelect Yes to run the End SQL statement irrespective of the number of records processed in the transaction. Select No to run the End SQL statement only if one or more records were processed in the transaction. Default: false
_before_afterSelect Yes to run SQL statements before and after data is accessed in the database.. Default: false
_begin_end_sqlSelect Yes to run SQL statements every time when a transaction begins and every time before a transaction ends. Default: false
_load_from_file._s3._file_name *Specify the location of the S3 file name from which data needs to be moved to table
_load_from_file._azure._sastoken *Specify the SAS token required for connecting to Azure Storage
schema_nameThe name of the schema that contains the table to write to
_load_from_file._s3._secret_key *Specify the Amazon Web Services secret key
_load_from_file._file_format._skip_byte_order_markSelect to skip the Byte Order Mark(BOM). Default: false
_load_from_file._file_format._snappy_compressionSelect to enable Snappy Compression. Default: false
_load_from_file._gcs._storage_integration *Specify the name of the Google cloud storage stage integration. This integration has to be created outside in Snowflake database
_load_from_file._staging_area_name *Specify the external staging area name
_load_from_file._staging_area_typeThe type of staging area either Snowflake managed or Externally managed one. Values: [external_azure, external_gcs, external_s3, internal_location]. Default: internal_location
static_statement *The SQL used for setup operations, for example a CREATE statement
_before_after._after_sql_node._fail_on_errorSelect Yes to stop the job if the After SQL (node) statement fails.. Default: true
_before_after._after_sql._fail_on_errorSelect Yes to stop the job if the After SQL statement fails.. Default: true
_before_after._before_sql_node._fail_on_errorSelect Yes to stop the job if the Before SQL (node) statement fails.. Default: true
_before_after._before_sql._fail_on_errorSelect Yes to stop the job if the Before SQL statement fails.. Default: true
_table_action._generate_create_statement._fail_on_errorSelect Yes to stop the job if the CREATE TABLE statement fails. Default: true
_table_action._generate_drop_statement._fail_on_errorSelect Yes to stop the job if the DROP TABLE statement fails.. Default: false
_table_action._generate_truncate_statement._fail_on_errorSelect Yes to stop the job if the TRUNCATE TABLE statement fails.. Default: true
_table_action *Select the action to complete before writing data to the table.. Values: [_append, _create, _replace, _truncate]. Default: _append
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
_table_name *Enter the fully qualified name of the table that you want to access in the data source.
table_nameThe name of the table to write to
_load_from_file._file_format._time_formatSpecify the Time Format
_load_from_file._file_format._timestamp_formatSpecify the Timestamp Format
_table_action._generate_truncate_statement._truncate_statement *Enter the TRUNCATE TABLE statement to run to truncate the target database table.
_update_statement *Enter an UPDATE statement or the fully qualified name of the file that contains an UPDATE statement. The statement is used to update rows in the database.
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
_load_from_file._azure._use_existing_file_format *Specify whether to use an existing file format or use the File format options when creating the stage. Default: true
_load_from_file._gcs._use_existing_file_format *Specify whether to use an existing file format or use the File format options when creating the stage. Default: true
_use_merge_statementSpecify Yes to use Snowflake merge statement functionality for Update, Delete, Insert then Update and Delete then Insert write modes. If No is selected Snowflake driver functionality for these write modes will be invoked. Default: true
_write_modeSelect the mode that you want to use to write to the data source. Values: [custom, delete, delete_insert, insert, insert_overwrite, insert_update, load_from_file, update]. Default: insert
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Tableau


Description: Tableau is an interactive data visualization platform.
Data source type ID: 9ebc33eb-8c01-43fd-be1e-7202cf5c2c82
Can be used as a source: Yes
Can be used as a target: No
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
access_token_name *The name of the personal access token to use
access_token_secret *The secret of the personal access token to use
auth_method. Values: [access_token, username_and_password]. Default: username_and_password
host *The hostname or IP address of the Tableau server
password *The password associated with the username for accessing the data source
portThe port of the tableau server
sslThe port is configured to accept SSL connections. Default: true
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
siteThe name of the Tableau site to use
username *The username for accessing the data source

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
file_name *The name of the file to read
infer_as_varcharTreat the data in all columns as VARCHARs. Default: false
infer_record_countThe number of records to process to obtain the structure of the data. Default: 1000
infer_schemaObtain the schema from the file. Default: false
invalid_data_handlingHow to handle values that are not valid: fail the job, null the column, or drop the row. Values: [column, fail, row]. Default: fail
read_modeThe method for reading files. Values: [read_single, read_raw, read_raw_multiple_wildcard, read_multiple_regex, read_multiple_wildcard]. Default: read_single
row_limitThe maximum number of rows to return
row_startIndicates the offset from the row where reading starts to treat as the start of the data
type_mappingOverrides the data types of specified columns in the file's inferred schema, for example, inferredType1:newType1;inferredType2:newType2

Interaction properties (when used as a target)

Name Type Description




Teradata


Description: Teradata provides database and analytics-related services and products.
Data source type ID: 96ec8f53-a818-4ba1-bd8d-c86cc33a0b45
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Supported but optional

Connection properties (connection asset)

Name Type Description
authentication_methodThe way of authentication. Values: [ldap, td2]. Default: td2
client_charsetThe name of the client character set to be used. This property should be used in specific cases. Although the CLIENT_CHARSET connection parameter can be used to override the Teradata JDBC Drivers normal mapping of Teradata session character sets to Java character sets, the CLIENT_CHARSET connection parameter is not intended for use in new Teradata deployments. Data corruption will occur if the wrong Java character set is specified with the CLIENT_CHARSET connection parameter. It is a legacy support feature and would be deprecated soon. If you do not provide a client character set, UTF16 will be used by default.
databaseThe name of the database
host *The hostname or IP address of the database
password *The password associated with the username for accessing the data source
portThe port of the database. Default: 1025
sslThe port is configured to accept SSL connections. Default: false
ssl_certificateThe SSL certificate of the host to be trusted which is only needed when the host certificate was not signed by a known certificate authority
username *The username for accessing the data source
ssl_certificate_validationValidate that the SSL certificate returned by the host is trusted

Interaction properties (when used as a source)

Name Type Description
byte_limitThe maximum number of bytes to return. Use any of following suffixes to change the unit: KB, MB, GB, or TB. A value of 0 returns all data.
call_statementThe SQL statement to execute the stored procedure
decimal_rounding_modeDifferent rounding modes for values in columns of decimal and numeric data types. Values: [ceiling, down, floor, halfdown, halfeven, halfup, up]. Default: floor
read_modeThe method for reading records from the table. Values: [general, select]. Default: general
row_limitThe maximum number of rows to return
sampling_percentagePercentage for each row or block to be included in the sample
sampling_typeIndicates which data sampling type should be used in the select statement. Values: [block, none]. Default: none
schema_nameThe name of the schema that contains the table to read from
select_statement *The SQL SELECT statement for retrieving data from the table
table_name *The name of the table to read from

Interaction properties (when used as a target)

Name Type Description
create_statementThe Create DDL statement for recreating the target table
key_column_namesA comma separated list of column names to override the primary key used during an update or merge
schema_nameThe name of the schema that contains the table to write to
static_statement *The SQL used for setup operations, for example a CREATE statement
table_actionThe action to take on the target table to handle the new data set. Values: [append, replace, truncate]. Default: append
table_nameThe name of the table to write to
update_statementThe SQL INSERT, UPDATE, MERGE, or DELETE statement for updating data in the table.
write_modeThe mode for writing records to the target table. Values: [insert, merge, static_statement, update, update_statement, update_statement_table_action]. Default: insert




Teradata (optimized)


Description: Connect to a Teradata database for the DataStage 'Teradata (optimized)' connector. Use the Teradata (optimized) connector in datstage flows to read,� write, or look up data in Teradata databases.� This connector also supports Parallel MultiLoad, bulk load and reject links for immediate lookups and writes. This connection can be used only in DataStage.
Data source type ID: 96441cf3-4edf-3eb8-89e7-0d16cab7ccec
Can be used as a source: Yes
Can be used as a target: Yes
Secure gateway: Not applicable / not supported

Connection properties (connection asset)

Name Type Description
accountUser account ID for resource accounting
auto_map_charset_encodingDefault is Yes. User can set it to No and specify the required ICU charset encoding. Default: true
queryband.read_from_file.character_setIANA character set name
client_character_setTeradata client character set. Default: UTF8
databaseDefault database
log_on_mechSelect the security mechanism to use to authenticate the user. Select Default to use the default logon mechanism of the Teradata server. Select TD2 to use the Teradata security mechanism. Select LDAP to use an LDAP security mechanism for external authentication.. Values: [default, ldap, td2]. Default: default
max_bytes_per_character *Maximum bytes per character
nls_map_name *Specify the ICU charset encoding name, to be used
password *Specify the password to use to connect to the database.
querybandSemicolon-separated list of name-value pairs used in the generated query band statement for the session
queryband.read_from_fileSelect YES to read the query band expression from the file that is specified in the Query band expression property.. Default: false
serverSpecify the Teradata Director Program ID.
transaction_modeSemantics for SQL transactions. Values: [ansi, teradata]. Default: ansi
unicode_pass_thruEnable or Disable Unicode pass through. Default: false
username *Specify the user name to use to connect to the database.

Interaction properties (when used as a source)

Name Type Description
access_methodSpecify whether to use immediate or bulk access. Values: [bulk, immediate]. Default: immediate
before_after.afterEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once after all data is processed.
before_after.after_nodeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once one each node after all data is processed on that node.
before_after.after_sql_fileFile on the conductor node that contains After SQL statements
session.array_sizeThe array size to be used for all read and write database operations. Default: 2000
before_after.beforeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once before any data is processed.
before_after.before_nodeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once on each node before any data is processed on that node.
before_after.before_sql_fileFile on the conductor node that contains Before SQL statements
before_afterBefore/After SQL properties. Default: false
before_after.after_sql_file.character_setIANA character set name
before_after.before_sql_file.character_setIANA character set name
session.pass_lob_locator.column *Use to choose columns containing LOBs to be passed by locator (reference)
source_temporal_support.transaction_time_qualifier.date_timestamp_expression *Specifies a date or timestamp expression for the AS OF qualifier
source_temporal_support.valid_time_qualifier.date_timestamp_expression *Specifies a date or timestamp expression for the AS OF qualifier
describe_strings_in_bytesDefault is False. Setting this to True, describes char-data as string instead of ustring. Default: false
disconnectEnter the condition under which the connection to the database shall be closed. Values: [never, period_of_inactivity]. Default: never
session.pass_lob_locatorEnables/disables the ability to specify LOB columns to be passed using locator (reference) information. LOB columns not specified will be passed inline. Default: false
enable_quoted_i_dsSpecifies whether or not to enclose database object names in quotes when generating DDL and DML. Default: true
transaction.end_of_wave.end_of_dataSpecifies whether to insert an EOW marker for the last set of records when the number is less than the specified transaction record count value. Default: false
transaction.end_of_waveSpecify settings for the end of wave handling. None means EOW markers are never inserted, Before means EOW markers are inserted before committing the transaction, After means EOW markers are inserted after committing the transaction. Values: [after, before, none]. Default: none
parallel_synchronization.end_timeoutMaximum number of seconds to wait for the other parallel instances to complete. Default: 0
before_after.after.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.after_node.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.after_sql_file.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.before.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.before_node.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.before_sql_file.fail_on_errorAbort the job if there is an error executing a command. Default: true
session.schema_reconciliation.fail_on_size_mismatchFail if sizes of numeric and string fields mismatch when validating the design schema against the database. Default: true
session.schema_reconciliation.fail_on_type_mismatchFail if the data types of the fields mismatch when validating the design schema against the database. Default: true
generate_sqlSpecifies whether to generate SQL statement(s) at runtime. Default: false
disconnect.inactivity_period *Enter the period of inactivity after which the connection should be closed. Default: 300
reconnect.retry_interval *Enter the interval in seconds to wait between attempts to establish a connection. Default: 10
session.isolation_levelDegree of isolation of an application process from concurrent application processes. Values: [default, read_committed, read_uncommitted, repeatable_read, serializable]. Default: default
limit_rows.limitEnter the maximum number of rows that will be returned by the connector.. Default: 1000
limit_rowsSelect Yes to limit the number of rows that are returned by the connector.. Default: false
lookup_typeLookup Type. Values: [empty, pxbridge]. Default: empty
limit_settings.max_buffer_sizeMaximum request or response buffer size. Default: 0
limit_settings.max_partition_sessionsMaximum number of connection sessions per partition. Default: 0
limit_settings.max_sessionsMaximum number of connection sessions. Default: 0
limit_settings.min_sessionsMinimum number of connection sessions. Default: 0
reconnect.retry_count *Enter the number of attempts to establish a connection. Default: 3
sql.other_clauseThe other clause predicate of the SQL statement
parallel_synchronizationParallel synchronization properties. Default: false
source_temporal_support.valid_time_qualifier.period_expressionSpecifies a period expression for the NONSEQUENCED or SEQUENCED qualifier
limit_settings.progress_intervalNumber of rows per partition before a progress message is displayed, or 0 for no messages. Default: 100000
reconnectSelect Yes to retry to establish connection to the database when the initial connection is unsuccessful or when the active connection is dropped.. Default: false
transaction.record_countNumber of records per transaction. The value 0 means all available records. Default: 2000
sql.select_statement *Statement to be executed when reading rows from the database
bulk_access.sleepNumber of minutes between logon retries. Default: 0
parallel_synchronization.sync_idSync table key value
parallel_synchronization.sync_databaseSync table database
parallel_synchronization.sync_passwordSync user password
parallel_synchronization.sync_pollNumber of seconds between retries to update the sync table. Default: 0
parallel_synchronization.sync_serverSync table server name
parallel_synchronization.sync_table *Sync table name
parallel_synchronization.sync_table_actionSelect the table action to perform on the sync table. Values: [append, create, replace, truncate]. Default: create
parallel_synchronization.sync_table_cleanupSelect the cleanup action to perform on the sync table. Values: [drop, keep]. Default: keep
parallel_synchronization.sync_table_write_modeThe mode to be used when writing to the sync table. Values: [delete_then_insert, insert]. Default: insert
parallel_synchronization.sync_timeoutMaximum number of seconds to retry an update of the sync table. Default: 0
parallel_synchronization.sync_userSync table user name
table_name *The table name to be used in generated SQL
source_temporal_support.temporal_columnsSpecifies the temporal columns in the table. Values: [bi-temporal, none, transaction_time, valid_time]. Default: none
source_temporal_supportSpecifies whether the source table has temporal columns. Default: false
bulk_access.tenacityMaximum number of hours to retry the logon operation. Default: 0
source_temporal_support.temporal_columns.transaction_time_column *Specifies the TRANSACTIONTIME column. If the Generate create statement at runtime property is set to Yes, the column will be designated as TRANSACTIONTIME in the generated CREATE TABLE statement
source_temporal_support.transaction_time_qualifierSpecifies the TRANSACTIONTIME qualifier. Values: [as_of, current, non-sequenced, none]. Default: none
session.schema_reconciliation.unused_field_actionSpecify whether to drop unused fields or abort the job. Values: [abort, drop, keep, warn]. Default: abort
source_temporal_support.temporal_columns.valid_time_column *Specifies the VALIDTIME column. If the Generate create statement at runtime property is set to Yes, the column will be designated as VALIDTIME in the generated CREATE TABLE statement
source_temporal_support.valid_time_qualifierSpecifies the VALIDTIME qualifier. Values: [as_of, current, non-sequenced, none, sequenced]. Default: none
sql.where_clauseThe where clause predicate of the SQL statement

Interaction properties (when used as a target)

Name Type Description
access_methodSpecify whether to use immediate or bulk access. Values: [bulk, immediate]. Default: immediate
before_after.afterEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once after all data is processed.
before_after.after_nodeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once one each node after all data is processed on that node.
before_after.after_sql_fileFile on the conductor node that contains After SQL statements
table_action.generate_create_statement.create_table_options.allow_duplicate_rowsControls whether to specify a SET or MULTISET qualifier. Values: [default, no, yes]. Default: default
session.array_sizeThe array size to be used for all read and write database operations. Default: 2000
before_after.beforeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once before any data is processed.
before_after.before_nodeEnter the SQL statement or the fully-qualified name of the file that contains the SQL statement to run once on each node before any data is processed on that node.
before_after.before_sql_fileFile on the conductor node that contains Before SQL statements
before_afterBefore/After SQL properties. Default: false
immediate_access.buffer_usageSpecify whether requests should share the same buffer or use separate buffers. Values: [separate, share]. Default: share
sql.user_defined.file.character_setIANA character set name
before_after.after_sql_file.character_setIANA character set name
before_after.before_sql_file.character_setIANA character set name
parallel_synchronization.checkpoint_timeoutMaximum number of seconds to wait for the other instances to reach the checkpoint. Default: 0
bulk_access.cleanup_modeSpecify whether to drop error tables and the work table if loading ends with an error that cannot be restarted. Values: [drop, keep]. Default: drop
logging.log_column_values.delimiterSpecifies the delimiter to use between columns. Values: [comma, newline, space, tab]. Default: space
session.pass_lob_locator.column *Use to choose columns containing LOBs to be passed by locator (reference)
table_action.generate_create_statement.create_statement *A statement to be executed when creating the target database table
table_action.generate_create_statement.create_table_options.data_block_sizeControls whether to specify a DATABLOCKSIZE clause. Default: 0
bulk_access.update_load.delete_multiple_rowsControls whether to use a delete task to delete multiple rows from a table. Default: false
sql.delete_statement *Statement to be executed when deleting rows from the database
disconnectEnter the condition under which the connection to the database shall be closed. Values: [never, period_of_inactivity]. Default: never
table_action.generate_drop_statement.drop_statement *A statement to be executed when dropping the target database table
bulk_access.error_control.duplicate_insert_rowsSpecify whether to reject or ignore duplicate rows in insert operations. Values: [default, ignore, reject]. Default: default
bulk_access.error_control.duplicate_update_rowsSpecify whether to reject or ignore duplicate rows in update operations. Values: [default, ignore, reject]. Default: default
session.pass_lob_locatorEnables/disables the ability to specify LOB columns to be passed using locator (reference) information. LOB columns not specified will be passed inline. Default: false
enable_quoted_i_dsSpecifies whether or not to enclose database object names in quotes when generating DDL and DML. Default: false
limit_settings.end_rowRow number to end loading. Default: 0
parallel_synchronization.end_timeoutMaximum number of seconds to wait for the other parallel instances to complete. Default: 0
bulk_access.error_limitMaximum number of rows rejected to error table 1. Default: 0
bulk_access.error_table1Table of rows rejected for SQL errors
bulk_access.error_table2Table of rows rejected for uniqueness violations
bulk_access.fail_on_mloa_derrsSet to Yes connector aborts the job if errors exist in the et/uv table and Set to No completes job even if errors exist in et/uv table and user should check tables. Default: true
sql.user_defined.request_type.fail_on_errorAbort the statement sequence when a statement fails. Default: true
before_after.after.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.after_node.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.after_sql_file.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.before.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.before_node.fail_on_errorAbort the job if there is an error executing a command. Default: true
before_after.before_sql_file.fail_on_errorAbort the job if there is an error executing a command. Default: true
table_action.generate_create_statement.fail_on_errorAbort the job if there is an error executing a command. Default: true
table_action.generate_drop_statement.fail_on_errorAbort the job if there is an error executing a command. Default: false
table_action.generate_truncate_statement.fail_on_errorAbort the job if there is an error executing a command. Default: true
sql.user_defined.file *File on the conductor node that contains SQL statements to be executed for each input row
table_action.generate_create_statement.create_table_options.table_free_space.free_space_percentPercent of free space to remain during loading operations. Default: 0
generate_sqlSpecifies whether to generate SQL statement(s) at runtime. Default: false
tmsmevents.generate_uowid *Select Yes to automatically generate the UOW ID by the TMSM at runtime. Select No to specify the UOW ID to use for the dual load.. Default: false
table_action.generate_create_statementSelect Yes to automatically generate the create statement at runtime.. Default: true
table_action.generate_drop_statementSpecifies whether to generate a drop table statement at runtime. Default: true
table_action.generate_truncate_statementSpecifies whether to generate a truncate table statement at runtime. Default: true
disconnect.inactivity_period *Enter the period of inactivity after which the connection should be closed. Default: 300
sql.insert_statement *Statement to be executed when inserting rows into the database
reconnect.retry_interval *Enter the interval in seconds to wait between attempts to establish a connection. Default: 10
bulk_access.load_typeType of bulk load. Values: [load, stream, update]. Default: load
logging.log_column_valuesSpecifies whether to log column values for the first row that fails to be written. Default: false
logging.log_column_values.log_keys_onlySpecifies whether to log key columns or all columns for failing statements. Default: false
bulk_access.log_tableRestart log table
bulk_access.stream_load.macro_databaseDatabase that contains macros used by the Stream load
table_action.generate_create_statement.create_table_options.make_duplicate_copiesControls whether to specify a FALLBACK clause. Values: [default, no, yes]. Default: default
limit_settings.max_buffer_sizeMaximum request or response buffer size. Default: 0
limit_settings.max_partition_sessionsMaximum number of connection sessions per partition. Default: 0
limit_settings.max_sessionsMaximum number of connection sessions. Default: 0
limit_settings.min_sessionsMinimum number of connection sessions. Default: 0
bulk_access.error_control.missing_delete_rowsSpecify whether to reject or ignore missing rows in delete operations. Values: [default, ignore, reject]. Default: default
bulk_access.error_control.missing_update_rowsSpecify whether to reject or ignore missing rows in update operations. Values: [default, ignore, reject]. Default: default
reconnect.retry_count *Enter the number of attempts to establish a connection. Default: 3
bulk_access.stream_load.pack_sizeNumber of statements per request. Default: 0
parallel_synchronizationParallel synchronization properties. Default: false
table_action.generate_create_statement.create_table_options.partition_by_expressionSpecifies the expression for the PARTITION BY clause
target_temporal_support.temporal_qualifier.period_expressionSpecifies a period expression for the SEQUENCED VALIDTIME qualifier. This expression can only be used with write modes that include UPDATE or DELETE.
table_action.generate_create_statement.create_table_options.primary_index_typePrimary index type for key columns. Values: [no_primary_index, non-unique, unique]. Default: non-unique
limit_settings.progress_intervalNumber of rows per partition before a progress message is displayed, or 0 for no messages. Default: 100000
reconnectSelect Yes to retry to establish connection to the database when the initial connection is unsuccessful or when the active connection is dropped.. Default: false
transaction.record_countNumber of records per transaction. The value 0 means all available records. Default: 2000
sql.user_defined.request_typeSpecify whether to separate the statements into individual requests or use a multi-statement request. Values: [individual, multi-statement]. Default: individual
bulk_access.stream_load.robustRobust restart logic. Values: [no, yes]. Default: yes
bulk_access.stream_load.serializeSerialize multiple statements. Values: [no, yes]. Default: yes
table_action.generate_create_statement.create_table_options.server_character_setServer character set for Char and VarChar columns
bulk_access.sleepNumber of minutes between logon retries. Default: 0
bulk_access.start_modeSpecify whether to drop error tables before the connector begins loading or to restart an aborted load. Values: [auto, clean, restart]. Default: clean
limit_settings.start_rowRow number to start loading. Default: 0
sql.user_defined.statements *SQL statements to be executed for each input row
parallel_synchronization.sync_idSync table key value
parallel_synchronization.sync_databaseSync table database
parallel_synchronization.sync_passwordSync user password
parallel_synchronization.sync_pollNumber of seconds between retries to update the sync table. Default: 0
parallel_synchronization.sync_serverSync table server name
parallel_synchronization.sync_table *Sync table name
parallel_synchronization.sync_table_actionSelect the table action to perform on the sync table. Values: [append, create, replace, truncate]. Default: create
parallel_synchronization.sync_table_cleanupSelect the cleanup action to perform on the sync table. Values: [drop, keep]. Default: keep
parallel_synchronization.sync_table_write_modeThe mode to be used when writing to the sync table. Values: [delete_then_insert, insert]. Default: insert
parallel_synchronization.sync_timeoutMaximum number of seconds to retry an update of the sync table. Default: 0
parallel_synchronization.sync_userSync table user name
tmsmeventsOptions for the TMSM events. Default: false
table_action *Select the action to perform on the database table. Values: [append, create, replace, truncate]. Default: append
table_action.generate_create_statement.create_table_options.table_free_spaceControls whether to specify a FREESPACE clause. Values: [default, yes]. Default: default
table_name *The table name to be used in generated SQL
target_temporal_support.temporal_columnsSpecifies the temporal columns in the table. Values: [bi-temporal, none, transaction_time, valid_time]. Default: none
target_temporal_support.temporal_qualifierSpecifies the temporal qualifier for generated SQL. Values: [current_valid_time, non-sequenced_valid_time, non-temporal, none, sequenced_valid_time]. Default: none
target_temporal_supportSpecifies whether the target table has temporal columns. Default: false
bulk_access.tenacityMaximum number of hours to retry the logon operation. Default: 0
target_temporal_support.temporal_columns.transaction_time_column *Specifies the TRANSACTIONTIME column. If the Generate create statement at runtime property is set to Yes, the column will be designated as TRANSACTIONTIME in the generated CREATE TABLE statement
table_action.generate_truncate_statement.truncate_statement *A statement to be executed when truncating the database table
tmsmevents.generate_uowid.uowid *Unique unit of work id
tmsmevents.uowclassThe classification of the unit of work
tmsmevents.uowsourcesystemThe name of the system the data is sourced from.
session.schema_reconciliation.unused_field_actionSpecify whether to drop unused fields or abort the job. Values: [abort, drop, keep, warn]. Default: abort
sql.update_statement *Statement to be executed when updating rows in the database
sql.user_defined *Source of the user-defined SQL statements. Values: [file, statements]. Default: statements
target_temporal_support.temporal_columns.valid_time_column *Specifies the VALIDTIME column. If the Generate create statement at runtime property is set to Yes, the column will be designated as VALIDTIME in the generated CREATE TABLE statement
bulk_access.work_tableWork table
write_mode *The mode to be used when writing to a database table. Values: [delete, delete_then_insert, insert, insert_then_update, update, update_then_insert, user-defined_sql]. Default: insert



Generated on: 2023-06-07T14:56:32.650Z
Generated from: GET /v2/datasource_types?interactive_properties=true&connection_properties=true