Configuration Reference for FTPS Source Connector for Confluent Platform¶
To use this connector, specify the name of the connector class in the connector.class
configuration property.
connector.class=io.confluent.connect.ftps.FtpsSourceConnector
Source connector-specific configuration properties are described below.
General¶
kafka.topic
The Kafka topic to publish records to.
- Type: string
- Importance: high
batch.size
The number of records that should be returned with each batch.
- Type: int
- Default: 1000
- Importance: medium
empty.poll.wait.ms
The amount of time in milliseconds to wait if a poll returns an empty list of records.
- Type: long
- Default: 10000
- Importance: low
Auto topic creation¶
For more information about Auto topic creation, see Configuring Auto Topic Creation for Source Connectors.
Configuration properties accept regular expressions (regex) that are defined as Java regex.
topic.creation.groups
A list of group aliases that are used to define per-group topic configurations for matching topics. A
default
group always exists and matches all topics.- Type: List of String types
- Default: empty
- Possible Values: The values of this property refer to any additional groups. A
default
group is always defined for topic configurations.
topic.creation.$alias.replication.factor
The replication factor for new topics created by the connector. This value must not be larger than the number of brokers in the Kafka cluster. If this value is larger than the number of Kafka brokers, an error occurs when the connector attempts to create a topic. This is a required property for the
default
group. This property is optional for any other group defined intopic.creation.groups
. Other groups use the Kafka broker default value.- Type: int
- Default: n/a
- Possible Values:
>= 1
for a specific valid value or-1
to use the Kafka broker’s default value.
topic.creation.$alias.partitions
The number of topic partitions created by this connector. This is a required property for the
default
group. This property is optional for any other group defined intopic.creation.groups
. Other groups use the Kafka broker default value.- Type: int
- Default: n/a
- Possible Values:
>= 1
for a specific valid value or-1
to use the Kafka broker’s default value.
topic.creation.$alias.include
A list of strings that represent regular expressions that match topic names. This list is used to include topics with matching values, and apply this group’s specific configuration to the matching topics.
$alias
applies to any group defined intopic.creation.groups
. This property does not apply to thedefault
group.- Type: List of String types
- Default: empty
- Possible Values: Comma-separated list of exact topic names or regular expressions.
topic.creation.$alias.exclude
A list of strings representing regular expressions that match topic names. This list is used to exclude topics with matching values from getting the group’s specfic configuration.
$alias
applies to any group defined intopic.creation.groups
. This property does not apply to thedefault
group. Note that exclusion rules override any inclusion rules for topics.- Type: List of String types
- Default: empty
- Possible Values: Comma-separated list of exact topic names or regular expressions.
topic.creation.$alias.${kafkaTopicSpecificConfigName}
Any of the Changing Broker Configurations Dynamically for the version of the Kafka broker where the records will be written. The broker’s topic-level configuration value is used if the configuration is not specified for the rule.
$alias
applies to thedefault
group as well as any group defined intopic.creation.groups
.- Type: property values
- Default: Kafka broker value
Schema¶
key.schema
The schema for the key written to Kafka.
- Type: string
- Default: {“name”:”default-key-schema”,”type”:”STRUCT”,”isOptional”:false,”fieldSchemas”:{“filePath”:{“type”:”STRING”,”isOptional”:true}}}
- Importance: medium
value.schema
The schema for the value written to Kafka.
- Type: string
- Default: “”
- Importance: medium
schema.generation.enabled
Flag to determine if schemas should be dynamically generated. If set to true,
key.schema
andvalue.schema
can be omitted, butschema.generation.key.name
andschema.generation.value.name
must be set. Nested JSON is currently not supported and will be ignored. Schemaless JSON parsing should be used instead.- Type: boolean
- Default: true
- Importance: medium
schema.generation.key.fields
The field(s) to use to build a key schema. This is used only during schema generation.
- Type: list
- Default: “”
- Importance: medium
schema.generation.key.name
The name of the generated key schema.
- Type: string
- Default: default_key_schema_name
- Importance: low
schema.generation.value.name
The name of the generated value schema.
- Type: string
- Default: default_value_schema_name
- Importance: low
Timestamp¶
ftps.timestamp.mode
Determines how the connector will set the timestamp for the Source Record. If set to
FIELD
, then the timestamp will be read from a field in the value. Specify the field inftps.timestamp.field
. If set toFILE_TIME
, then the last modified time of the file will be used. If set toPROCESS_TIME
, the time the record is read will be used.- Type: string
- Default: PROCESS_TIME
- Valid Values: one of [FIELD, FILE_TIME, PROCESS_TIME]
- Importance: medium
ftps.timestamp.field
The field in the value schema that will contain the parsed timestamp for the record. This field cannot be marked as optional and must be a Timestamp.
- Type: string
- Default: “”
- Importance: medium
File server connection¶
ftps.host
The address of the FTPS server.
- Type: string
- Importance: high
ftps.username
The username used to authenticate with the file server.
- Type: string
- Importance: high
ftps.password
The password used to authenticate with the file server.
- Type: password
- Importance: high
ftps.port
Port number of the FTPS server.
- Type: int
- Default: 21
- Valid Values: [0,…]
- Importance: medium
ftps.security.mode
The mode used to connect with the server. This is related to the establishment of the TLS connection and is up to the client.
- Type: string
- Default: EXPLICIT
- Valid Values: one of [EXPLICIT, IMPLICIT]
- Importance: low
ftps.server.timezone
The timezone on the FTPS server. This is used when calculating the age of a file.
- Type: string
- Default: UTC
- Valid Values: Should be one of the identifiers used by
java.util.TimeZone
to refer to time zones, for example, America/Chicago or UTC. - Importance: low
File System¶
ftps.input.path
The FTPS directory to read files that will be processed.
- Type: string
- Importance: high
ftps.input.file.pattern
Regular expression to check input file names against. This expression must match the entire filename.
- Type: string
- Default: .+
- Importance: high
ftps.error.path
The FTPS directory to place files that have errors.
- Type: string
- Default: “”
- Importance: high
ftps.finished.path
The FTPS directory to place files that have been successfully processed.
- Type: string
- Default: “”
- Importance: high
ftps.behavior.on.error
Specifies whether the task should halt when it encounters an error or continue to the next file.
- Type: string
- Default: LOG
- Valid Values: one of [LOG, IGNORE, FAIL]
- Importance: medium
ftps.schemaless.json.parsing
To specify whether JSON files should be parsed without using any schema. Needed only when JSON is one of file types specified in file.types.to.parse.
- Type: boolean
- Default: false
- Importance: medium
ftps.ignore.file.extension
By default, the parser to use for a file is chosen based on the file’s extension. Instead, if a single parser is to be used for each file, regardless of extension, this property should be set to true. In that case, the parser specified in parser.type will be used for parsing all files.
- Type: boolean
- Default: false
- Importance: medium
ftps.parser.type
The parser to use for parsing files. Required only if
ftps.ignore.file.extension
istrue
.- Type: string
- Default: TXT
- Valid Values: one of [TXT, CSV, JSON]
- Importance: medium
ftps.cleanup.policy
Determines how the connector should cleanup the files that have been successfully processed. NONE leaves the files in place which could cause them to be reprocessed if the connector is restarted. DELETE removes the file from the filesystem. MOVE will move the file to a finished directory.
- Type: string
- Default: NONE
- Valid Values: one of [DELETE, MOVE, NONE]
- Importance: medium
ftps.file.minimum.age.ms
The amount of time in milliseconds after the file was last written to before the file can be processed.
- Type: long
- Default: 0
- Valid Values: [0,…]
- Importance: medium
Security¶
ftps.ssl.key.password
The password of the private key in the key store file. This is optional for client.
- Type: password
- Default: null
- Importance: high
ftps.ssl.keystore.location
The location of the key store file. This is optional for client and can be used for two-way authentication for client.
- Type: string
- Default: null
- Importance: high
ftps.ssl.keystore.password
The store password for the key store file. This is optional for client and only needed if
ssl.keystore.location
is configured.- Type: password
- Default: null
- Importance: high
ftps.ssl.truststore.location
The location of the truststore file.
- Type: string
- Default: null
- Importance: high
ftps.ssl.truststore.password
The password for the truststore file. If a password is not set access to the truststore is still available, but integrity checking is disabled.
- Type: password
- Default: null
- Importance: high
ftps.ssl.enabled.protocols
The list of protocols enabled for SSL connections.
- Type: list
- Default: TLSv1.2,TLSv1.1,TLSv1
- Importance: medium
ftps.ssl.keystore.type
The file format of the key store file. This is optional for client.
- Type: string
- Default: JKS
- Importance: medium
ftps.ssl.protocol
The SSL protocol used to generate the SSLContext. Default setting is TLS, which is fine for most cases. Allowed values in recent JVMs are TLS, TLSv1.1 and TLSv1.2. SSL, SSLv2 and SSLv3 may be supported in older JVMs, but their usage is discouraged due to known security vulnerabilities.
- Type: string
- Default: TLS
- Importance: medium
ftps.ssl.provider
The name of the security provider used for SSL connections. Default value is the default security provider of the JVM.
- Type: string
- Default: null
- Importance: medium
ftps.ssl.truststore.type
The file format of the truststore file.
- Type: string
- Default: JKS
- Importance: medium
ftps.ssl.cipher.suites
A list of cipher suites. This is a named combination of authentication, encryption, MAC and key exchange algorithm used to negotiate the security settings for a network connection using TLS or SSL network protocol. By default all the available cipher suites are supported.
- Type: list
- Default: null
- Importance: low
ftps.ssl.endpoint.identification.algorithm
The endpoint identification algorithm to validate server hostname using server certificate.
- Type: string
- Default: https
- Importance: low
ftps.ssl.keymanager.algorithm
The algorithm used by key manager factory for SSL connections. Default value is the key manager factory algorithm configured for the Java Virtual Machine.
- Type: string
- Default: SunX509
- Importance: low
ftps.ssl.secure.random.implementation
The SecureRandom PRNG implementation to use for SSL cryptography operations.
- Type: string
- Default: null
- Importance: low
ftps.ssl.trustmanager.algorithm
The algorithm used by trust manager factory for SSL connections. Default value is the trust manager factory algorithm configured for the Java Virtual Machine.
- Type: string
- Default: PKIX
- Importance: low
TXT configs¶
ftps.txt.file.charset
Character set to read a text file with.
- Type: string
- Default: UTF-8
- Valid Values: Should be a valid charset such as UTF-8, UTF-16, IBM01147 etc.
- Importance: low
CSV configs¶
ftps.csv.first.row.as.header
Flag to indicate if the fist row of data contains the header of the file. If true the position of the columns will be determined by the first row to the CSV. The column position will be inferred from the position of the schema supplied in
value.schema
. If set to true the number of columns must be greater than or equal to the number of fields in the schema and the first row, which is the header, will be used to set the field names in the generated schema ifschema.generation.enabled
is set to true.- Type: boolean
- Default: false
- Importance: medium
ftps.csv.separator.char
The character that separates each field in the form of an integer. Typically in a CSV this is a ,(44) character. A TSV would use a tab(9) character. If
ftps.csv.separator.char
is defined as a null(0), then the RFC 4180 parser must be utilized by default. This is the equivalent offtps.csv.rfc.4180.parser.enabled = true
.- Type: int
- Default: 44
- Importance: low
ftps.csv.case.sensitive.field.names
Flag to determine if the field names in the header row should be treated as case sensitive.
- Type: boolean
- Default: false
- Importance: low
ftps.csv.rfc.4180.parser.enabled
Flag to determine if the RFC 4180 parser should be used instead of the default parser.
- Type: boolean
- Default: false
- Importance: low
ftps.csv.escape.char
The character as an integer to use when a special character is encountered. The default escape character is typically a (92).
- Type: int
- Default: 92
- Importance: low
ftps.csv.file.charset
Character set to read the file with.
- Type: string
- Default: UTF-8
- Valid Values: Should be a valid charset such as UTF-8, UTF-16, IBM01147 etc.
- Importance: low
ftps.csv.ignore.leading.whitespace
Sets the ignore leading whitespace setting - if true, white space in front of a quote in a field is ignored.
- Type: boolean
- Default: true
- Importance: low
ftps.csv.ignore.quotations
Sets the ignore quotations mode - if true, quotations are ignored.
- Type: boolean
- Default: false
- Importance: low
ftps.csv.keep.carriage.return
Flag to determine if the carriage return at the end of the line should be maintained.
- Type: boolean
- Default: false
- Importance: low
ftps.csv.null.field.indicator
Indicator to determine how the CSV Reader can determine if a field is null. Valid values are EMPTY_SEPARATORS, EMPTY_QUOTES, BOTH, or NEITHER. For more information see http://opencsv.sourceforge.net/apidocs/com/opencsv/enums/CSVReaderNullFieldIndicator.html.
- Type: string
- Default: NEITHER
- Valid Values: one of [NEITHER, EMPTY_SEPARATORS, EMPTY_QUOTES, BOTH]
- Importance: low
ftps.csv.quote.char
The character that is used to quote a field. This typically happens when the ftps.csv.separator.char character is within the data.
- Type: int
- Default: 34
- Importance: low
ftps.csv.skip.lines
Number of lines to skip in the beginning of the file.
- Type: int
- Default: 0
- Importance: low
ftps.csv.strict.quotes
Sets the strict quotes setting - if true, characters outside the quotes are ignored.
- Type: boolean
- Default: false
- Importance: low
Confluent Licensing¶
confluent.license
Confluent will issue a license key to each subscriber. The license key will be a short snippet of text that you can copy and paste. Without the license key, you can use the connector for a 30-day trial period. If you are a subscriber, contact Confluent Support for more information.
- Type: string
- Default: “”
- Valid Values: Confluent Platform license
- Importance: high
confluent.topic.bootstrap.servers
A list of host/port pairs to use for establishing the initial connection to the Kafka cluster used for licensing. All servers in the cluster will be discovered from the initial connection. This list should be in the form
host1:port1,host2:port2,...
. These servers are used only for the initial connection to discover the full cluster membership, which may change dynamically, so this list need not contain the full set of servers. You may want more than one, in case a server is down.- Type: list
- Importance: high
confluent.topic
Name of the Kafka topic used for Confluent Platform configuration, including licensing information.
- Type: string
- Default: _confluent-command
- Importance: low
confluent.topic.replication.factor
The replication factor for the Kafka topic used for Confluent Platform configuration, including licensing information. This is used only if the topic does not already exist, and the default of 3 is appropriate for production use. If you are using a development environment with less than 3 brokers, you must set this to the number of brokers (often 1).
- Type: int
- Default: 3
- Importance: low
Confluent license properties¶
You can put license-related properties in the connector configuration, or starting with Confluent Platform version 6.0, you can put license-related properties in the Connect worker configuration instead of in each connector configuration.
This connector is proprietary and requires a license. The license information is stored in the _confluent-command
topic. If the broker requires SSL for connections, you must include the security-related confluent.topic.*
properties
as described below.
confluent.license
Confluent issues enterprise license keys to each subscriber. The license key is text that you can copy and paste as the value for
confluent.license
. A trial license allows using the connector for a 30-day trial period. A developer license allows using the connector indefinitely for single-broker development environments.If you are a subscriber, contact Confluent Support for more information.
- Type: string
- Default: “”
- Valid Values: Confluent Platform license
- Importance: high
confluent.topic.ssl.truststore.location
The location of the trust store file.
- Type: string
- Default: null
- Importance: high
confluent.topic.ssl.truststore.password
The password for the trust store file. If a password is not set access to the truststore is still available, but integrity checking is disabled.
- Type: password
- Default: null
- Importance: high
confluent.topic.ssl.keystore.location
The location of the key store file. This is optional for client and can be used for two-way authentication for client.
- Type: string
- Default: null
- Importance: high
confluent.topic.ssl.keystore.password
The store password for the key store file. This is optional for client and only needed if ssl.keystore.location is configured.
- Type: password
- Default: null
- Importance: high
confluent.topic.ssl.key.password
The password of the private key in the key store file. This is optional for client.
- Type: password
- Default: null
- Importance: high
confluent.topic.security.protocol
Protocol used to communicate with brokers. Valid values are: PLAINTEXT, SSL, SASL_PLAINTEXT, SASL_SSL.
- Type: string
- Default: “PLAINTEXT”
- Importance: medium
License topic configuration¶
A Confluent enterprise license is stored in the _confluent-command
topic.
This topic is created by default and contains the license that corresponds to
the license key supplied through the confluent.license
property. No public
keys are stored in Kafka topics.
The following describes how the default _confluent-command
topic is
generated under different scenarios:
- A 30-day trial license is automatically generated for the
_confluent command
topic if you do not add theconfluent.license
property or leave this property empty (for example,confluent.license=
). - Adding a valid license key (for example,
confluent.license=<valid-license-key>
) adds a valid license in the_confluent-command
topic.
Here is an example of the minimal properties for development and testing.
You can change the name of the _confluent-command
topic using the
confluent.topic
property (for instance, if your environment has strict
naming conventions). The example below shows this change and the configured
Kafka bootstrap server.
confluent.topic=foo_confluent-command
confluent.topic.bootstrap.servers=localhost:9092
The example above shows the minimally required bootstrap server property that
you can use for development and testing. For a production environment, you add
the normal producer, consumer, and topic configuration properties to the
connector properties, prefixed with confluent.topic.
.
Override Default Configuration Properties¶
You can override the replication factor using
confluent.topic.replication.factor
. For example, when using a Kafka cluster
as a destination with less than three brokers (for development and testing) you
should set the confluent.topic.replication.factor
property to 1
.
You can override producer-specific properties by using the
producer.override.*
prefix (for source connectors) and consumer-specific
properties by using the consumer.override.*
prefix (for sink connectors).
You can use the defaults or customize the other properties as well. For example,
the confluent.topic.client.id
property defaults to the name of the connector
with -licensing
suffix. You can specify the configuration settings for
brokers that require SSL or SASL for client connections using this prefix.
You cannot override the cleanup policy of a topic because the topic always has a single partition and is compacted. Also, do not specify serializers and deserializers using this prefix; they are ignored if added.