Configuration Reference for FTPS Source Connector for Confluent Platform

To use this connector, specify the name of the connector class in the connector.class configuration property.

connector.class=io.confluent.connect.ftps.FtpsSourceConnector

Source connector-specific configuration properties are described below.

General

kafka.topic

The Kafka topic to publish records to.

  • Type: string
  • Importance: high
batch.size

The number of records that should be returned with each batch.

  • Type: int
  • Default: 1000
  • Importance: medium
empty.poll.wait.ms

The amount of time in milliseconds to wait if a poll returns an empty list of records.

  • Type: long
  • Default: 10000
  • Importance: low

Auto topic creation

For more information about Auto topic creation, see Configuring Auto Topic Creation for Source Connectors.

Configuration properties accept regular expressions (regex) that are defined as Java regex.

topic.creation.groups

A list of group aliases that are used to define per-group topic configurations for matching topics. A default group always exists and matches all topics.

  • Type: List of String types
  • Default: empty
  • Possible Values: The values of this property refer to any additional groups. A default group is always defined for topic configurations.
topic.creation.$alias.replication.factor

The replication factor for new topics created by the connector. This value must not be larger than the number of brokers in the Kafka cluster. If this value is larger than the number of Kafka brokers, an error occurs when the connector attempts to create a topic. This is a required property for the default group. This property is optional for any other group defined in topic.creation.groups. Other groups use the Kafka broker default value.

  • Type: int
  • Default: n/a
  • Possible Values: >= 1 for a specific valid value or -1 to use the Kafka broker’s default value.
topic.creation.$alias.partitions

The number of topic partitions created by this connector. This is a required property for the default group. This property is optional for any other group defined in topic.creation.groups. Other groups use the Kafka broker default value.

  • Type: int
  • Default: n/a
  • Possible Values: >= 1 for a specific valid value or -1 to use the Kafka broker’s default value.
topic.creation.$alias.include

A list of strings that represent regular expressions that match topic names. This list is used to include topics with matching values, and apply this group’s specific configuration to the matching topics. $alias applies to any group defined in topic.creation.groups. This property does not apply to the default group.

  • Type: List of String types
  • Default: empty
  • Possible Values: Comma-separated list of exact topic names or regular expressions.
topic.creation.$alias.exclude

A list of strings representing regular expressions that match topic names. This list is used to exclude topics with matching values from getting the group’s specfic configuration. $alias applies to any group defined in topic.creation.groups. This property does not apply to the default group. Note that exclusion rules override any inclusion rules for topics.

  • Type: List of String types
  • Default: empty
  • Possible Values: Comma-separated list of exact topic names or regular expressions.
topic.creation.$alias.${kafkaTopicSpecificConfigName}

Any of the Changing Broker Configurations Dynamically for the version of the Kafka broker where the records will be written. The broker’s topic-level configuration value is used if the configuration is not specified for the rule. $alias applies to the default group as well as any group defined in topic.creation.groups.

  • Type: property values
  • Default: Kafka broker value

Schema

key.schema

The schema for the key written to Kafka.

  • Type: string
  • Default: {“name”:”default-key-schema”,”type”:”STRUCT”,”isOptional”:false,”fieldSchemas”:{“filePath”:{“type”:”STRING”,”isOptional”:true}}}
  • Importance: medium
value.schema

The schema for the value written to Kafka.

  • Type: string
  • Default: “”
  • Importance: medium
schema.generation.enabled

Flag to determine if schemas should be dynamically generated. If set to true, key.schema and value.schema can be omitted, but schema.generation.key.name and schema.generation.value.name must be set. Nested JSON is currently not supported and will be ignored. Schemaless JSON parsing should be used instead.

  • Type: boolean
  • Default: true
  • Importance: medium
schema.generation.key.fields

The field(s) to use to build a key schema. This is used only during schema generation.

  • Type: list
  • Default: “”
  • Importance: medium
schema.generation.key.name

The name of the generated key schema.

  • Type: string
  • Default: default_key_schema_name
  • Importance: low
schema.generation.value.name

The name of the generated value schema.

  • Type: string
  • Default: default_value_schema_name
  • Importance: low

Timestamp

ftps.timestamp.mode

Determines how the connector will set the timestamp for the Source Record. If set to FIELD, then the timestamp will be read from a field in the value. Specify the field in ftps.timestamp.field. If set to FILE_TIME, then the last modified time of the file will be used. If set to PROCESS_TIME, the time the record is read will be used.

  • Type: string
  • Default: PROCESS_TIME
  • Valid Values: one of [FIELD, FILE_TIME, PROCESS_TIME]
  • Importance: medium
ftps.timestamp.field

The field in the value schema that will contain the parsed timestamp for the record. This field cannot be marked as optional and must be a Timestamp.

  • Type: string
  • Default: “”
  • Importance: medium

File server connection

ftps.host

The address of the FTPS server.

  • Type: string
  • Importance: high
ftps.username

The username used to authenticate with the file server.

  • Type: string
  • Importance: high
ftps.password

The password used to authenticate with the file server.

  • Type: password
  • Importance: high
ftps.port

Port number of the FTPS server.

  • Type: int
  • Default: 21
  • Valid Values: [0,…]
  • Importance: medium
ftps.security.mode

The mode used to connect with the server. This is related to the establishment of the TLS connection and is up to the client.

  • Type: string
  • Default: EXPLICIT
  • Valid Values: one of [EXPLICIT, IMPLICIT]
  • Importance: low
ftps.server.timezone

The timezone on the FTPS server. This is used when calculating the age of a file.

  • Type: string
  • Default: UTC
  • Valid Values: Should be one of the identifiers used by java.util.TimeZone to refer to time zones, for example, America/Chicago or UTC.
  • Importance: low

File System

ftps.input.path

The FTPS directory to read files that will be processed.

  • Type: string
  • Importance: high
ftps.input.file.pattern

Regular expression to check input file names against. This expression must match the entire filename.

  • Type: string
  • Default: .+
  • Importance: high
ftps.error.path

The FTPS directory to place files that have errors.

  • Type: string
  • Default: “”
  • Importance: high
ftps.finished.path

The FTPS directory to place files that have been successfully processed.

  • Type: string
  • Default: “”
  • Importance: high
ftps.behavior.on.error

Specifies whether the task should halt when it encounters an error or continue to the next file.

  • Type: string
  • Default: LOG
  • Valid Values: one of [LOG, IGNORE, FAIL]
  • Importance: medium
ftps.schemaless.json.parsing

To specify whether JSON files should be parsed without using any schema. Needed only when JSON is one of file types specified in file.types.to.parse.

  • Type: boolean
  • Default: false
  • Importance: medium
ftps.ignore.file.extension

By default, the parser to use for a file is chosen based on the file’s extension. Instead, if a single parser is to be used for each file, regardless of extension, this property should be set to true. In that case, the parser specified in parser.type will be used for parsing all files.

  • Type: boolean
  • Default: false
  • Importance: medium
ftps.parser.type

The parser to use for parsing files. Required only if ftps.ignore.file.extension is true.

  • Type: string
  • Default: TXT
  • Valid Values: one of [TXT, CSV, JSON]
  • Importance: medium
ftps.cleanup.policy

Determines how the connector should cleanup the files that have been successfully processed. NONE leaves the files in place which could cause them to be reprocessed if the connector is restarted. DELETE removes the file from the filesystem. MOVE will move the file to a finished directory.

  • Type: string
  • Default: NONE
  • Valid Values: one of [DELETE, MOVE, NONE]
  • Importance: medium
ftps.file.minimum.age.ms

The amount of time in milliseconds after the file was last written to before the file can be processed.

  • Type: long
  • Default: 0
  • Valid Values: [0,…]
  • Importance: medium

Security

ftps.ssl.key.password

The password of the private key in the key store file. This is optional for client.

  • Type: password
  • Default: null
  • Importance: high
ftps.ssl.keystore.location

The location of the key store file. This is optional for client and can be used for two-way authentication for client.

  • Type: string
  • Default: null
  • Importance: high
ftps.ssl.keystore.password

The store password for the key store file. This is optional for client and only needed if ssl.keystore.location is configured.

  • Type: password
  • Default: null
  • Importance: high
ftps.ssl.truststore.location

The location of the truststore file.

  • Type: string
  • Default: null
  • Importance: high
ftps.ssl.truststore.password

The password for the truststore file. If a password is not set access to the truststore is still available, but integrity checking is disabled.

  • Type: password
  • Default: null
  • Importance: high
ftps.ssl.enabled.protocols

The list of protocols enabled for SSL connections.

  • Type: list
  • Default: TLSv1.2,TLSv1.1,TLSv1
  • Importance: medium
ftps.ssl.keystore.type

The file format of the key store file. This is optional for client.

  • Type: string
  • Default: JKS
  • Importance: medium
ftps.ssl.protocol

The SSL protocol used to generate the SSLContext. Default setting is TLS, which is fine for most cases. Allowed values in recent JVMs are TLS, TLSv1.1 and TLSv1.2. SSL, SSLv2 and SSLv3 may be supported in older JVMs, but their usage is discouraged due to known security vulnerabilities.

  • Type: string
  • Default: TLS
  • Importance: medium
ftps.ssl.provider

The name of the security provider used for SSL connections. Default value is the default security provider of the JVM.

  • Type: string
  • Default: null
  • Importance: medium
ftps.ssl.truststore.type

The file format of the truststore file.

  • Type: string
  • Default: JKS
  • Importance: medium
ftps.ssl.cipher.suites

A list of cipher suites. This is a named combination of authentication, encryption, MAC and key exchange algorithm used to negotiate the security settings for a network connection using TLS or SSL network protocol. By default all the available cipher suites are supported.

  • Type: list
  • Default: null
  • Importance: low
ftps.ssl.endpoint.identification.algorithm

The endpoint identification algorithm to validate server hostname using server certificate.

  • Type: string
  • Default: https
  • Importance: low
ftps.ssl.keymanager.algorithm

The algorithm used by key manager factory for SSL connections. Default value is the key manager factory algorithm configured for the Java Virtual Machine.

  • Type: string
  • Default: SunX509
  • Importance: low
ftps.ssl.secure.random.implementation

The SecureRandom PRNG implementation to use for SSL cryptography operations.

  • Type: string
  • Default: null
  • Importance: low
ftps.ssl.trustmanager.algorithm

The algorithm used by trust manager factory for SSL connections. Default value is the trust manager factory algorithm configured for the Java Virtual Machine.

  • Type: string
  • Default: PKIX
  • Importance: low

TXT configs

ftps.txt.file.charset

Character set to read a text file with.

  • Type: string
  • Default: UTF-8
  • Valid Values: Should be a valid charset such as UTF-8, UTF-16, IBM01147 etc.
  • Importance: low

CSV configs

ftps.csv.first.row.as.header

Flag to indicate if the fist row of data contains the header of the file. If true the position of the columns will be determined by the first row to the CSV. The column position will be inferred from the position of the schema supplied in value.schema. If set to true the number of columns must be greater than or equal to the number of fields in the schema and the first row, which is the header, will be used to set the field names in the generated schema if schema.generation.enabled is set to true.

  • Type: boolean
  • Default: false
  • Importance: medium
ftps.csv.separator.char

The character that separates each field in the form of an integer. Typically in a CSV this is a ,(44) character. A TSV would use a tab(9) character. If ftps.csv.separator.char is defined as a null(0), then the RFC 4180 parser must be utilized by default. This is the equivalent of ftps.csv.rfc.4180.parser.enabled = true.

  • Type: int
  • Default: 44
  • Importance: low
ftps.csv.case.sensitive.field.names

Flag to determine if the field names in the header row should be treated as case sensitive.

  • Type: boolean
  • Default: false
  • Importance: low
ftps.csv.rfc.4180.parser.enabled

Flag to determine if the RFC 4180 parser should be used instead of the default parser.

  • Type: boolean
  • Default: false
  • Importance: low
ftps.csv.escape.char

The character as an integer to use when a special character is encountered. The default escape character is typically a (92).

  • Type: int
  • Default: 92
  • Importance: low
ftps.csv.file.charset

Character set to read the file with.

  • Type: string
  • Default: UTF-8
  • Valid Values: Should be a valid charset such as UTF-8, UTF-16, IBM01147 etc.
  • Importance: low
ftps.csv.ignore.leading.whitespace

Sets the ignore leading whitespace setting - if true, white space in front of a quote in a field is ignored.

  • Type: boolean
  • Default: true
  • Importance: low
ftps.csv.ignore.quotations

Sets the ignore quotations mode - if true, quotations are ignored.

  • Type: boolean
  • Default: false
  • Importance: low
ftps.csv.keep.carriage.return

Flag to determine if the carriage return at the end of the line should be maintained.

  • Type: boolean
  • Default: false
  • Importance: low
ftps.csv.null.field.indicator

Indicator to determine how the CSV Reader can determine if a field is null. Valid values are EMPTY_SEPARATORS, EMPTY_QUOTES, BOTH, or NEITHER. For more information see http://opencsv.sourceforge.net/apidocs/com/opencsv/enums/CSVReaderNullFieldIndicator.html.

  • Type: string
  • Default: NEITHER
  • Valid Values: one of [NEITHER, EMPTY_SEPARATORS, EMPTY_QUOTES, BOTH]
  • Importance: low
ftps.csv.quote.char

The character that is used to quote a field. This typically happens when the ftps.csv.separator.char character is within the data.

  • Type: int
  • Default: 34
  • Importance: low
ftps.csv.skip.lines

Number of lines to skip in the beginning of the file.

  • Type: int
  • Default: 0
  • Importance: low
ftps.csv.strict.quotes

Sets the strict quotes setting - if true, characters outside the quotes are ignored.

  • Type: boolean
  • Default: false
  • Importance: low

Confluent Licensing

confluent.license

Confluent will issue a license key to each subscriber. The license key will be a short snippet of text that you can copy and paste. Without the license key, you can use the connector for a 30-day trial period. If you are a subscriber, contact Confluent Support for more information.

  • Type: string
  • Default: “”
  • Valid Values: Confluent Platform license
  • Importance: high
confluent.topic.bootstrap.servers

A list of host/port pairs to use for establishing the initial connection to the Kafka cluster used for licensing. All servers in the cluster will be discovered from the initial connection. This list should be in the form host1:port1,host2:port2,.... These servers are used only for the initial connection to discover the full cluster membership, which may change dynamically, so this list need not contain the full set of servers. You may want more than one, in case a server is down.

  • Type: list
  • Importance: high
confluent.topic

Name of the Kafka topic used for Confluent Platform configuration, including licensing information.

  • Type: string
  • Default: _confluent-command
  • Importance: low
confluent.topic.replication.factor

The replication factor for the Kafka topic used for Confluent Platform configuration, including licensing information. This is used only if the topic does not already exist, and the default of 3 is appropriate for production use. If you are using a development environment with less than 3 brokers, you must set this to the number of brokers (often 1).

  • Type: int
  • Default: 3
  • Importance: low

Confluent license properties

You can put license-related properties in the connector configuration, or starting with Confluent Platform version 6.0, you can put license-related properties in the Connect worker configuration instead of in each connector configuration.

This connector is proprietary and requires a license. The license information is stored in the _confluent-command topic. If the broker requires SSL for connections, you must include the security-related confluent.topic.* properties as described below.

confluent.license

Confluent issues enterprise license keys to each subscriber. The license key is text that you can copy and paste as the value for confluent.license. A trial license allows using the connector for a 30-day trial period. A developer license allows using the connector indefinitely for single-broker development environments.

If you are a subscriber, contact Confluent Support for more information.

  • Type: string
  • Default: “”
  • Valid Values: Confluent Platform license
  • Importance: high
confluent.topic.ssl.truststore.location

The location of the trust store file.

  • Type: string
  • Default: null
  • Importance: high
confluent.topic.ssl.truststore.password

The password for the trust store file. If a password is not set access to the truststore is still available, but integrity checking is disabled.

  • Type: password
  • Default: null
  • Importance: high
confluent.topic.ssl.keystore.location

The location of the key store file. This is optional for client and can be used for two-way authentication for client.

  • Type: string
  • Default: null
  • Importance: high
confluent.topic.ssl.keystore.password

The store password for the key store file. This is optional for client and only needed if ssl.keystore.location is configured.

  • Type: password
  • Default: null
  • Importance: high
confluent.topic.ssl.key.password

The password of the private key in the key store file. This is optional for client.

  • Type: password
  • Default: null
  • Importance: high
confluent.topic.security.protocol

Protocol used to communicate with brokers. Valid values are: PLAINTEXT, SSL, SASL_PLAINTEXT, SASL_SSL.

  • Type: string
  • Default: “PLAINTEXT”
  • Importance: medium

License topic configuration

A Confluent enterprise license is stored in the _confluent-command topic. This topic is created by default and contains the license that corresponds to the license key supplied through the confluent.license property. No public keys are stored in Kafka topics.

The following describes how the default _confluent-command topic is generated under different scenarios:

  • A 30-day trial license is automatically generated for the _confluent command topic if you do not add the confluent.license property or leave this property empty (for example, confluent.license=).
  • Adding a valid license key (for example, confluent.license=<valid-license-key>) adds a valid license in the _confluent-command topic.

Here is an example of the minimal properties for development and testing.

You can change the name of the _confluent-command topic using the confluent.topic property (for instance, if your environment has strict naming conventions). The example below shows this change and the configured Kafka bootstrap server.

confluent.topic=foo_confluent-command
confluent.topic.bootstrap.servers=localhost:9092

The example above shows the minimally required bootstrap server property that you can use for development and testing. For a production environment, you add the normal producer, consumer, and topic configuration properties to the connector properties, prefixed with confluent.topic..

Override Default Configuration Properties

You can override the replication factor using confluent.topic.replication.factor. For example, when using a Kafka cluster as a destination with less than three brokers (for development and testing) you should set the confluent.topic.replication.factor property to 1.

You can override producer-specific properties by using the producer.override.* prefix (for source connectors) and consumer-specific properties by using the consumer.override.* prefix (for sink connectors).

You can use the defaults or customize the other properties as well. For example, the confluent.topic.client.id property defaults to the name of the connector with -licensing suffix. You can specify the configuration settings for brokers that require SSL or SASL for client connections using this prefix.

You cannot override the cleanup policy of a topic because the topic always has a single partition and is compacted. Also, do not specify serializers and deserializers using this prefix; they are ignored if added.