Important
You are viewing documentation for an older version of Confluent Platform. For the latest, click here.
FTPS Source Connector Configuration Properties¶
To use this connector, specify the name of the connector class in the connector.class
configuration property.
connector.class=io.confluent.connect.ftps.FtpsSourceConnector
Connector-specific configuration properties are described below.
General¶
kafka.topic
The Kafka topic to publish records to.
- Type: string
- Importance: high
batch.size
The number of records that should be returned with each batch.
- Type: int
- Default: 1000
- Importance: medium
empty.poll.wait.ms
The amount of time in milliseconds to wait if a poll returns an empty list of records.
- Type: long
- Default: 10000
- Importance: low
Schema¶
key.schema
The schema for the key written to Kafka.
- Type: string
- Default: {“name”:”default-key-schema”,”type”:”STRUCT”,”isOptional”:false,”fieldSchemas”:{“filePath”:{“type”:”STRING”,”isOptional”:true}}}
- Importance: medium
value.schema
The schema for the value written to Kafka.
- Type: string
- Default: “”
- Importance: medium
schema.generation.enabled
Flag to determine if schemas should be dynamically generated. If set to true,
key.schema
andvalue.schema
can be omitted, butschema.generation.key.name
andschema.generation.value.name
must be set. Nested JSON is currently not supported and will be ignored. Schemaless JSON parsing should be used instead.- Type: boolean
- Default: true
- Importance: medium
schema.generation.key.fields
The field(s) to use to build a key schema. This is used only during schema generation.
- Type: list
- Default: “”
- Importance: medium
schema.generation.key.name
The name of the generated key schema.
- Type: string
- Default: default_key_schema_name
- Importance: low
schema.generation.value.name
The name of the generated value schema.
- Type: string
- Default: default_value_schema_name
- Importance: low
Timestamp¶
ftps.timestamp.mode
Determines how the connector will set the timestamp for the Source Record. If set to
FIELD
, then the timestamp will be read from a field in the value. Specify the field inftps.timestamp.field
. If set toFILE_TIME
, then the last modified time of the file will be used. If set toPROCESS_TIME
, the time the record is read will be used.- Type: string
- Default: PROCESS_TIME
- Valid Values: one of [FIELD, FILE_TIME, PROCESS_TIME]
- Importance: medium
ftps.timestamp.field
The field in the value schema that will contain the parsed timestamp for the record. This field cannot be marked as optional and must be a Timestamp.
- Type: string
- Default: “”
- Importance: medium
File server connection¶
ftps.host
The address of the FTPS server.
- Type: string
- Importance: high
ftps.username
The username used to authenticate with the file server.
- Type: string
- Importance: high
ftps.password
The password used to authenticate with the file server.
- Type: password
- Importance: high
ftps.port
Port number of the FTPS server.
- Type: int
- Default: 21
- Valid Values: [0,…]
- Importance: medium
ftps.security.mode
The mode used to connect with the server. This is related to the establishment of the TLS connection and is up to to the client.
- Type: string
- Default: EXPLICIT
- Valid Values: one of [EXPLICIT, IMPLICIT]
- Importance: low
ftps.server.timezone
The timezone on the FTPS server. This is used when calculating the age of a file.
- Type: string
- Default: UTC
- Valid Values: Should be one of the identifiers used by
java.util.TimeZone
to refer to time zones, for example, America/Chicago or UTC. - Importance: low
File System¶
ftps.input.path
The FTPS directory to read files that will be processed.
- Type: string
- Importance: high
ftps.input.file.pattern
Regular expression to check input file names against. This expression must match the entire filename.
- Type: string
- Default: .+
- Importance: high
ftps.error.path
The FTPS directory to place files that have errors.
- Type: string
- Default: “”
- Importance: high
ftps.finished.path
The FTPS directory to place files that have been successfully processed.
- Type: string
- Default: “”
- Importance: high
ftps.behavior.on.error
Specifies whether the task should halt when it encounters an error or continue to the next file.
- Type: string
- Default: LOG
- Valid Values: one of [LOG, IGNORE, FAIL]
- Importance: medium
ftps.schemaless.json.parsing
To specify whether JSON files should be parsed without using any schema. Needed only when JSON is one of file types specified in file.types.to.parse.
- Type: boolean
- Default: false
- Importance: medium
ftps.ignore.file.extension
By default, the parser to use for a file is chosen based on the file’s extension. Instead, if a single parser is to be used for each file, regardless of extension, this property should be set to true. In that case, the parser specified in parser.type will be used for parsing all files.
- Type: boolean
- Default: false
- Importance: medium
ftps.parser.type
The parser to use for parsing files. Required only if
ftps.ignore.file.extension
istrue
.- Type: string
- Default: TXT
- Valid Values: one of [TXT, CSV, JSON]
- Importance: medium
ftps.cleanup.policy
Determines how the connector should cleanup the files that have been successfully processed. NONE leaves the files in place which could cause them to be reprocessed if the connector is restarted. DELETE removes the file from the filesystem. MOVE will move the file to a finished directory.
- Type: string
- Default: NONE
- Valid Values: one of [DELETE, MOVE, NONE]
- Importance: medium
ftps.file.minimum.age.ms
The amount of time in milliseconds after the file was last written to before the file can be processed.
- Type: long
- Default: 0
- Valid Values: [0,…]
- Importance: medium
Security¶
ftps.ssl.key.password
The password of the private key in the key store file. This is optional for client.
- Type: password
- Default: null
- Importance: high
ftps.ssl.keystore.location
The location of the key store file. This is optional for client and can be used for two-way authentication for client.
- Type: string
- Default: null
- Importance: high
ftps.ssl.keystore.password
The store password for the key store file. This is optional for client and only needed if
ssl.keystore.location
is configured.- Type: password
- Default: null
- Importance: high
ftps.ssl.truststore.location
The location of the truststore file.
- Type: string
- Default: null
- Importance: high
ftps.ssl.truststore.password
The password for the truststore file. If a password is not set access to the truststore is still available, but integrity checking is disabled.
- Type: password
- Default: null
- Importance: high
ftps.ssl.enabled.protocols
The list of protocols enabled for SSL connections.
- Type: list
- Default: TLSv1.2,TLSv1.1,TLSv1
- Importance: medium
ftps.ssl.keystore.type
The file format of the key store file. This is optional for client.
- Type: string
- Default: JKS
- Importance: medium
ftps.ssl.protocol
The SSL protocol used to generate the SSLContext. Default setting is TLS, which is fine for most cases. Allowed values in recent JVMs are TLS, TLSv1.1 and TLSv1.2. SSL, SSLv2 and SSLv3 may be supported in older JVMs, but their usage is discouraged due to known security vulnerabilities.
- Type: string
- Default: TLS
- Importance: medium
ftps.ssl.provider
The name of the security provider used for SSL connections. Default value is the default security provider of the JVM.
- Type: string
- Default: null
- Importance: medium
ftps.ssl.truststore.type
The file format of the truststore file.
- Type: string
- Default: JKS
- Importance: medium
ftps.ssl.cipher.suites
A list of cipher suites. This is a named combination of authentication, encryption, MAC and key exchange algorithm used to negotiate the security settings for a network connection using TLS or SSL network protocol. By default all the available cipher suites are supported.
- Type: list
- Default: null
- Importance: low
ftps.ssl.endpoint.identification.algorithm
The endpoint identification algorithm to validate server hostname using server certificate.
- Type: string
- Default: https
- Importance: low
ftps.ssl.keymanager.algorithm
The algorithm used by key manager factory for SSL connections. Default value is the key manager factory algorithm configured for the Java Virtual Machine.
- Type: string
- Default: SunX509
- Importance: low
ftps.ssl.secure.random.implementation
The SecureRandom PRNG implementation to use for SSL cryptography operations.
- Type: string
- Default: null
- Importance: low
ftps.ssl.trustmanager.algorithm
The algorithm used by trust manager factory for SSL connections. Default value is the trust manager factory algorithm configured for the Java Virtual Machine.
- Type: string
- Default: PKIX
- Importance: low
TXT configs¶
ftps.txt.file.charset
Character set to read a text file with.
- Type: string
- Default: UTF-8
- Valid Values: Should be a valid charset such as UTF-8, UTF-16, IBM01147 etc.
- Importance: low
CSV configs¶
ftps.csv.first.row.as.header
Flag to indicate if the fist row of data contains the header of the file. If true the position of the columns will be determined by the first row to the CSV. The column position will be inferred from the position of the schema supplied in
value.schema
. If set to true the number of columns must be greater than or equal to the number of fields in the schema and the first row, which is the header, will be used to set the field names in the generated schema ifschema.generation.enabled
is set to true.- Type: boolean
- Default: false
- Importance: medium
ftps.csv.separator.char
The character that separates each field in the form of an integer. Typically in a CSV this is a ,(44) character. A TSV would use a tab(9) character. If
ftps.csv.separator.char
is defined as a null(0), then the RFC 4180 parser must be utilized by default. This is the equivalent offtps.csv.rfc.4180.parser.enabled = true
.- Type: int
- Default: 44
- Importance: low
ftps.csv.case.sensitive.field.names
Flag to determine if the field names in the header row should be treated as case sensitive.
- Type: boolean
- Default: false
- Importance: low
ftps.csv.rfc.4180.parser.enabled
Flag to determine if the RFC 4180 parser should be used instead of the default parser.
- Type: boolean
- Default: false
- Importance: low
ftps.csv.escape.char
The character as an integer to use when a special character is encountered. The default escape character is typically a (92).
- Type: int
- Default: 92
- Importance: low
ftps.csv.file.charset
Character set to read the file with.
- Type: string
- Default: UTF-8
- Valid Values: Should be a valid charset such as UTF-8, UTF-16, IBM01147 etc.
- Importance: low
ftps.csv.ignore.leading.whitespace
Sets the ignore leading whitespace setting - if true, white space in front of a quote in a field is ignored.
- Type: boolean
- Default: true
- Importance: low
ftps.csv.ignore.quotations
Sets the ignore quotations mode - if true, quotations are ignored.
- Type: boolean
- Default: false
- Importance: low
ftps.csv.keep.carriage.return
Flag to determine if the carriage return at the end of the line should be maintained.
- Type: boolean
- Default: false
- Importance: low
ftps.csv.null.field.indicator
Indicator to determine how the CSV Reader can determine if a field is null. Valid values are EMPTY_SEPARATORS, EMPTY_QUOTES, BOTH, or NEITHER. For more information see http://opencsv.sourceforge.net/apidocs/com/opencsv/enums/CSVReaderNullFieldIndicator.html.
- Type: string
- Default: NEITHER
- Valid Values: one of [NEITHER, EMPTY_SEPARATORS, EMPTY_QUOTES, BOTH]
- Importance: low
ftps.csv.quote.char
The character that is used to quote a field. This typically happens when the ftps.csv.separator.char character is within the data.
- Type: int
- Default: 34
- Importance: low
ftps.csv.skip.lines
Number of lines to skip in the beginning of the file.
- Type: int
- Default: 0
- Importance: low
ftps.csv.strict.quotes
Sets the strict quotes setting - if true, characters outside the quotes are ignored.
- Type: boolean
- Default: false
- Importance: low
Confluent Licensing¶
confluent.license
Confluent will issue a license key to each subscriber. The license key will be a short snippet of text that you can copy and paste. Without the license key, you can use the connector for a 30-day trial period. If you are a subscriber, please contact Confluent Support for more information.
- Type: string
- Default: “”
- Valid Values: Confluent Platform license
- Importance: high
confluent.topic.bootstrap.servers
A list of host/port pairs to use for establishing the initial connection to the Kafka cluster used for licensing. All servers in the cluster will be discovered from the initial connection. This list should be in the form
host1:port1,host2:port2,...
. These servers are used only for the initial connection to discover the full cluster membership, which may change dynamically, so this list need not contain the full set of servers. You may want more than one, in case a server is down.- Type: list
- Importance: high
confluent.topic
Name of the Kafka topic used for Confluent Platform configuration, including licensing information.
- Type: string
- Default: _confluent-command
- Importance: low
confluent.topic.replication.factor
The replication factor for the Kafka topic used for Confluent Platform configuration, including licensing information. This is used only if the topic does not already exist, and the default of 3 is appropriate for production use. If you are using a development environment with less than 3 brokers, you must set this to the number of brokers (often 1).
- Type: int
- Default: 3
- Importance: low
Confluent license properties¶
Note
This connector is proprietary and requires a license. The license information
is stored in the _confluent-command
topic. If the broker requires SSL for
connections, you must include the security-related confluent.topic.*
properties as described below.
confluent.license
Confluent issues enterprise license keys to each subscriber. The license key is text that you can copy and paste as the value for
confluent.license
. A trial license allows using the connector for a 30-day trial period. A developer license allows using the connector indefinitely for single-broker development environments.If you are a subscriber, please contact Confluent Support for more information.
- Type: string
- Default: “”
- Valid Values: Confluent Platform license
- Importance: high
confluent.topic.ssl.truststore.location
The location of the trust store file.
- Type: string
- Default: null
- Importance: high
confluent.topic.ssl.truststore.password
The password for the trust store file. If a password is not set access to the truststore is still available, but integrity checking is disabled.
- Type: password
- Default: null
- Importance: high
confluent.topic.ssl.keystore.location
The location of the key store file. This is optional for client and can be used for two-way authentication for client.
- Type: string
- Default: null
- Importance: high
confluent.topic.ssl.keystore.password
The store password for the key store file. This is optional for client and only needed if ssl.keystore.location is configured.
- Type: password
- Default: null
- Importance: high
confluent.topic.ssl.key.password
The password of the private key in the key store file. This is optional for client.
- Type: password
- Default: null
- Importance: high
confluent.topic.security.protocol
Protocol used to communicate with brokers. Valid values are: PLAINTEXT, SSL, SASL_PLAINTEXT, SASL_SSL.
- Type: string
- Default: “PLAINTEXT”
- Importance: medium
License topic configuration¶
A Confluent enterprise license is stored in the _confluent-command
topic.
This topic is created by default and contains the license that corresponds to
the license key supplied through the confluent.license
property.
Note
No public keys are stored in Kafka topics.
The following describes how the default _confluent-command
topic is
generated under different scenarios:
- A 30-day trial license is automatically generated for the
_confluent command
topic if you do not add theconfluent.license
property or leave this property empty (for example,confluent.license=
). - Adding a valid license key (for example,
confluent.license=<valid-license-key>
) adds a valid license in the_confluent-command
topic.
Here is an example of the minimal properties for development and testing.
You can change the name of the _confluent-command
topic using the
confluent.topic
property (for instance, if your environment has strict
naming conventions). The example below shows this change and the configured
Kafka bootstrap server.
confluent.topic=foo_confluent-command
confluent.topic.bootstrap.servers=localhost:9092
The example above shows the minimally required bootstrap server property that
you can use for development and testing. For a production environment, you add
the normal producer, consumer, and topic configuration properties to the
connector properties, prefixed with confluent.topic.
.
Overriding Default Configuration Properties¶
You can override the replication factor using
confluent.topic.replication.factor
. For example, when using a Kafka cluster
as a destination with less than three brokers (for development and testing) you
should set the confluent.topic.replication.factor
property to 1
.
You can override producer-specific properties by using the
confluent.topic.producer.
prefix and consumer-specific properties by using
the confluent.topic.consumer.
prefix.
You can use the defaults or customize the other properties as well. For example,
the confluent.topic.client.id
property defaults to the name of the connector
with -licensing
suffix. You can specify the configuration settings for
brokers that require SSL or SASL for client connections using this prefix.
You cannot override the cleanup policy of a topic because the topic always has a single partition and is compacted. Also, do not specify serializers and deserializers using this prefix; they are ignored if added.