Confluent Cloud Cluster Types¶
Some features are supported by all Confluent clusters, regardless of type. However, the Confluent Cloud cluster type you choose determines the full set of features, capabilities, and the price of the cluster.
Cluster types include:
Important
The capabilities provided in this topic are for planning purposes, and are not a guarantee of performance, which varies depending on each unique configuration.
For per-topic settings and limits, see Manage Kafka Cluster and Topic Configuration Settings in Confluent Cloud. For quotas that apply to organizations, environments, clusters, and accounts, see Service Quotas for Confluent Cloud. To monitor the performance of your clusters, see Metrics API.
All Confluent Cloud cluster types support the following features:
Feature | Details |
---|---|
Kafka ACLs | For details, see Use Access Control Lists (ACLs) for Confluent Cloud. |
Encryption at rest | Confluent Cloud uses encrypted volumes for all data storage at rest. |
Encryption in motion | Confluent Cloud requires SSL/TLS 1.2 for all connections. |
Exactly Once Semantics | For details, see Processing Guarantees. |
Fully-managed replica placement | Confluent Cloud manages replica placement, ensuring that replicas are spread across different brokers. For multi-zone clusters, Confluent Cloud ensures replicas are also spread across different Availability Zones. |
Key based compacted storage | For details, see Log Compaction. |
UI: Consumer lag | For details, see Monitor Consumer Lag in Confluent Cloud. |
Topic management | For details, see Create, Edit, and Delete Topics in Confluent Cloud. |
Uptime SLA | For details, see the Confluent Cloud Service Level Agreement . |
Basic clusters¶
Basic clusters are designed for development use-cases. Basic clusters support the following:
- 99.5% uptime SLA
- Up to 1 GB/s of throughput and 5 TB of storage.
- You only pay for the ingress, egress, storage, and partitions. There is no base cluster price.
- Can be upgraded to a single-zone Standard cluster at any time using the Confluent Cloud Console.
Note
You can view connector events in Confluent Cloud Console with a basic cluster, but you can’t consume events from a topic using Confluent CLI, Java, or C/C++. For more information, see View Connector Events.
Basic cluster limits per cluster¶
Dimension | Capability | Additional details |
---|---|---|
Ingress | Max 250 MBps | Number of bytes that can be produced to the cluster in one second. Available in the Metrics API as If you are self-managing Kafka, you can look at the producer outgoing-byte-rate metrics and broker kafka.server:type=BrokerTopicMetrics,name=BytesInPerSec metrics to understand your throughput. To reduce usage on this dimension, you can compress your messages.
|
Egress | Max 750 MBps | Number of bytes that can be consumed from the cluster in one second. Available in the the Metrics API as If you are self-managing Kafka, you can look at the consumer incoming-byte-rate metrics and broker kafka.server:type=BrokerTopicMetrics,name=BytesOutPerSec to understand your throughput. To reduce usage on this dimension, you can compress your messages
and ensure each consumer is only consuming from the topics it requires. |
Storage (pre-replication) | Max 5 TB | Number of bytes retained on the cluster, pre-replication. Available in the Metrics API as You can configure policy settings If you are self-managing Kafka, you can look at how much disk space your cluster is using to understand your storage needs. To reduce usage on this dimension, you can
compress your messages
and reduce your retention settings.
|
Partitions (pre-replication) | Max 4096 | Maximum number of partitions that can exist on the cluster at one time, before replication.
All topics that the customer creates as well as internal topics that are automatically created
by Confluent Platform components–such as ksqlDB, Kafka Streams, Connect, and Control Center–count towards the cluster
partition limit. The automatically created topics are prefixed with an underscore ( Available in the Metrics API as Attempts to create additional partitions beyond this limit will fail with an error message. If you are self-managing Kafka, you can look at the To reduce usage on this dimension, you can delete unused topics and create new topics with fewer partitions. You can use the Kafka Admin interface to increase the partition count of an existing topic if the initial partition count is too low. |
Total client connections | Max 1000 | Number of TCP connections to the cluster that can be open at one time. Available in the Metrics API as If you are self-managing Kafka, you can look at the broker kafka.server:type=socket-server-metrics,listener={listener_name},networkProcessor={#},name=connection-count metrics to understand how many connections you are using. This value can vary widely based on several factors, including number of producer clients, number of consumer clients, partition keying strategy, produce patterns per client, and consume patterns per client. To reduce usage on this dimension, you can reduce the total number of clients connecting to the cluster. |
Connection attempts | Max 80 per second | Maximum number of new TCP connections to the cluster that can be created in one second. This means successful authentications plus unsuccessful authentication attempts. Available in the Metrics API as If you are self-managing Kafka, you can look at the rate of change for the
kafka.server:type=socket-server-metrics,listener={listener_name},networkProcessor={#},name=connection-count metric and
the Consumer To reduce usage on this dimension, you can use longer-lived connections to the cluster. |
Requests | Max ~15000 per second | Number of client requests to the cluster in one second. Available in the the Metrics API as If you are self-managing Kafka, you can look at the broker kafka.network:type=RequestMetrics,name=RequestsPerSec,request={Produce FetchConsumer FetchFollower} metrics and client request-rate metrics to understand your request volume. To reduce usage on this dimension, you can adjust producer batching configurations, consumer client batching configurations, and shut down otherwise inactive clients. |
Message size | Max 8 MB | For message size defaults at the topic level see Custom topic settings for all cluster types. |
Client version | Minimum 0.11.0 | None |
Request size | Max 100 MB | None |
Fetch bytes | Max 55 MB | None |
API keys | Max 50 | None |
Partition creation and deletion | Max 250 per 5 minute period | The following occurs when partition creation and deletion rate limit is reached:
|
Connector tasks per Kafka cluster | Max 250 | 1 task per connector |
ACLs | Max 1000 | None |
Kafka REST Produce v3 | Max throughput: 10 MBps Max connection requests: 25 connection requests per second Max streamed requests: 1000 requests per second Max message size for Kafka REST Produce API: 1 MB |
Client applications can connect over the REST API to Produce records directly to the Confluent Cloud cluster. To learn more, see the Kafka REST concepts overview. To learn more about the Kafka REST Produce API streaming mode, see the examples and explanation of streaming mode in the concept docs, and the Produce example in the quick start. |
Basic cluster limits per partition¶
The partition capabilities that follow are based on benchmarking and intended as practical guidelines for planning purposes. Performance per partition will vary depending on your individual configuration, and these benchmarks do not guarantee performance.
Dimension | Capability |
---|---|
Ingress per Partition | ~5 MBps |
Egress per Partition | ~15 MBps |
Storage per Partition | Unlimited |
Storage per Partition for Compacted Topics | Unlimited |
Cluster Linking capabilities on a Basic cluster¶
- A Basic cluster can be a source cluster of a cluster link (can send data).
- A Basic cluster cannot be a destination cluster (cannot receive data).
To learn more, see Supported Cluster Types in the Cluster Linking documentation.
Standard clusters¶
Standard clusters are designed for production-ready features and functionality. Standard clusters support the following:
- Uptime SLA: 99.95% for Single-Zone, 99.99% for Multi-Zone
- Up to 1 GB/s of throughput and unlimited storage.
- Multi-zone high availability (optional). A multi-zone cluster is spread across three availability zones for better resiliency.
- Charged an hourly base price in addition to the ingress, egress, storage, and partitions.
Standard cluster limits per cluster¶
Dimension | Capability | Additional details |
---|---|---|
Ingress | Max 250 MBps | Number of bytes that can be produced to the cluster in one second. Available in the Metrics API as If you are self-managing Kafka, you can look at the producer outgoing-byte-rate metrics and broker kafka.server:type=BrokerTopicMetrics,name=BytesInPerSec metrics to understand your throughput. To reduce usage on this dimension, you can compress your messages.
|
Egress | Max 750 MBps | Number of bytes that can be consumed from the cluster in one second. Available in the the Metrics API as If you are self-managing Kafka, you can look at the consumer incoming-byte-rate metrics and broker kafka.server:type=BrokerTopicMetrics,name=BytesOutPerSec to understand your throughput. To reduce usage on this dimension, you can compress your messages
and ensure each consumer is only consuming from the topics it requires. |
Storage (pre-replication) | Infinite | Number of bytes retained on the cluster, pre-replication. Standard Confluent Cloud clusters support infinite storage. This means there is no maximum size limit for the amount of data that can be stored on the cluster. Available in the Metrics API as You can configure policy settings If you are self-managing Kafka, you can look at how much disk space your cluster is using to understand your storage needs. To reduce usage on this dimension, you can
compress your messages
and reduce your retention settings.
|
Partitions (pre-replication) | Max 4096 | Maximum number of partitions that can exist on the cluster at one time, before replication.
All topics that the customer creates as well as internal topics that are automatically created
by Confluent Platform components–such as ksqlDB, Kafka Streams, Connect, and Control Center–count towards the cluster
partition limit. The automatically created topics are prefixed with an underscore ( Available in the Metrics API as Attempts to create additional partitions beyond this limit will fail with an error message. If you are self-managing Kafka, you can look at the To reduce usage on this dimension, you can delete unused topics and create new topics with fewer partitions. You can use the Kafka Admin interface to increase the partition count of an existing topic if the initial partition count is too low. |
Total client connections | Max 1000 | Number of TCP connections to the cluster that can be open at one time. Available in the Metrics API as If you are self-managing Kafka, you can look at the broker kafka.server:type=socket-server-metrics,listener={listener_name},networkProcessor={#},name=connection-count metrics to understand how many connections you are using. This value can vary widely based on several factors, including number of producer clients, number of consumer clients, partition keying strategy, produce patterns per client, and consume patterns per client. To reduce usage on this dimension, you can reduce the total number of clients connecting to the cluster. |
Connection attempts | Max 80 per second | Maximum number of new TCP connections to the cluster that can be created in one second. This means successful authentications plus unsuccessful authentication attempts. Available in the Metrics API as If you are self-managing Kafka, you can look at the rate of change for the
kafka.server:type=socket-server-metrics,listener={listener_name},networkProcessor={#},name=connection-count metric and
the Consumer To reduce usage on this dimension, you can use longer-lived connections to the cluster. |
Requests | Max ~15000 per second | Number of client requests to the cluster in one second. Available in the the Metrics API as If you are self-managing Kafka, you can look at the broker kafka.network:type=RequestMetrics,name=RequestsPerSec,request={Produce FetchConsumer FetchFollower} metrics and client request-rate metrics to understand your request volume. To reduce usage on this dimension, you can adjust producer batching configurations, consumer client batching configurations, and shut down otherwise inactive clients. |
Message size | Max 8 MB | For message size defaults at the topic level, see Custom topic settings for all cluster types. |
Client version | Minimum 0.11.0 | None |
Request size | Max 100 MB | None |
Fetch bytes | Max 55 MB | None |
API keys | Max 100 | To request an increase for this limit, contact Confluent Support. |
Partition creation and deletion | Max 500 per 5 minute period | The following occurs when partition creation and deletion rate limit is reached:
|
Connector tasks per Kafka cluster | Max 250 | None |
ACLs | Max 1000 | None |
Kafka REST Produce v3 | Max throughput: 10 MBps Max connection requests: 25 connection requests per second Max streamed requests: 1000 requests per second Max message size for Kafka REST Produce API: 1 MB |
Client applications can connect over the REST API to Produce records directly to the Confluent Cloud cluster. To learn more, see the Kafka REST concepts overview. To learn more about the Kafka REST Produce API streaming mode, see the examples and explanation of streaming mode in the concept docs, and the Produce example in the quick start. |
Standard cluster limits per partition¶
The partition capabilities that follow are based on benchmarking and intended as practical guidelines for planning purposes. Performance per partition will vary depending on your individual configuration, and these benchmarks do not guarantee performance.
Dimension | Capability |
---|---|
Ingress per Partition | ~5 MBps |
Egress per Partition | ~15 MBps |
Storage per Partition | Unlimited |
Storage per Partition for Compacted Topics | Unlimited |
Cluster Linking capabilities on a Standard cluster¶
- A Standard cluster can be a source cluster of a cluster link (can send data).
- A Standard cluster cannot be a destination cluster (cannot receive data).
To learn more, see Supported Cluster Types in the Cluster Linking documentation.
Dedicated clusters¶
Dedicated clusters are designed for critical production workloads with high traffic or private networking requirements. Dedicated clusters support the following:
- Single-tenant deployments with a 99.95% uptime SLA for Single-Zone, and 99.99% for Multi-Zone
- Private networking options including VPC peering, AWS Transit Gateway, AWS PrivateLink, and Azure PrivateLink.
- Self-managed keys when AWS, Azure, or Google Cloud is the cloud service provider.
- Multi-zone high availability (optional). A multi-zone cluster is spread across three availability zones for better resiliency.
- Can be scaled to achieve gigabytes per second of ingress.
- Simple scaling in terms of CKUs.
- Cluster expansion, and Cluster shrinking.
- Cluster Linking for fully-managed replication (multiregion, multicloud, hybrid cloud, and inter-organization).
Dedicated clusters are provisioned and billed in terms of Confluent Unit for Kafka (CKU). CKUs are a unit of horizontal scalability in Confluent Cloud that provide a pre-allocated amount of resources. How much you can ingest and stream per CKU depends on a variety of factors including client application design and partitioning strategy.
CKU limits per cluster¶
Dedicated clusters can be purchased in any whole number of CKUs up to a limit.
- For organizations with credit card billing, the upper limit is 4 CKUs per dedicated cluster. Clusters up to 100 CKUs are available by request.
- For organizations with integrated cloud provider billing or payment using an invoice, the upper limit is 24 CKUs per dedicated cluster. Clusters up to 100 CKUs are available by request.
To provision a cluster with more CKUs than the upper limit, contact Confluent Support.
Single-zone clusters can have 1 or more CKUs, whereas multi-zone clusters, which are spread across three availability zones, require a minimum of 2 CKUs. Zone availability cannot be changed after the cluster is created.
Limits per CKU¶
CKUs determine the capacity of your cluster. For a Confluent Cloud cluster, the expected performance for any given workload is dependent on a variety of dimensions, such as message size and number of partitions.
There are two categories of CKU dimensions:
- Dimensions with a fixed limit that cannot be exceeded.
- Dimensions with a more flexible guideline that may be exceeded depending on the overall cluster load.
The recommended guideline for a dimension is calculated for a workload optimized across the dimensions, enabling high levels of CKU utilization as measured by the cluster load metric. You may exceed the recommended guideline for a dimension, and achieve higher performance for that dimension, usually only if your usage of other dimensions is less than the recommended guideline or fixed limit.
Also note that usage patterns across all dimensions affect the workload and you may not achieve the suggested guideline for a particular dimension. For example, if you reach the partition limit, you will not likely reach the maximum CKU throughput guideline.
You should monitor the cluster load metric for your cluster to see how your usage pattern correlates with cluster utilization.
When a cluster’s load metric is high, the cluster may delay new connections and/or throttle clients in an attempt to ensure the cluster remains available. This throttling would register as non-zero values for the producer client produce-throttle-time-max and produce-throttle-time-avg metrics and consumer client fetch-throttle-time-max and fetch-throttle-time-avg metrics.
Use the information in the following tables to determine the minimum number of CKUs to use for a given workload, how to monitor a dimension and suggestions to reduce your use of a particular dimension.
Dimensions with fixed limits¶
The following table lists dimensions that have a fixed maximum limit that cannot be exceeded.
Dimension | Maximum per CKU | Details |
---|---|---|
Storage (pre-replication) | Infinite | Number of bytes retained on the cluster, pre-replication. Dedicated clusters have infinite storage, which means there is no maximum size limit for the amount of data that can be stored on the cluster. Available in the Metrics API as You can configure policy settings If you are self-managing Kafka, you can look at how much disk space your cluster is using to understand your storage needs. To reduce usage on this dimension, you can
compress your messages
and reduce your retention settings.
|
Partitions (pre-replication) | 4,500 partitions (100,000 partitions maximum across all CKUs) | Maximum number of partitions that can exist on the cluster at one time, before replication.
All topics that the customer creates as well as internal topics that are automatically created
by Confluent Platform components–such as ksqlDB, Kafka Streams, Connect, and Control Center–count towards the cluster
partition limit. The automatically created topics are prefixed with an underscore ( Available in the Metrics API as Attempts to create additional partitions beyond this limit will fail with an error message. If you are self-managing Kafka, you can look at the To reduce usage on this dimension, you can delete unused topics and create new topics with fewer partitions. You can use the Kafka Admin interface to increase the partition count of an existing topic if the initial partition count is too low. |
Connection attempts | 500 connection attempts per second | Maximum number of new TCP connections to the cluster that can be created in one second. This means successful authentications plus unsuccessful authentication attempts. If you exceed the maximum, connection attempts may be refused. If you are self-managing Kafka, you can look at the rate of change for the
kafka.server:type=socket-server-metrics,listener={listener_name},networkProcessor={#},name=connection-count metric and
the Consumer |
Kafka REST Produce v3 | Max throughput: Max 50 MBps Max connection requests: Max 300 connection requests per second Max streamed requests: Max 3000 requests per second |
Client applications can connect over the REST API to Produce records directly to the Confluent Cloud cluster. To learn more, see the Kafka REST concepts overview. To learn more about the Kafka REST Produce API streaming mode, see the examples and explanation of streaming mode in the concept docs, and the Produce example in the quick start. |
Dimensions with a recommended guideline¶
The following table lists dimensions with a recommended guideline.
These dimensions provide guidelines for capacity planning. The ability to fully utilize these dimensions depend on the workload and utilization of other dimensions. See more about measuring load in cluster load metric and for the maximum bandwidth for each cloud provider (AWS, GCP, Azure), are available in Benchmark Your Dedicated Apache Kafka Cluster on Confluent Cloud.
Dimension | Guideline per CKU | Details |
---|---|---|
Ingress | ~50 megabytes per second (MBps) | Number of bytes that can be produced to the cluster in one second. Available in the Metrics API as If you are self-managing Kafka, you can look at the producer outgoing-byte-rate metrics and broker kafka.server:type=BrokerTopicMetrics,name=BytesInPerSec metrics to understand your throughput. To reduce usage on this dimension, you can compress your messages.
|
Egress | ~150 megabytes per second (MBps) | Number of bytes that can be consumed from the cluster in one second. Available in the the Metrics API as If you are self-managing Kafka, you can look at the consumer incoming-byte-rate metrics and broker kafka.server:type=BrokerTopicMetrics,name=BytesOutPerSec to understand your throughput. To reduce usage on this dimension, you can compress your messages
and ensure each consumer is only consuming from the topics it requires. |
Total client connections | ~9,000 connections | Number of TCP connections to the cluster that can be open at one time. Available in the Metrics API as If you are self-managing Kafka, you can look at the broker kafka.server:type=socket-server-metrics,listener={listener_name},networkProcessor={#},name=connection-count metrics to understand how many connections you are using. This value can vary widely based on several factors, including number of producer clients, number of consumer clients, partition keying strategy, produce patterns per client, and consume patterns per client. To reduce usage on this dimension, you can reduce the total number of clients connecting to the cluster. A high number of client connections results in increased load on the cluster. |
Requests | ~15,000 requests per second | Number of client requests to the cluster in one second. Available in the the Metrics API as If you are self-managing Kafka, you can look at the broker kafka.network:type=RequestMetrics,name=RequestsPerSec,request={Produce FetchConsumer FetchFollower} metrics and client request-rate metrics to understand your request volume. To reduce usage on this dimension, you can adjust producer batching configurations, consumer client batching configurations, and shut down otherwise inactive clients. A high number of requests per second results in increased load on the cluster. |
Dedicated cluster limits per cluster¶
You can add CKUs to a dedicated cluster to meet the capacity for your high traffic workloads. However, the limits shown in this table will not change as you increase the number of CKUs. For more information about quotas and limits, see Service Quotas for Confluent Cloud and Custom topic settings for all cluster types.
Dimension | Capability | Additional details |
---|---|---|
Message size | Max 20 MB | For message size defaults at the topic level, see Custom topic settings for all cluster types. |
Client version | Minimum 0.11.0 | None |
Request size | Max 100 MB | None |
Fetch bytes | Max 55 MB | None |
Partitions | Depends on number of CKUs, absolute max 100,000 | None |
API keys | Max 2000 | To request an increase for this limit, contact Confluent Support. |
Partition creation and deletion | Max 5000 per 5 minute period | The following occurs when partition creation and deletion rate limit is reached:
|
Connector tasks per Kafka cluster | Max 250 | None |
ACLs | Max 10000 | None |
Dedicated cluster limits per partition¶
You can add CKUs to a dedicated cluster to meet the capacity for your high traffic workloads. However, the limits shown in this table will not change as you increase the number of CKUs.
The partition capabilities that follow are based on benchmarking and intended as practical guidelines for planning purposes. Performance per partition will vary depending on your individual configuration, and these benchmarks do not guarantee performance.
Dimension | Capability |
---|---|
Ingress per Partition | Max 10 MBps (aggregate producer throughput) |
Egress per Partition | Max 30 MBps (aggregate consumer throughput) |
Storage per Partition | Unlimited |
Storage per Partition for Compacted Topics | Unlimited |
Dedicated cluster provisioning time¶
The estimated time to initially provision a dedicated cluster depends on the cluster’s size and your choice of cloud provider. Regardless of the cloud provider, sometimes provisioning can take 24 hours or more. Contact Confluent Support if provisioning takes longer than 24 hours.
Note that provisioning time is excluded from the Confluent SLA.
- AWS: the expected provisioning time is one hour per CKU. A dedicated 2 CKU cluster on AWS is estimated to complete provisioning in about two hours.
- GCP: the expected provisioning time is one hour per CKU. A dedicated 2 CKU cluster on GCP is estimated to complete provisioning in two hours.
- Azure: the expected provisioning time is two and a half hours per CKU. A dedicated 2 CKU cluster on Azure is estimated to complete provisioning in five hours.
You will receive an email when provisioning is complete.
Dedicated cluster resizing time¶
When a cluster is not heavily loaded, expansion and shrink times are generally similar to expected provisioning times (single-digit hours). When a cluster is under heavy load, resizing can take longer.
Resizing a cluster should take about 1-2 hours per CKU.
During a resize operation, your applications may see leader elections, but otherwise performance will not suffer. Supported Kafka clients will gracefully handle these changes. You will receive an email when the resize operation is complete.
Cluster Linking capabilities on a Dedicated cluster¶
A Dedicated cluster can be a source or destination of a cluster link, dependent on the networking type and the other cluster involved. To learn more, see Supported Cluster Types in the Cluster Linking documentation.