Confluent Certified Administrator for Apache Kafka
Last Update Oct 15, 2025
Total Questions : 54
We are offering FREE CCAAK Confluent exam questions. All you do is to just go and sign up. Give your details, prepare CCAAK free exam questions and then go for complete pool of Confluent Certified Administrator for Apache Kafka test questions that will help you more.
A topic ’recurring payments’ is created on a Kafka cluster with three brokers (broker id '0', ’1’, ‘2’) and nine partitions. The min.insync.replicas is set to three, and the producer is set with acks=all. Kafka Broker with id '0' is down.
Which statement is correct?
If the Controller detects the failure of a broker that was the leader for some partitions, which actions will be taken? (Choose two.)
A developer is working for a company with internal best practices that dictate that there is no single point of failure for all data stored.
What is the best approach to make sure the developer is complying with this best practice when creating Kafka topics?
You have an existing topic t1 that you want to delete because there are no more producers writing to it or consumers reading from it.
What is the recommended way to delete the topic?
Kafka broker supports which Simple Authentication and Security Layer (SASL) mechanisms for authentication? (Choose three.)
An employee in the reporting department needs assistance because their data feed is slowing down. You start by quickly checking the consumer lag for the clients on the data stream.
Which command will allow you to quickly check for lag on the consumers?
Your organization has a mission-critical Kafka cluster that must be highly available. A Disaster Recovery (DR) cluster has been set up using Replicator, and data is continuously being replicated from source cluster to the DR cluster. However, you notice that the message on offset 1002 on source cluster does not seem to match with offset 1002 on the destination DR cluster.
Which statement is correct?
In certain scenarios, it is necessary to weigh the trade-off between latency and throughput. One method to increase throughput is to configure batching of messages.
In addition to batch.size, what other producer property can be used to accomplish this?
Kafka Connect is running on a two node cluster in distributed mode. The connector is a source connector that pulls data from Postgres tables (users/payment/orders), writes to topics with two partitions, and with replication factor two. The development team notices that the data is lagging behind.
What should be done to reduce the data lag*?
The Connector definition is listed below:
{
"name": "confluent-postgresql-source",
"connector class": "PostgresSource",
"topic.prefix": "postgresql_",
& nbsp;& nbsp;& nbsp;…
"db.name": "postgres",
"table.whitelist": "users.payment.orders”,
"timestamp.column.name": "created_at",
"output.data format": "JSON",
"db.timezone": "UTC",
"tasks.max": "1"
}
How can authentication for both internal component traffic and external client traffic be accomplished?