aiotestking uk

CCDAK Exam Questions - Online Test


CCDAK Premium VCE File

Learn More 100% Pass Guarantee - Dumps Verified - Instant Download
150 Lectures, 20 Hours

Your success in Confluent CCDAK is our sole target and we develop all our CCDAK braindumps in a way that facilitates the attainment of this target. Not only is our CCDAK study material the best you can find, it is also the most detailed and the most updated. CCDAK Practice Exams for Confluent CCDAK are written to the highest standards of technical accuracy.

Confluent CCDAK Free Dumps Questions Online, Read and Test Now.

NEW QUESTION 1
Which KSQL queries write to Kafka?

  • A. COUNT and JOIN
  • B. SHOW STREAMS and EXPLAIN <query> statements
  • C. CREATE STREAM WITH <topic> and CREATE TABLE WITH <topic>
  • D. CREATE STREAM AS SELECT and CREATE TABLE AS SELECT

Answer: CD

Explanation:
SHOW STREAMS and EXPLAIN <query> statements run against the KSQL server that the KSQL client is connected to. They don't communicate directly with Kafka. CREATE STREAM WITH <topic> and CREATE TABLE WITH <topic> write metadata to the KSQL command topic. Persistent queries based on CREATE STREAM AS SELECT and CREATE TABLE AS SELECT read and write to Kafka topics. Non-persistent queries based on SELECT that are stateless only read from Kafka topics, for example SELECT … FROM foo WHERE …. Non-persistent queries that are stateful read and write to Kafka, for example, COUNT and JOIN. The data in Kafka is deleted automatically when you terminate the query with CTRL-C.

NEW QUESTION 2
You are running a Kafka Streams application in a Docker container managed by Kubernetes, and upon application restart, it takes a long time for the docker container to replicate the state and get back to processing the data. How can you improve dramatically the application restart?

  • A. Mount a persistent volume for your RocksDB
  • B. Increase the number of partitions in your inputs topic
  • C. Reduce the Streams caching property
  • D. Increase the number of Streams threads

Answer: A

Explanation:
Although any Kafka Streams application is stateless as the state is stored in Kafka, it can take a while and lots of resources to recover the state from Kafka. In order to speed up recovery, it is advised to store the Kafka Streams state on a persistent volume, so that only the missing part of the state needs to be recovered.

NEW QUESTION 3
What exceptions may be caught by the following producer? (select two) ProducerRecord<String, String> record =
new ProducerRecord<>("topic1", "key1", "value1"); try {
producer.send(record);
} catch (Exception e) { e.printStackTrace();
}

  • A. BrokerNotAvailableException
  • B. SerializationException
  • C. InvalidPartitionsException
  • D. BufferExhaustedException

Answer: BD

Explanation:
These are the client side exceptions that may be encountered before message is sent to the broker, and before a future is returned by the .send() method.

NEW QUESTION 4
A producer is sending messages with null key to a topic with 6 partitions using the DefaultPartitioner. Where will the messages be stored?

  • A. Partition 5
  • B. Any of the topic partitions
  • C. The partition for the null key
  • D. Partition 0

Answer: A

Explanation:
Message with no keys will be stored with round-robin strategy among partitions.

NEW QUESTION 5
To produce data to a topic, a producer must provide the Kafka client with...

  • A. the list of brokers that have the data, the topic name and the partitions list
  • B. any broker from the cluster and the topic name and the partitions list
  • C. all the brokers from the cluster and the topic name
  • D. any broker from the cluster and the topic name

Answer: D

Explanation:
All brokers can respond to a Metadata request, so a client can connect to any broker in the cluster and then figure out on its own which brokers to send data to.

NEW QUESTION 6
You want to send a message of size 3 MB to a topic with default message size configuration. How does KafkaProducer handle large messages?

  • A. KafkaProducer divides messages into sizes of max.request.size and sends them in order
  • B. KafkaProducer divides messages into sizes of message.max.bytes and sends them in order
  • C. MessageSizeTooLarge exception will be thrown, KafkaProducer will not retry and return exception immediately
  • D. MessageSizeTooLarge exception will be thrown, KafkaProducer retries until the number of retries are exhausted

Answer: C

Explanation:
MessageSizeTooLarge is not a retryable exception.

NEW QUESTION 7
To import data from external databases, I should use

  • A. Confluent REST Proxy
  • B. Kafka Connect Sink
  • C. Kafka Streams
  • D. Kafka Connect Source

Answer: D

Explanation:
Kafka Connect Sink is used to export data from Kafka to external databases and Kafka
Connect Source is used to import from external databases into Kafka.

NEW QUESTION 8
A consumer is configured with enable.auto.commit=false. What happens when close() is called on the consumer object?

  • A. The uncommitted offsets are committed
  • B. A rebalance in the consumer group will happen immediately
  • C. The group coordinator will discover that the consumer stopped sending heartbeat
  • D. It will cause rebalance after session.timeout.ms

Answer: B

Explanation:
Calling close() on consumer immediately triggers a partition rebalance as the consumer will not be available anymore.

NEW QUESTION 9
How will you set the retention for the topic named ‚Ä??my-topic‚Ä?? to 1 hour?

  • A. Set the broker config log.retention.ms to 3600000
  • B. Set the consumer config retention.ms to 3600000
  • C. Set the topic config retention.ms to 3600000
  • D. Set the producer config retention.ms to 3600000

Answer: C

Explanation:
retention.ms can be configured at topic level while creating topic or by altering topic. It shouldn't be set at the broker level (log.retention.ms) as this would impact all the topics in the cluster, not just the one we are interested in

NEW QUESTION 10
The exactly once guarantee in the Kafka Streams is for which flow of data?

  • A. Kafka => Kafka
  • B. Kafka => External
  • C. External => Kafka

Answer: A

Explanation:
Kafka Streams can only guarantee exactly once processing if you have a Kafka to Kafka topology.

NEW QUESTION 11
What is a generic unique id that I can use for messages I receive from a consumer?

  • A. topic + partition + timestamp
  • B. topic + partition + offset
  • C. topic + timestamp

Answer: B

Explanation:
(Topic,Partition,Offset) uniquely identifies a message in Kafka

NEW QUESTION 12
The Controller is a broker that is... (select two)

  • A. elected by Zookeeper ensemble
  • B. is responsible for partition leader election
  • C. elected by broker majority
  • D. is responsible for consumer group rebalances

Answer: AB

Explanation:
Controller is a broker that in addition to usual broker functions is responsible for partition leader election. The election of that broker happens thanks to Zookeeper and at any time only one broker can be a controller

NEW QUESTION 13
By default, which replica will be elected as a partition leader? (select two)

  • A. Preferred leader broker if it is in-sync and auto.leader.rebalance.enable=true
  • B. Any of the replicas
  • C. Preferred leader broker if it is in-sync and auto.leader.rebalance.enable=false
  • D. An in-sync replica

Answer: BD

Explanation:
Preferred leader is a broker that was leader when topic was created. It is preferred because when partitions are first created, the leaders are balanced between brokers. Otherwise, any of the in-sync replicas (ISR) will be elected leader, as long as unclean.leader.election=false (by default)

NEW QUESTION 14
StreamsBuilder builder = new StreamsBuilder();
KStream<String, String> textLines = builder.stream("word-count-input"); KTable<String, Long> wordCounts = textLines
.mapValues(textLine -> textLine.toLowerCase())
.flatMapValues(textLine -> Arrays.asList(textLine.split("\W+")))
.selectKey((key, word) -> word)
.groupByKey()
.count(Materialized.as("Counts"));
wordCounts.toStream().to("word-count-output", Produced.with(Serdes.String(), Serdes.Long()));
builder.build();
What is an adequate topic configuration for the topic word-count-output?

  • A. max.message.bytes=10000000
  • B. cleanup.policy=delete
  • C. compression.type=lz4
  • D. cleanup.policy=compact

Answer: D

Explanation:
Result is aggregated into a table with key as the unique word and value its frequency. We have to enable log compaction for this topic to align the topic's cleanup policy with KTable semantics.

NEW QUESTION 15
What Java library is KSQL based on?

  • A. Kafka Streams
  • B. REST Proxy
  • C. Schema Registry
  • D. Kafka Connect

Answer: A

Explanation:
KSQL is based on Kafka Streams and allows you to express transformations in the SQL language that get automatically converted to a Kafka Streams program in the backend

NEW QUESTION 16
How much should be the heap size of a broker in a production setup on a machine with 256 GB of RAM, in PLAINTEXT mode?

  • A. 4 GB
  • B. 128 GB
  • C. 16 GB
  • D. 512 MB

Answer: A

Explanation:
In Kafka, a small heap size is needed, while the rest of the RAM goes automatically to the page cache (managed by the OS). The heap size goes slightly up if you need to enable SSL

NEW QUESTION 17
To prevent network-induced duplicates when producing to Kafka, I should use

  • A. max.in.flight.requests.per.connection=1
  • B. enable.idempotence=true
  • C. retries=200000
  • D. batch.size=1

Answer: B

Explanation:
Producer idempotence helps prevent the network introduced duplicates. More details herehttps://cwiki.apache.org/confluence/display/KAFKA/Idempotent+Producer

NEW QUESTION 18
......

P.S. Dumps-hub.com now are offering 100% pass ensure CCDAK dumps! All CCDAK exam questions have been updated with correct answers: https://www.dumps-hub.com/CCDAK-dumps.html (150 New Questions)