You are creating a Kafka Streams application to process retail data.
Match the input data streams with the appropriate Kafka Streams object.
You need to consume messages from Kafka using the command-line interface (CLI).
Which command should you use?
This schema excerpt is an example of which schema format?
package com.mycorp.mynamespace;
message SampleRecord {
int32 Stock = 1;
double Price = 2;
string Product_Name = 3;
}
Your application is consuming from a topic with one consumer group.
The number of running consumers is equal to the number of partitions.
Application logs show that some consumers are leaving the consumer group during peak time, triggering a rebalance. You also notice that your application is processing many duplicates.
You need to stop consumers from leaving the consumer group.
What should you do?
You need to collect logs from a host and write them to a Kafka topic named 'logs-topic'. You decide to use Kafka Connect File Source connector for this task.
What is the preferred deployment mode for this connector?
You create a topic named loT-Data with 10 partitions and replication factor of three.
A producer sends 1 MB messages compressed with Gzip.
Which two statements are true in this scenario?
(Select two.)
A stream processing application is consuming from a topic with five partitions. You run three instances of the application. Each instance has num.stream.threads=5.
You need to identify the number of stream tasks that will be created and how many will actively consume messages from the input topic.
What is a consequence of increasing the number of partitions in an existing Kafka topic?
Which partition assignment minimizes partition movements between two assignments?
Your Kafka cluster has five brokers. The topic t1 on the cluster has:
Two partitions
Replication factor = 4
min.insync.replicas = 3You need strong durability guarantees for messages written to topic t1.You configure a producer acks=all and all the replicas for t1 are in-sync.How many brokers need to acknowledge a message before it is considered committed?
You have a consumer group with default configuration settings reading messages from your Kafka cluster.
You need to optimize throughput so the consumer group processes more messages in the same amount of time.
Which change should you make?
Match the topic configuration setting with the reason the setting affects topic durability.
(You are given settings like unclean.leader.election.enable=false, replication.factor, min.insync.replicas=2)
What is the default maximum size of a message the Apache Kafka broker can accept?
Which two statements about Kafka Connect Single Message Transforms (SMTs) are correct?
(Select two.)