summaryrefslogtreecommitdiffstats
path: root/src/fluent-bit/lib/librdkafka-2.1.0/tests/0122-buffer_cleaning_after_rebalance.c
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-05-05 11:19:16 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-05-05 12:07:37 +0000
commitb485aab7e71c1625cfc27e0f92c9509f42378458 (patch)
treeae9abe108601079d1679194de237c9a435ae5b55 /src/fluent-bit/lib/librdkafka-2.1.0/tests/0122-buffer_cleaning_after_rebalance.c
parentAdding upstream version 1.44.3. (diff)
downloadnetdata-b485aab7e71c1625cfc27e0f92c9509f42378458.tar.xz
netdata-b485aab7e71c1625cfc27e0f92c9509f42378458.zip
Adding upstream version 1.45.3+dfsg.
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'src/fluent-bit/lib/librdkafka-2.1.0/tests/0122-buffer_cleaning_after_rebalance.c')
-rw-r--r--src/fluent-bit/lib/librdkafka-2.1.0/tests/0122-buffer_cleaning_after_rebalance.c226
1 files changed, 226 insertions, 0 deletions
diff --git a/src/fluent-bit/lib/librdkafka-2.1.0/tests/0122-buffer_cleaning_after_rebalance.c b/src/fluent-bit/lib/librdkafka-2.1.0/tests/0122-buffer_cleaning_after_rebalance.c
new file mode 100644
index 000000000..4f8727017
--- /dev/null
+++ b/src/fluent-bit/lib/librdkafka-2.1.0/tests/0122-buffer_cleaning_after_rebalance.c
@@ -0,0 +1,226 @@
+/*
+ * librdkafka - Apache Kafka C library
+ *
+ * Copyright (c) 2021, Magnus Edenhill
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions are met:
+ *
+ * 1. Redistributions of source code must retain the above copyright notice,
+ * this list of conditions and the following disclaimer.
+ * 2. Redistributions in binary form must reproduce the above copyright notice,
+ * this list of conditions and the following disclaimer in the documentation
+ * and/or other materials provided with the distribution.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
+ * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+ * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
+ * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
+ * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
+ * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
+ * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
+ * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
+ * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
+ * POSSIBILITY OF SUCH DAMAGE.
+ */
+
+#include "test.h"
+/* Typical include path would be <librdkafka/rdkafka.h>, but this program
+ * is built from within the librdkafka source tree and thus differs. */
+#include "rdkafka.h" /* for Kafka driver */
+
+typedef struct consumer_s {
+ const char *what;
+ rd_kafka_queue_t *rkq;
+ int timeout_ms;
+ int consume_msg_cnt;
+ int expected_msg_cnt;
+ rd_kafka_t *rk;
+ uint64_t testid;
+ test_msgver_t *mv;
+ struct test *test;
+} consumer_t;
+
+static int consumer_batch_queue(void *arg) {
+ consumer_t *arguments = arg;
+ int msg_cnt = 0;
+ int i;
+ test_timing_t t_cons;
+
+ rd_kafka_queue_t *rkq = arguments->rkq;
+ int timeout_ms = arguments->timeout_ms;
+ const int consume_msg_cnt = arguments->consume_msg_cnt;
+ rd_kafka_t *rk = arguments->rk;
+ uint64_t testid = arguments->testid;
+ rd_kafka_message_t **rkmessage =
+ malloc(consume_msg_cnt * sizeof(*rkmessage));
+
+ if (arguments->test)
+ test_curr = arguments->test;
+
+ TEST_SAY(
+ "%s calling consume_batch_queue(timeout=%d, msgs=%d) "
+ "and expecting %d messages back\n",
+ rd_kafka_name(rk), timeout_ms, consume_msg_cnt,
+ arguments->expected_msg_cnt);
+
+ TIMING_START(&t_cons, "CONSUME");
+ msg_cnt = (int)rd_kafka_consume_batch_queue(rkq, timeout_ms, rkmessage,
+ consume_msg_cnt);
+ TIMING_STOP(&t_cons);
+
+ TEST_SAY("%s consumed %d/%d/%d message(s)\n", rd_kafka_name(rk),
+ msg_cnt, arguments->consume_msg_cnt,
+ arguments->expected_msg_cnt);
+ TEST_ASSERT(msg_cnt == arguments->expected_msg_cnt,
+ "consumed %d messages, expected %d", msg_cnt,
+ arguments->expected_msg_cnt);
+
+ for (i = 0; i < msg_cnt; i++) {
+ if (test_msgver_add_msg(rk, arguments->mv, rkmessage[i]) == 0)
+ TEST_FAIL(
+ "The message is not from testid "
+ "%" PRId64 " \n",
+ testid);
+ rd_kafka_message_destroy(rkmessage[i]);
+ }
+
+ free(rkmessage);
+
+ return 0;
+}
+
+
+/**
+ * @brief Produce 400 messages and consume 500 messages totally by 2 consumers
+ * using batch queue method, verify if there isn't any missed or
+ * duplicate messages received by the two consumers.
+ * The reasons for setting the consume messages number is higher than
+ * or equal to the produce messages number are:
+ * 1) Make sure each consumer can at most receive half of the produced
+ * messages even though the consumers expect more.
+ * 2) If the consume messages number is smaller than the produce
+ * messages number, it's hard to verify that the messages returned
+ * are added to the batch queue before or after the rebalancing.
+ * But if the consume messages number is larger than the produce
+ * messages number, and we still received half of the produced
+ * messages by each consumer, we can make sure that the buffer
+ * cleaning is happened during the batch queue process to guarantee
+ * only received messages added to the batch queue after the
+ * rebalance.
+ *
+ * 1. Produce 100 messages to each of the 4 partitions
+ * 2. First consumer subscribes to the topic, wait for it's assignment
+ * 3. The first consumer consumes 500 messages using the batch queue
+ * method
+ * 4. Second consumer subscribes to the topic, wait for it's assignment
+ * 5. Rebalance happenes
+ * 6. The second consumer consumes 500 messages using the batch queue
+ * method
+ * 7. Each consumer receives 200 messages finally
+ * 8. Combine all the messages received by the 2 consumers and
+ * verify if there isn't any missed or duplicate messages
+ *
+ */
+static void do_test_consume_batch(const char *strategy) {
+ const int partition_cnt = 4;
+ rd_kafka_queue_t *rkq1, *rkq2;
+ const char *topic;
+ rd_kafka_t *c1;
+ rd_kafka_t *c2;
+ int p;
+ const int timeout_ms = 12000; /* Must be > rebalance time */
+ uint64_t testid;
+ const int consume_msg_cnt = 500;
+ const int produce_msg_cnt = 400;
+ rd_kafka_conf_t *conf;
+ consumer_t c1_args = RD_ZERO_INIT;
+ consumer_t c2_args = RD_ZERO_INIT;
+ test_msgver_t mv;
+ thrd_t thread_id;
+
+ SUB_TEST("partition.assignment.strategy = %s", strategy);
+
+ test_conf_init(&conf, NULL, 60);
+ test_conf_set(conf, "enable.auto.commit", "false");
+ test_conf_set(conf, "auto.offset.reset", "earliest");
+ test_conf_set(conf, "partition.assignment.strategy", strategy);
+
+ testid = test_id_generate();
+ test_msgver_init(&mv, testid);
+
+ /* Produce messages */
+ topic = test_mk_topic_name("0122-buffer_cleaning", 1);
+
+ for (p = 0; p < partition_cnt; p++)
+ test_produce_msgs_easy(topic, testid, p,
+ produce_msg_cnt / partition_cnt);
+
+ /* Create consumers */
+ c1 = test_create_consumer(topic, NULL, rd_kafka_conf_dup(conf), NULL);
+ c2 = test_create_consumer(topic, NULL, conf, NULL);
+
+ test_consumer_subscribe(c1, topic);
+ test_consumer_wait_assignment(c1, rd_false);
+
+ /* Create generic consume queue */
+ rkq1 = rd_kafka_queue_get_consumer(c1);
+
+ c1_args.what = "C1.PRE";
+ c1_args.rkq = rkq1;
+ c1_args.timeout_ms = timeout_ms;
+ c1_args.consume_msg_cnt = consume_msg_cnt;
+ c1_args.expected_msg_cnt = produce_msg_cnt / 2;
+ c1_args.rk = c1;
+ c1_args.testid = testid;
+ c1_args.mv = &mv;
+ c1_args.test = test_curr;
+ if (thrd_create(&thread_id, consumer_batch_queue, &c1_args) !=
+ thrd_success)
+ TEST_FAIL("Failed to create thread for %s", "C1.PRE");
+
+ test_consumer_subscribe(c2, topic);
+ test_consumer_wait_assignment(c2, rd_false);
+
+ thrd_join(thread_id, NULL);
+
+ /* Create generic consume queue */
+ rkq2 = rd_kafka_queue_get_consumer(c2);
+
+ c2_args.what = "C2.PRE";
+ c2_args.rkq = rkq2;
+ /* Second consumer should be able to consume all messages right away */
+ c2_args.timeout_ms = 5000;
+ c2_args.consume_msg_cnt = consume_msg_cnt;
+ c2_args.expected_msg_cnt = produce_msg_cnt / 2;
+ c2_args.rk = c2;
+ c2_args.testid = testid;
+ c2_args.mv = &mv;
+
+ consumer_batch_queue(&c2_args);
+
+ test_msgver_verify("C1.PRE + C2.PRE", &mv,
+ TEST_MSGVER_ORDER | TEST_MSGVER_DUP, 0,
+ produce_msg_cnt);
+ test_msgver_clear(&mv);
+
+ rd_kafka_queue_destroy(rkq1);
+ rd_kafka_queue_destroy(rkq2);
+
+ test_consumer_close(c1);
+ test_consumer_close(c2);
+
+ rd_kafka_destroy(c1);
+ rd_kafka_destroy(c2);
+
+ SUB_TEST_PASS();
+}
+
+
+int main_0122_buffer_cleaning_after_rebalance(int argc, char **argv) {
+ do_test_consume_batch("range");
+ do_test_consume_batch("cooperative-sticky");
+ return 0;
+}