diff options
Diffstat (limited to 'fluent-bit/lib/librdkafka-2.1.0/tests/0054-offset_time.cpp')
-rw-r--r-- | fluent-bit/lib/librdkafka-2.1.0/tests/0054-offset_time.cpp | 236 |
1 files changed, 236 insertions, 0 deletions
diff --git a/fluent-bit/lib/librdkafka-2.1.0/tests/0054-offset_time.cpp b/fluent-bit/lib/librdkafka-2.1.0/tests/0054-offset_time.cpp new file mode 100644 index 00000000..58c88b4a --- /dev/null +++ b/fluent-bit/lib/librdkafka-2.1.0/tests/0054-offset_time.cpp @@ -0,0 +1,236 @@ +/* + * librdkafka - Apache Kafka C library + * + * Copyright (c) 2012-2015, Magnus Edenhill + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, + * this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright notice, + * this list of conditions and the following disclaimer in the documentation + * and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE + * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + * POSSIBILITY OF SUCH DAMAGE. + */ + +#include <iostream> +#include "testcpp.h" + +/** + * Test offset_for_times (KIP-79): time-based offset lookups. + */ + + +static int verify_offset(const RdKafka::TopicPartition *tp, + int64_t timestamp, + int64_t exp_offset, + RdKafka::ErrorCode exp_err) { + int fails = 0; + if (tp->err() != exp_err) { + Test::FailLater(tostr() + << " " << tp->topic() << " [" << tp->partition() << "] " + << "expected error " << RdKafka::err2str(exp_err) + << ", got " << RdKafka::err2str(tp->err()) << "\n"); + fails++; + } + + if (!exp_err && tp->offset() != exp_offset) { + Test::FailLater(tostr() + << " " << tp->topic() << " [" << tp->partition() << "] " + << "expected offset " << exp_offset << " for timestamp " + << timestamp << ", got " << tp->offset() << "\n"); + fails++; + } + + return fails; +} + + +static void test_offset_time(void) { + std::vector<RdKafka::TopicPartition *> query_parts; + std::string topic = Test::mk_topic_name("0054-offset_time", 1); + RdKafka::Conf *conf, *tconf; + int64_t timestamps[] = { + /* timestamp, expected offset */ + 1234, + 0, + 999999999999, + 1, + }; + const int timestamp_cnt = 2; + int fails = 0; + std::string errstr; + + Test::conf_init(&conf, &tconf, 0); + + /* Need acks=all to make sure OffsetRequest correctly reads fully + * written Produce record. */ + Test::conf_set(tconf, "acks", "all"); + Test::conf_set(conf, "api.version.request", "true"); + conf->set("dr_cb", &Test::DrCb, errstr); + conf->set("default_topic_conf", tconf, errstr); + + RdKafka::Producer *p = RdKafka::Producer::create(conf, errstr); + if (!p) + Test::Fail("Failed to create Producer: " + errstr); + + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 97, timestamps[0])); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 98, timestamps[0])); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 99, timestamps[0])); + + /* First query timestamps before topic exists, should fail. */ + Test::Say("Attempting first offsetsForTimes() query (should fail)\n"); + RdKafka::ErrorCode err = p->offsetsForTimes(query_parts, tmout_multip(10000)); + Test::Say("offsetsForTimes #1 with non-existing partitions returned " + + RdKafka::err2str(err) + "\n"); + Test::print_TopicPartitions("offsetsForTimes #1", query_parts); + + if (err != RdKafka::ERR__UNKNOWN_PARTITION) + Test::Fail( + "offsetsForTimes #1 should have failed with UNKNOWN_PARTITION, " + "not " + + RdKafka::err2str(err)); + + Test::Say("Producing to " + topic + "\n"); + for (int partition = 0; partition < 2; partition++) { + for (int ti = 0; ti < timestamp_cnt * 2; ti += 2) { + err = p->produce(topic, partition, RdKafka::Producer::RK_MSG_COPY, + (void *)topic.c_str(), topic.size(), NULL, 0, + timestamps[ti], NULL); + if (err != RdKafka::ERR_NO_ERROR) + Test::Fail("Produce failed: " + RdKafka::err2str(err)); + } + } + + if (p->flush(tmout_multip(5000)) != 0) + Test::Fail("Not all messages flushed"); + + + for (int ti = 0; ti < timestamp_cnt * 2; ti += 2) { + RdKafka::TopicPartition::destroy(query_parts); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 0, timestamps[ti])); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 1, timestamps[ti])); + + Test::Say(tostr() << "Attempting offsetsForTimes() for timestamp " + << timestamps[ti] << "\n"); + err = p->offsetsForTimes(query_parts, tmout_multip(5000)); + Test::print_TopicPartitions("offsetsForTimes", query_parts); + if (err != RdKafka::ERR_NO_ERROR) + Test::Fail("offsetsForTimes failed: " + RdKafka::err2str(err)); + + fails += verify_offset(query_parts[0], timestamps[ti], timestamps[ti + 1], + RdKafka::ERR_NO_ERROR); + fails += verify_offset(query_parts[1], timestamps[ti], timestamps[ti + 1], + RdKafka::ERR_NO_ERROR); + } + + /* repeat test with -1 timeout */ + for (int ti = 0; ti < timestamp_cnt * 2; ti += 2) { + RdKafka::TopicPartition::destroy(query_parts); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 0, timestamps[ti])); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 1, timestamps[ti])); + + Test::Say(tostr() << "Attempting offsetsForTimes() for timestamp " + << timestamps[ti] << " with a timeout of -1\n"); + err = p->offsetsForTimes(query_parts, -1); + Test::print_TopicPartitions("offsetsForTimes", query_parts); + if (err != RdKafka::ERR_NO_ERROR) + Test::Fail("offsetsForTimes failed: " + RdKafka::err2str(err)); + + fails += verify_offset(query_parts[0], timestamps[ti], timestamps[ti + 1], + RdKafka::ERR_NO_ERROR); + fails += verify_offset(query_parts[1], timestamps[ti], timestamps[ti + 1], + RdKafka::ERR_NO_ERROR); + } + + /* And a negative test with a request that should timeout instantly. */ + for (int ti = 0; ti < timestamp_cnt * 2; ti += 2) { + RdKafka::TopicPartition::destroy(query_parts); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 0, timestamps[ti])); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 1, timestamps[ti])); + + Test::Say(tostr() << "Attempting offsetsForTimes() for timestamp " + << timestamps[ti] + << " with minimal timeout (should fail)\n"); + err = p->offsetsForTimes(query_parts, 0); + Test::print_TopicPartitions("offsetsForTimes", query_parts); + if (err != RdKafka::ERR__TIMED_OUT) + Test::Fail( + "expected offsetsForTimes(timeout=0) to fail with TIMED_OUT, not " + + RdKafka::err2str(err)); + } + + /* Include non-existent partitions */ + for (int ti = 0; ti < timestamp_cnt * 2; ti += 2) { + RdKafka::TopicPartition::destroy(query_parts); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 0, timestamps[ti])); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 1, timestamps[ti])); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 2, timestamps[ti])); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 20, timestamps[ti])); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 3, timestamps[ti])); + query_parts.push_back( + RdKafka::TopicPartition::create(topic, 21, timestamps[ti])); + Test::Say("Attempting offsetsForTimes() with non-existent partitions\n"); + err = p->offsetsForTimes(query_parts, -1); + Test::print_TopicPartitions("offsetsForTimes", query_parts); + if (err != RdKafka::ERR_NO_ERROR) + Test::Fail("expected offsetsForTimes(timeout=0) to succeed, not " + + RdKafka::err2str(err)); + fails += verify_offset(query_parts[0], timestamps[ti], timestamps[ti + 1], + RdKafka::ERR_NO_ERROR); + fails += verify_offset(query_parts[1], timestamps[ti], timestamps[ti + 1], + RdKafka::ERR_NO_ERROR); + fails += verify_offset(query_parts[2], timestamps[ti], -1, + RdKafka::ERR_NO_ERROR); + fails += verify_offset(query_parts[3], timestamps[ti], -1, + RdKafka::ERR__UNKNOWN_PARTITION); + fails += verify_offset(query_parts[4], timestamps[ti], -1, + RdKafka::ERR_NO_ERROR); + fails += verify_offset(query_parts[5], timestamps[ti], -1, + RdKafka::ERR__UNKNOWN_PARTITION); + } + + + if (fails > 0) + Test::Fail(tostr() << "See " << fails << " previous error(s)"); + + RdKafka::TopicPartition::destroy(query_parts); + + delete p; + delete conf; + delete tconf; +} + +extern "C" { +int main_0054_offset_time(int argc, char **argv) { + test_offset_time(); + return 0; +} +} |