/* * librdkafka - Apache Kafka C/C++ library * * Copyright (c) 2015 Magnus Edenhill * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * 1. Redistributions of source code must retain the above copyright notice, * this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright notice, * this list of conditions and the following disclaimer in the documentation * and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE * POSSIBILITY OF SUCH DAMAGE. */ #include #include #include "rdkafkacpp_int.h" RdKafka::KafkaConsumer::~KafkaConsumer() { } RdKafka::KafkaConsumer *RdKafka::KafkaConsumer::create( const RdKafka::Conf *conf, std::string &errstr) { char errbuf[512]; const RdKafka::ConfImpl *confimpl = dynamic_cast(conf); RdKafka::KafkaConsumerImpl *rkc = new RdKafka::KafkaConsumerImpl(); rd_kafka_conf_t *rk_conf = NULL; size_t grlen; if (!confimpl || !confimpl->rk_conf_) { errstr = "Requires RdKafka::Conf::CONF_GLOBAL object"; delete rkc; return NULL; } if (rd_kafka_conf_get(confimpl->rk_conf_, "group.id", NULL, &grlen) != RD_KAFKA_CONF_OK || grlen <= 1 /* terminating null only */) { errstr = "\"group.id\" must be configured"; delete rkc; return NULL; } rkc->set_common_config(confimpl); rk_conf = rd_kafka_conf_dup(confimpl->rk_conf_); rd_kafka_t *rk; if (!(rk = rd_kafka_new(RD_KAFKA_CONSUMER, rk_conf, errbuf, sizeof(errbuf)))) { errstr = errbuf; // rd_kafka_new() takes ownership only if succeeds rd_kafka_conf_destroy(rk_conf); delete rkc; return NULL; } rkc->rk_ = rk; /* Redirect handle queue to cgrp's queue to provide a single queue point */ rd_kafka_poll_set_consumer(rk); return rkc; } RdKafka::ErrorCode RdKafka::KafkaConsumerImpl::subscribe( const std::vector &topics) { rd_kafka_topic_partition_list_t *c_topics; rd_kafka_resp_err_t err; c_topics = rd_kafka_topic_partition_list_new((int)topics.size()); for (unsigned int i = 0; i < topics.size(); i++) rd_kafka_topic_partition_list_add(c_topics, topics[i].c_str(), RD_KAFKA_PARTITION_UA); err = rd_kafka_subscribe(rk_, c_topics); rd_kafka_topic_partition_list_destroy(c_topics); return static_cast(err); } RdKafka::ErrorCode RdKafka::KafkaConsumerImpl::unsubscribe() { return static_cast(rd_kafka_unsubscribe(this->rk_)); } RdKafka::Message *RdKafka::KafkaConsumerImpl::consume(int timeout_ms) { rd_kafka_message_t *rkmessage; rkmessage = rd_kafka_consumer_poll(this->rk_, timeout_ms); if (!rkmessage) return new RdKafka::MessageImpl(RD_KAFKA_CONSUMER, NULL, RdKafka::ERR__TIMED_OUT); return new RdKafka::MessageImpl(RD_KAFKA_CONSUMER, rkmessage); } RdKafka::ErrorCode RdKafka::KafkaConsumerImpl::assignment( std::vector &partitions) { rd_kafka_topic_partition_list_t *c_parts; rd_kafka_resp_err_t err; if ((err = rd_kafka_assignment(rk_, &c_parts))) return static_cast(err); partitions.resize(c_parts->cnt); for (int i = 0; i < c_parts->cnt; i++) partitions[i] = new RdKafka::TopicPartitionImpl(&c_parts->elems[i]); rd_kafka_topic_partition_list_destroy(c_parts); return RdKafka::ERR_NO_ERROR; } bool RdKafka::KafkaConsumerImpl::assignment_lost() { return rd_kafka_assignment_lost(rk_) ? true : false; } RdKafka::ErrorCode RdKafka::KafkaConsumerImpl::subscription( std::vector &topics) { rd_kafka_topic_partition_list_t *c_topics; rd_kafka_resp_err_t err; if ((err = rd_kafka_subscription(rk_, &c_topics))) return static_cast(err); topics.resize(c_topics->cnt); for (int i = 0; i < c_topics->cnt; i++) topics[i] = std::string(c_topics->elems[i].topic); rd_kafka_topic_partition_list_destroy(c_topics); return RdKafka::ERR_NO_ERROR; } RdKafka::ErrorCode RdKafka::KafkaConsumerImpl::assign( const std::vector &partitions) { rd_kafka_topic_partition_list_t *c_parts; rd_kafka_resp_err_t err; c_parts = partitions_to_c_parts(partitions); err = rd_kafka_assign(rk_, c_parts); rd_kafka_topic_partition_list_destroy(c_parts); return static_cast(err); } RdKafka::ErrorCode RdKafka::KafkaConsumerImpl::unassign() { return static_cast(rd_kafka_assign(rk_, NULL)); } RdKafka::Error *RdKafka::KafkaConsumerImpl::incremental_assign( const std::vector &partitions) { rd_kafka_topic_partition_list_t *c_parts; rd_kafka_error_t *c_error; c_parts = partitions_to_c_parts(partitions); c_error = rd_kafka_incremental_assign(rk_, c_parts); rd_kafka_topic_partition_list_destroy(c_parts); if (c_error) return new ErrorImpl(c_error); return NULL; } RdKafka::Error *RdKafka::KafkaConsumerImpl::incremental_unassign( const std::vector &partitions) { rd_kafka_topic_partition_list_t *c_parts; rd_kafka_error_t *c_error; c_parts = partitions_to_c_parts(partitions); c_error = rd_kafka_incremental_unassign(rk_, c_parts); rd_kafka_topic_partition_list_destroy(c_parts); if (c_error) return new ErrorImpl(c_error); return NULL; } RdKafka::ErrorCode RdKafka::KafkaConsumerImpl::committed( std::vector &partitions, int timeout_ms) { rd_kafka_topic_partition_list_t *c_parts; rd_kafka_resp_err_t err; c_parts = partitions_to_c_parts(partitions); err = rd_kafka_committed(rk_, c_parts, timeout_ms); if (!err) { update_partitions_from_c_parts(partitions, c_parts); } rd_kafka_topic_partition_list_destroy(c_parts); return static_cast(err); } RdKafka::ErrorCode RdKafka::KafkaConsumerImpl::position( std::vector &partitions) { rd_kafka_topic_partition_list_t *c_parts; rd_kafka_resp_err_t err; c_parts = partitions_to_c_parts(partitions); err = rd_kafka_position(rk_, c_parts); if (!err) { update_partitions_from_c_parts(partitions, c_parts); } rd_kafka_topic_partition_list_destroy(c_parts); return static_cast(err); } RdKafka::ErrorCode RdKafka::KafkaConsumerImpl::seek( const RdKafka::TopicPartition &partition, int timeout_ms) { const RdKafka::TopicPartitionImpl *p = dynamic_cast(&partition); rd_kafka_topic_t *rkt; if (!(rkt = rd_kafka_topic_new(rk_, p->topic_.c_str(), NULL))) return static_cast(rd_kafka_last_error()); /* FIXME: Use a C API that takes a topic_partition_list_t instead */ RdKafka::ErrorCode err = static_cast( rd_kafka_seek(rkt, p->partition_, p->offset_, timeout_ms)); rd_kafka_topic_destroy(rkt); return err; } RdKafka::ErrorCode RdKafka::KafkaConsumerImpl::close() { return static_cast(rd_kafka_consumer_close(rk_)); } RdKafka::Error *RdKafka::KafkaConsumerImpl::close(Queue *queue) { QueueImpl *queueimpl = dynamic_cast(queue); rd_kafka_error_t *c_error; c_error = rd_kafka_consumer_close_queue(rk_, queueimpl->queue_); if (c_error) return new ErrorImpl(c_error); return NULL; } RdKafka::ConsumerGroupMetadata::~ConsumerGroupMetadata() { }