/* * Copyright (c) 2017, Matias Fontanini * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are * met: * * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following disclaimer * in the documentation and/or other materials provided with the * distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * */ #ifndef CPPKAFKA_KAFKA_HANDLE_BASE_H #define CPPKAFKA_KAFKA_HANDLE_BASE_H #include #include #include #include #include #include #include #include #include #include "group_information.h" #include "topic_partition.h" #include "topic_partition_list.h" #include "topic_configuration.h" #include "configuration.h" #include "macros.h" #include "logging.h" #include "queue.h" namespace cppkafka { class Topic; class Metadata; class TopicMetadata; class GroupInformation; /** * Base class for kafka consumer/producer */ class CPPKAFKA_API KafkaHandleBase { public: using OffsetTuple = std::tuple; using TopicPartitionsTimestampsMap = std::map; virtual ~KafkaHandleBase() = default; KafkaHandleBase(const KafkaHandleBase&) = delete; KafkaHandleBase(KafkaHandleBase&&) = delete; KafkaHandleBase& operator=(const KafkaHandleBase&) = delete; KafkaHandleBase& operator=(KafkaHandleBase&&) = delete; /** * \brief Pauses consumption/production from the given topic/partition list * * This translates into a call to rd_kafka_pause_partitions * * \param topic_partitions The topic/partition list to pause consuming/producing from/to */ void pause_partitions(const TopicPartitionList& topic_partitions); /** * \brief Pauses consumption/production for this topic * * \param topic The topic name */ void pause(const std::string& topic); /** * \brief Resumes consumption/production from the given topic/partition list * * This translates into a call to rd_kafka_resume_partitions * * \param topic_partitions The topic/partition list to resume consuming/producing from/to */ void resume_partitions(const TopicPartitionList& topic_partitions); /** * \brief Resumes consumption/production for this topic * * \param topic The topic name */ void resume(const std::string& topic); /** * \brief Sets the timeout for operations that require a timeout * * This timeout is applied to operations like polling, querying for offsets, etc * * \param timeout The timeout to be set */ void set_timeout(std::chrono::milliseconds timeout); /** * \brief Sets the log level */ void set_log_level(LogLevel level); /** * \brief Adds one or more brokers to this handle's broker list * * This calls rd_kafka_brokers_add using the provided broker list. * * \param brokers The broker list endpoint string */ void add_brokers(const std::string& brokers); /** * \brief Queries the offset for the given topic/partition * * This translates into a call to rd_kafka_query_watermark_offsets * * \param topic_partition The topic/partition to be queried * * \return A pair of watermark offsets {low, high} */ OffsetTuple query_offsets(const TopicPartition& topic_partition) const; /** * \brief Queries the offset for the given topic/partition with a given timeout * * This translates into a call to rd_kafka_query_watermark_offsets * * \param topic_partition The topic/partition to be queried * * \timeout The timeout for this operation. This supersedes the default handle timeout. * * \return A pair of watermark offsets {low, high} */ OffsetTuple query_offsets(const TopicPartition& topic_partition, std::chrono::milliseconds timeout) const; /** * \brief Gets the rdkafka handle * * \return The rdkafka handle */ rd_kafka_t* get_handle() const; /** * \brief Creates a topic handle * * This translates into a call to rd_kafka_topic_new. This will use the default topic * configuration provided in the Configuration object for this consumer/producer handle, * if any. * * \param name The name of the topic to be created * * \return A topic */ Topic get_topic(const std::string& name); /** * \brief Creates a topic handle * * This translates into a call to rd_kafka_topic_new. * * \param name The name of the topic to be created * \param config The configuration to be used for the new topic * * \return A topic */ Topic get_topic(const std::string& name, TopicConfiguration config); /** * \brief Gets metadata for brokers, topics, partitions, etc * * This translates into a call to rd_kafka_metadata * * \param all_topics Whether to fetch metadata about all topics or only locally known ones * * \return The metadata */ Metadata get_metadata(bool all_topics = true) const; /** * \brief Gets metadata for brokers, topics, partitions, etc with a timeout * * This translates into a call to rd_kafka_metadata * * \param all_topics Whether to fetch metadata about all topics or only locally known ones * * \param timeout The timeout for this operation. Supersedes the default handle timeout. * * \return The metadata */ Metadata get_metadata(bool all_topics, std::chrono::milliseconds timeout) const; /** * \brief Gets general metadata but only fetches metadata for the given topic rather than * all of them * * This translates into a call to rd_kafka_metadata * * \param topic The topic to fetch information for * * \return The topic metadata */ TopicMetadata get_metadata(const Topic& topic) const; /** * \brief Gets general metadata but only fetches metadata for the given topic rather than * all of them. Uses a timeout to limit the operation execution time. * * This translates into a call to rd_kafka_metadata * * \param topic The topic to fetch information for * * \param timeout The timeout for this operation. Supersedes the default handle timeout. * * \return The topic metadata */ TopicMetadata get_metadata(const Topic& topic, std::chrono::milliseconds timeout) const; /** * \brief Gets the consumer group information * * \param name The name of the consumer group to look up * * \return The group information */ GroupInformation get_consumer_group(const std::string& name); /** * \brief Gets the consumer group information with a timeout * * \param name The name of the consumer group to look up * * \param timeout The timeout for this operation. Supersedes the default handle timeout. * * \return The group information */ GroupInformation get_consumer_group(const std::string& name, std::chrono::milliseconds timeout); /** * \brief Gets all consumer groups * * \return A list of consumer groups */ GroupInformationList get_consumer_groups(); /** * \brief Gets all consumer groups with a timeout * * \param timeout The timeout for this operation. Supersedes the default handle timeout. * * \return A list of consumer groups */ GroupInformationList get_consumer_groups(std::chrono::milliseconds timeout); /** * \brief Gets topic/partition offsets based on timestamps * * This translates into a call to rd_kafka_offsets_for_times * * \param queries A map from topic/partition to the timestamp to be used * * \return A topic partition list */ TopicPartitionList get_offsets_for_times(const TopicPartitionsTimestampsMap& queries) const; /** * \brief Gets topic/partition offsets based on timestamps with a timeout * * This translates into a call to rd_kafka_offsets_for_times * * \param queries A map from topic/partition to the timestamp to be used * * \param timeout The timeout for this operation. This supersedes the default handle timeout. * * \return A topic partition list */ TopicPartitionList get_offsets_for_times(const TopicPartitionsTimestampsMap& queries, std::chrono::milliseconds timeout) const; /** * \brief Get the kafka handle name * * \return The handle name */ std::string get_name() const; /** * \brief Gets the configured timeout. * * \return The configured timeout * * \sa KafkaHandleBase::set_timeout */ std::chrono::milliseconds get_timeout() const; /** * \brief Gets the handle's configuration * * \return A reference to the configuration object */ const Configuration& get_configuration() const; #if RD_KAFKA_VERSION >= RD_KAFKA_ADMIN_API_SUPPORT_VERSION /** * \brief Gets the background queue * * This translates into a call to rd_kafka_queue_get_background * * \return The background queue */ Queue get_background_queue() const { return Queue::make_queue(rd_kafka_queue_get_background(handle_.get())); } #endif /** * \brief Gets the length of the out queue * * This calls rd_kafka_outq_len * * \return The length of the queue */ int get_out_queue_length() const; #if RD_KAFKA_VERSION >= RD_KAFKA_DESTROY_FLAGS_SUPPORT_VERSION /** * \brief Sets flags for rd_kafka_destroy_flags() */ void set_destroy_flags(int destroy_flags); /** * \brief Returns flags for rd_kafka_destroy_flags() */ int get_destroy_flags() const; #endif /** * \brief Cancels the current callback dispatcher * * This calls rd_kafka_yield */ void yield() const; protected: KafkaHandleBase(Configuration config); void set_handle(rd_kafka_t* handle); void check_error(rd_kafka_resp_err_t error) const; void check_error(rd_kafka_resp_err_t error, const rd_kafka_topic_partition_list_t* list_ptr) const; rd_kafka_conf_t* get_configuration_handle(); private: static const std::chrono::milliseconds DEFAULT_TIMEOUT; struct HandleDeleter { explicit HandleDeleter(const KafkaHandleBase* handle_base_ptr) : handle_base_ptr_{handle_base_ptr} {} void operator()(rd_kafka_t* handle); private: const KafkaHandleBase * handle_base_ptr_; }; using HandlePtr = std::unique_ptr; using TopicConfigurationMap = std::unordered_map; Topic get_topic(const std::string& name, rd_kafka_topic_conf_t* conf); Metadata get_metadata(bool all_topics, rd_kafka_topic_t* topic_ptr, std::chrono::milliseconds timeout) const; GroupInformationList fetch_consumer_groups(const char* name, std::chrono::milliseconds timeout); void save_topic_config(const std::string& topic_name, TopicConfiguration config); std::chrono::milliseconds timeout_ms_; Configuration config_; TopicConfigurationMap topic_configurations_; std::mutex topic_configurations_mutex_; HandlePtr handle_; int destroy_flags_; }; } // cppkafka #endif // CPPKAFKA_KAFKA_HANDLE_BASE_H