226 lines
5.9 KiB
C
226 lines
5.9 KiB
C
/*
|
|
* librdkafka - Apache Kafka C library
|
|
*
|
|
* Copyright (c) 2012, Magnus Edenhill
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* 1. Redistributions of source code must retain the above copyright notice,
|
|
* this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright notice,
|
|
* this list of conditions and the following disclaimer in the documentation
|
|
* and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
/**
|
|
* Apache Kafka consumer & producer example programs
|
|
* using the Kafka driver from librdkafka
|
|
* (https://github.com/edenhill/librdkafka)
|
|
*/
|
|
|
|
#include <ctype.h>
|
|
#include <signal.h>
|
|
|
|
/* Typical include path would be <librdkafka/rdkafkah>, but this program
|
|
* is builtin from within the librdkafka source tree and thus differs. */
|
|
#include "rdkafka.h" /* for Kafka driver */
|
|
|
|
|
|
static int run = 1;
|
|
|
|
static void stop (int sig) {
|
|
run = 0;
|
|
}
|
|
|
|
|
|
static void hexdump (FILE *fp, const char *name, const void *ptr, size_t len) {
|
|
const char *p = (const char *)ptr;
|
|
int of = 0;
|
|
|
|
|
|
if (name)
|
|
fprintf(fp, "%s hexdump (%lu bytes):\n", name, len);
|
|
|
|
for (of = 0 ; of < len ; of += 16) {
|
|
char hexen[16*3+1];
|
|
char charen[16+1];
|
|
int hof = 0;
|
|
|
|
int cof = 0;
|
|
int i;
|
|
|
|
for (i = of ; i < of + 16 && i < len ; i++) {
|
|
hof += sprintf(hexen+hof, "%02x ", p[i] & 0xff);
|
|
cof += sprintf(charen+cof, "%c",
|
|
isprint(p[i]) ? p[i] : '.');
|
|
}
|
|
fprintf(fp, "%08x: %-48s %-16s\n",
|
|
of, hexen, charen);
|
|
}
|
|
}
|
|
|
|
|
|
|
|
|
|
int main (int argc, char **argv) {
|
|
rd_kafka_t *rk;
|
|
char *broker = NULL;
|
|
char mode = 'C';
|
|
char *topic = NULL;
|
|
int partition = 0;
|
|
int opt;
|
|
|
|
|
|
while ((opt = getopt(argc, argv, "PCt:p:b:")) != -1) {
|
|
switch (opt) {
|
|
case 'P':
|
|
case 'C':
|
|
mode = opt;
|
|
break;
|
|
case 't':
|
|
topic = optarg;
|
|
break;
|
|
case 'p':
|
|
partition = atoi(optarg);
|
|
break;
|
|
case 'b':
|
|
broker = optarg;
|
|
break;
|
|
default:
|
|
goto usage;
|
|
}
|
|
}
|
|
|
|
if (!topic || optind != argc) {
|
|
usage:
|
|
fprintf(stderr,
|
|
"Usage: %s [-C|-P] -t <topic> "
|
|
"[-p <partition>] [-b <broker>]\n"
|
|
"\n"
|
|
" Options:\n"
|
|
" -C | -P Consumer or Producer mode\n"
|
|
" -t <topic> Topic to fetch / produce\n"
|
|
" -p <num> Partition (defaults to 0)\n"
|
|
" -b <broker> Broker address (localhost:9092)\n"
|
|
"\n"
|
|
" In Consumer mode:\n"
|
|
" writes fetched messages to stdout\n"
|
|
" In Producer mode:\n"
|
|
" reads messages from stdin and sends to broker\n"
|
|
"\n",
|
|
argv[0]);
|
|
exit(1);
|
|
}
|
|
|
|
|
|
signal(SIGINT, stop);
|
|
|
|
if (mode == 'P') {
|
|
/*
|
|
* Producer
|
|
*/
|
|
char buf[1024];
|
|
|
|
/* Create Kafka handle */
|
|
if (!(rk = rd_kafka_new(RD_KAFKA_PRODUCER, broker, NULL))) {
|
|
perror("kafka_new producer");
|
|
exit(1);
|
|
}
|
|
|
|
fprintf(stderr, "%% Type stuff and hit enter to send\n");
|
|
while (run && (fgets(buf, sizeof(buf), stdin))) {
|
|
int len = strlen(buf);
|
|
/* Send/Produce message. */
|
|
rd_kafka_produce(rk, topic, partition, 0, buf, len);
|
|
fprintf(stderr, "%% Sent %i bytes to topic "
|
|
"%s partition %i\n", len, topic, partition);
|
|
}
|
|
|
|
/* Destroy the handle */
|
|
rd_kafka_destroy(rk);
|
|
|
|
} else if (mode == 'C') {
|
|
/*
|
|
* Consumer
|
|
*/
|
|
rd_kafka_op_t *rko;
|
|
/* Base our configuration on the default config. */
|
|
rd_kafka_conf_t conf = rd_kafka_defaultconf;
|
|
|
|
|
|
/* The offset storage file is optional but its presence
|
|
* avoids starting all over from offset 0 again when
|
|
* the program restarts.
|
|
* ZooKeeper functionality will be implemented in future
|
|
* versions and then the offset will be stored there instead. */
|
|
conf.consumer.offset_file = "."; /* current directory */
|
|
|
|
/* Indicate to rdkafka that the application is responsible
|
|
* for storing the offset. This allows the application to
|
|
* succesfully handle a message before storing the offset.
|
|
* If this flag is not set rdkafka will store the offset
|
|
* just prior to returning the message from rd_kafka_consume().
|
|
*/
|
|
conf.flags |= RD_KAFKA_CONF_F_APP_OFFSET_STORE;
|
|
|
|
|
|
|
|
/* Use the consumer convenience function
|
|
* to create a Kafka handle. */
|
|
if (!(rk = rd_kafka_new_consumer(broker, topic,
|
|
(uint32_t)partition,
|
|
0, &conf))) {
|
|
perror("kafka_new_consumer");
|
|
exit(1);
|
|
}
|
|
|
|
while (run) {
|
|
/* Fetch an "op" which is one of:
|
|
* - a kafka message (if rko_len>0 && rko_err==0)
|
|
* - an error (if rko_err)
|
|
*/
|
|
if (!(rko = rd_kafka_consume(rk, 1000/*timeout ms*/)))
|
|
continue;
|
|
|
|
if (rko->rko_err)
|
|
fprintf(stderr, "%% Error: %.*s\n",
|
|
rko->rko_len, rko->rko_payload);
|
|
else if (rko->rko_len) {
|
|
fprintf(stderr, "%% Message with "
|
|
"next-offset %"PRIu64" is %i bytes\n",
|
|
rko->rko_offset, rko->rko_len);
|
|
hexdump(stdout, "Message",
|
|
rko->rko_payload, rko->rko_len);
|
|
}
|
|
|
|
/* rko_offset contains the offset of the _next_
|
|
* message. We store it when we're done processing
|
|
* the current message. */
|
|
if (rko->rko_offset)
|
|
rd_kafka_offset_store(rk, rko->rko_offset);
|
|
|
|
/* Destroy the op */
|
|
rd_kafka_op_destroy(rk, rko);
|
|
}
|
|
|
|
/* Destroy the handle */
|
|
rd_kafka_destroy(rk);
|
|
}
|
|
|
|
return 0;
|
|
}
|