mirror of
https://github.com/rsyslog/rsyslog.git
synced 2025-12-19 02:10:42 +01:00
1082 lines
32 KiB
C
1082 lines
32 KiB
C
/* omkafka.c
|
|
* This output plugin make rsyslog talk to Apache Kafka.
|
|
*
|
|
* Copyright 2014 by Adiscon GmbH.
|
|
*
|
|
* This file is part of rsyslog.
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
* -or-
|
|
* see COPYING.ASL20 in the source distribution
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
#include "config.h"
|
|
#include "rsyslog.h"
|
|
#include <stdio.h>
|
|
#include <stdarg.h>
|
|
#include <stdlib.h>
|
|
#include <string.h>
|
|
#include <assert.h>
|
|
#include <errno.h>
|
|
#include <fcntl.h>
|
|
#include <pthread.h>
|
|
#include <sys/uio.h>
|
|
#include <librdkafka/rdkafka.h>
|
|
#include "conf.h"
|
|
#include "syslogd-types.h"
|
|
#include "srUtils.h"
|
|
#include "template.h"
|
|
#include "module-template.h"
|
|
#include "errmsg.h"
|
|
#include "atomic.h"
|
|
#include "statsobj.h"
|
|
#include "unicode-helper.h"
|
|
|
|
MODULE_TYPE_OUTPUT
|
|
MODULE_TYPE_NOKEEP
|
|
MODULE_CNFNAME("omkafka")
|
|
|
|
/* internal structures
|
|
*/
|
|
DEF_OMOD_STATIC_DATA
|
|
DEFobjCurrIf(errmsg)
|
|
DEFobjCurrIf(statsobj)
|
|
|
|
statsobj_t *kafkaStats;
|
|
int ctrQueueSize;
|
|
STATSCOUNTER_DEF(ctrTopicSubmit, mutCtrTopicSubmit);
|
|
STATSCOUNTER_DEF(ctrKafkaFail, mutCtrKafkaFail);
|
|
STATSCOUNTER_DEF(ctrCacheMiss, mutCtrCacheMiss);
|
|
STATSCOUNTER_DEF(ctrCacheEvict, mutCtrCacheEvict);
|
|
STATSCOUNTER_DEF(ctrCacheSkip, mutCtrCacheSkip);
|
|
|
|
#define MAX_ERRMSG 1024 /* max size of error messages that we support */
|
|
|
|
#define NO_FIXED_PARTITION -1 /* signifies that no fixed partition config exists */
|
|
|
|
struct kafka_params {
|
|
const char *name;
|
|
const char *val;
|
|
};
|
|
|
|
#if HAVE_ATOMIC_BUILTINS64
|
|
static uint64 clockTopicAccess = 0;
|
|
#else
|
|
static unsigned clockTopicAccess = 0;
|
|
#endif
|
|
/* and the "tick" function */
|
|
#ifndef HAVE_ATOMIC_BUILTINS
|
|
static pthread_mutex_t mutClock;
|
|
#endif
|
|
static inline uint64
|
|
getClockTopicAccess(void)
|
|
{
|
|
#if HAVE_ATOMIC_BUILTINS64
|
|
return ATOMIC_INC_AND_FETCH_uint64(&clockTopicAccess, &mutClock);
|
|
#else
|
|
return ATOMIC_INC_AND_FETCH_unsigned(&clockTopicAccess, &mutClock);
|
|
#endif
|
|
}
|
|
|
|
static int closeTimeout = 1000;
|
|
static pthread_mutex_t closeTimeoutMut = PTHREAD_MUTEX_INITIALIZER;
|
|
|
|
/* dynamic topic cache */
|
|
struct s_dynaTopicCacheEntry {
|
|
uchar *pName;
|
|
rd_kafka_topic_t *pTopic;
|
|
uint64 clkTickAccessed;
|
|
pthread_rwlock_t lock;
|
|
};
|
|
typedef struct s_dynaTopicCacheEntry dynaTopicCacheEntry;
|
|
|
|
typedef struct _instanceData {
|
|
uchar *topic;
|
|
sbool dynaTopic;
|
|
dynaTopicCacheEntry **dynCache;
|
|
pthread_mutex_t mutDynCache;
|
|
rd_kafka_topic_t *pTopic;
|
|
int iCurrElt;
|
|
int iCurrCacheSize;
|
|
int bReportErrs;
|
|
int iDynaTopicCacheSize;
|
|
uchar *tplName; /* assigned output template */
|
|
char *brokers;
|
|
sbool autoPartition;
|
|
int fixedPartition;
|
|
int nPartitions;
|
|
uint32_t currPartition;
|
|
int nConfParams;
|
|
struct kafka_params *confParams;
|
|
int nTopicConfParams;
|
|
struct kafka_params *topicConfParams;
|
|
uchar *errorFile;
|
|
uchar *key;
|
|
int fdErrFile; /* error file fd or -1 if not open */
|
|
pthread_mutex_t mutErrFile;
|
|
int bIsOpen;
|
|
pthread_rwlock_t rkLock;
|
|
rd_kafka_t *rk;
|
|
int closeTimeout;
|
|
} instanceData;
|
|
|
|
typedef struct wrkrInstanceData {
|
|
instanceData *pData;
|
|
} wrkrInstanceData_t;
|
|
|
|
|
|
/* tables for interfacing with the v6 config system */
|
|
/* action (instance) parameters */
|
|
static struct cnfparamdescr actpdescr[] = {
|
|
{ "topic", eCmdHdlrString, CNFPARAM_REQUIRED },
|
|
{ "dynatopic", eCmdHdlrBinary, 0 },
|
|
{ "dynatopic.cachesize", eCmdHdlrInt, 0 },
|
|
{ "partitions.auto", eCmdHdlrBinary, 0 }, /* use librdkafka's automatic partitioning function */
|
|
{ "partitions.number", eCmdHdlrPositiveInt, 0 },
|
|
{ "partitions.usefixed", eCmdHdlrNonNegInt, 0 }, /* expert parameter, "nails" partition */
|
|
{ "broker", eCmdHdlrArray, 0 },
|
|
{ "confparam", eCmdHdlrArray, 0 },
|
|
{ "topicconfparam", eCmdHdlrArray, 0 },
|
|
{ "errorfile", eCmdHdlrGetWord, 0 },
|
|
{ "key", eCmdHdlrGetWord, 0 },
|
|
{ "template", eCmdHdlrGetWord, 0 },
|
|
{ "closeTimeout", eCmdHdlrPositiveInt, 0 }
|
|
};
|
|
static struct cnfparamblk actpblk =
|
|
{ CNFPARAMBLK_VERSION,
|
|
sizeof(actpdescr)/sizeof(struct cnfparamdescr),
|
|
actpdescr
|
|
};
|
|
|
|
BEGINinitConfVars /* (re)set config variables to default values */
|
|
CODESTARTinitConfVars
|
|
ENDinitConfVars
|
|
|
|
static inline uint32_t
|
|
getPartition(instanceData *const __restrict__ pData)
|
|
{
|
|
if (pData->autoPartition) {
|
|
return RD_KAFKA_PARTITION_UA;
|
|
} else {
|
|
return (pData->fixedPartition == NO_FIXED_PARTITION) ?
|
|
ATOMIC_INC_AND_FETCH_unsigned(&pData->currPartition,
|
|
&pData->mutCurrPartition) % pData->nPartitions
|
|
: (unsigned) pData->fixedPartition;
|
|
}
|
|
}
|
|
|
|
/* must always be called with appropriate locks taken */
|
|
static void
|
|
d_free_topic(rd_kafka_topic_t **topic)
|
|
{
|
|
if (*topic != NULL) {
|
|
DBGPRINTF("omkafka: closing topic %s\n", rd_kafka_topic_name(*topic));
|
|
rd_kafka_topic_destroy(*topic);
|
|
*topic = NULL;
|
|
}
|
|
}
|
|
|
|
/* destroy topic item */
|
|
/* must be called with write(rkLock) */
|
|
static void
|
|
closeTopic(instanceData *__restrict__ const pData)
|
|
{
|
|
d_free_topic(&pData->pTopic);
|
|
}
|
|
|
|
/* these dynaTopic* functions are only slightly modified versions of those found in omfile.c.
|
|
* check the sources in omfile.c for more descriptive comments about each of these functions.
|
|
* i will only put the bare descriptions in this one. 2015-01-09 - Tait Clarridge
|
|
*/
|
|
|
|
/* delete a cache entry from the dynamic topic cache */
|
|
/* must be called with lock(mutDynCache) */
|
|
static rsRetVal
|
|
dynaTopicDelCacheEntry(instanceData *__restrict__ const pData, const int iEntry, const int bFreeEntry)
|
|
{
|
|
dynaTopicCacheEntry **pCache = pData->dynCache;
|
|
DEFiRet;
|
|
ASSERT(pCache != NULL);
|
|
|
|
if(pCache[iEntry] == NULL)
|
|
FINALIZE;
|
|
pthread_rwlock_wrlock(&pCache[iEntry]->lock);
|
|
|
|
DBGPRINTF("Removing entry %d for topic '%s' from dynaCache.\n", iEntry,
|
|
pCache[iEntry]->pName == NULL ? UCHAR_CONSTANT("[OPEN FAILED]") : pCache[iEntry]->pName);
|
|
|
|
if(pCache[iEntry]->pName != NULL) {
|
|
d_free(pCache[iEntry]->pName);
|
|
pCache[iEntry]->pName = NULL;
|
|
}
|
|
|
|
pthread_rwlock_unlock(&pCache[iEntry]->lock);
|
|
|
|
if(bFreeEntry) {
|
|
pthread_rwlock_destroy(&pCache[iEntry]->lock);
|
|
d_free(pCache[iEntry]);
|
|
pCache[iEntry] = NULL;
|
|
}
|
|
|
|
finalize_it:
|
|
RETiRet;
|
|
}
|
|
|
|
/* clear the entire dynamic topic cache */
|
|
static inline void
|
|
dynaTopicFreeCacheEntries(instanceData *__restrict__ const pData)
|
|
{
|
|
register int i;
|
|
ASSERT(pData != NULL);
|
|
|
|
BEGINfunc;
|
|
pthread_mutex_lock(&pData->mutDynCache);
|
|
for(i = 0 ; i < pData->iCurrCacheSize ; ++i) {
|
|
dynaTopicDelCacheEntry(pData, i, 1);
|
|
}
|
|
pData->iCurrElt = -1; /* invalidate current element */
|
|
pthread_mutex_unlock(&pData->mutDynCache);
|
|
ENDfunc;
|
|
}
|
|
|
|
/* create the topic object */
|
|
/* must be called with _atleast_ read(rkLock) */
|
|
static rsRetVal
|
|
createTopic(instanceData *__restrict__ const pData, const uchar *__restrict__ const newTopicName, rd_kafka_topic_t** topic) {
|
|
/* Get a new topic conf */
|
|
rd_kafka_topic_conf_t *const topicconf = rd_kafka_topic_conf_new();
|
|
char errstr[MAX_ERRMSG];
|
|
rd_kafka_topic_t *rkt = NULL;
|
|
DEFiRet;
|
|
|
|
*topic = NULL;
|
|
|
|
if(topicconf == NULL) {
|
|
errmsg.LogError(0, RS_RET_KAFKA_ERROR,
|
|
"omkafka: error creating kafka topic conf obj: %s\n",
|
|
rd_kafka_err2str(rd_kafka_errno2err(errno)));
|
|
ABORT_FINALIZE(RS_RET_KAFKA_ERROR);
|
|
}
|
|
for(int i = 0 ; i < pData->nTopicConfParams ; ++i) {
|
|
if(rd_kafka_topic_conf_set(topicconf,
|
|
pData->topicConfParams[i].name,
|
|
pData->topicConfParams[i].val,
|
|
errstr, sizeof(errstr))
|
|
!= RD_KAFKA_CONF_OK) {
|
|
if(pData->bReportErrs) {
|
|
errmsg.LogError(0, RS_RET_PARAM_ERROR, "error in kafka "
|
|
"topic conf parameter '%s=%s': %s",
|
|
pData->topicConfParams[i].name,
|
|
pData->topicConfParams[i].val, errstr);
|
|
}
|
|
ABORT_FINALIZE(RS_RET_PARAM_ERROR);
|
|
}
|
|
}
|
|
rkt = rd_kafka_topic_new(pData->rk, (char *)newTopicName, topicconf);
|
|
if(rkt == NULL) {
|
|
errmsg.LogError(0, RS_RET_KAFKA_ERROR,
|
|
"omkafka: error creating kafka topic: %s\n",
|
|
rd_kafka_err2str(rd_kafka_errno2err(errno)));
|
|
ABORT_FINALIZE(RS_RET_KAFKA_ERROR);
|
|
}
|
|
*topic = rkt;
|
|
finalize_it:
|
|
RETiRet;
|
|
}
|
|
|
|
/* create the topic object */
|
|
/* must be called with write(rkLock) */
|
|
static rsRetVal
|
|
prepareTopic(instanceData *__restrict__ const pData, const uchar *__restrict__ const newTopicName)
|
|
{
|
|
DEFiRet;
|
|
CHKiRet(createTopic(pData, newTopicName, &pData->pTopic));
|
|
finalize_it:
|
|
if(iRet != RS_RET_OK) {
|
|
if(pData->pTopic != NULL) {
|
|
closeTopic(pData);
|
|
}
|
|
}
|
|
RETiRet;
|
|
}
|
|
|
|
/* check dynamic topic cache for existence of the already created topic.
|
|
* if it does not exist, create a new one, or if we are currently using it
|
|
* as of the last message, keep using it.
|
|
*
|
|
* important: topic is returned read-locked, user must unlock it after using it.
|
|
*
|
|
* must be called with read(rkLock)
|
|
*/
|
|
static inline rsRetVal
|
|
prepareDynTopic(instanceData *__restrict__ const pData, const uchar *__restrict__ const newTopicName,
|
|
rd_kafka_topic_t** topic, pthread_rwlock_t** lock)
|
|
{
|
|
uint64 ctOldest;
|
|
int iOldest;
|
|
int i;
|
|
int iFirstFree;
|
|
rsRetVal localRet;
|
|
dynaTopicCacheEntry **pCache;
|
|
dynaTopicCacheEntry *entry = NULL;
|
|
rd_kafka_topic_t *tmpTopic = NULL;
|
|
DEFiRet;
|
|
ASSERT(pData != NULL);
|
|
ASSERT(newTopicName != NULL);
|
|
|
|
pthread_mutex_lock(&pData->mutDynCache);
|
|
|
|
pCache = pData->dynCache;
|
|
/* first check, if we still have the current topic */
|
|
if ((pData->iCurrElt != -1)
|
|
&& !ustrcmp(newTopicName, pCache[pData->iCurrElt]->pName)) {
|
|
/* great, we are all set */
|
|
pCache[pData->iCurrElt]->clkTickAccessed = getClockTopicAccess();
|
|
entry = pCache[pData->iCurrElt];
|
|
STATSCOUNTER_INC(ctrCacheSkip, mutCtrCacheSkip);
|
|
FINALIZE;
|
|
}
|
|
|
|
/* ok, no luck. Now let's search the table if we find a matching spot.
|
|
* While doing so, we also prepare for creation of a new one.
|
|
*/
|
|
pData->iCurrElt = -1;
|
|
iFirstFree = -1;
|
|
iOldest = 0;
|
|
ctOldest = getClockTopicAccess();
|
|
for(i = 0 ; i < pData->iCurrCacheSize ; ++i) {
|
|
if(pCache[i] == NULL || pCache[i]->pName == NULL) {
|
|
if(iFirstFree == -1)
|
|
iFirstFree = i;
|
|
} else { /*got an element, let's see if it matches */
|
|
if(!ustrcmp(newTopicName, pCache[i]->pName)) {
|
|
/* we found our element! */
|
|
entry = pCache[i];
|
|
pData->iCurrElt = i;
|
|
pCache[i]->clkTickAccessed = getClockTopicAccess(); /* update "timestamp" for LRU */
|
|
FINALIZE;
|
|
}
|
|
/* did not find it - so lets keep track of the counters for LRU */
|
|
if(pCache[i]->clkTickAccessed < ctOldest) {
|
|
ctOldest = pCache[i]->clkTickAccessed;
|
|
iOldest = i;
|
|
}
|
|
}
|
|
}
|
|
STATSCOUNTER_INC(ctrCacheMiss, mutCtrCacheMiss);
|
|
|
|
/* invalidate iCurrElt as we may error-exit out of this function when the currrent
|
|
* iCurrElt has been freed or otherwise become unusable. This is a precaution, and
|
|
* performance-wise it may be better to do that in each of the exits. However, that
|
|
* is error-prone, so I prefer to do it here. -- rgerhards, 2010-03-02
|
|
*/
|
|
pData->iCurrElt = -1;
|
|
|
|
if(iFirstFree == -1 && (pData->iCurrCacheSize < pData->iDynaTopicCacheSize)) {
|
|
/* there is space left, so set it to that index */
|
|
iFirstFree = pData->iCurrCacheSize++;
|
|
}
|
|
|
|
if(iFirstFree == -1) {
|
|
dynaTopicDelCacheEntry(pData, iOldest, 0);
|
|
STATSCOUNTER_INC(ctrCacheEvict, mutCtrCacheEvict);
|
|
iFirstFree = iOldest; /* this one *is* now free ;) */
|
|
} else {
|
|
/* we need to allocate memory for the cache structure */
|
|
CHKmalloc(pCache[iFirstFree] = (dynaTopicCacheEntry*) calloc(1, sizeof(dynaTopicCacheEntry)));
|
|
CHKiRet(pthread_rwlock_init(&pCache[iFirstFree]->lock, NULL));
|
|
}
|
|
|
|
/* Ok, we finally can open the topic */
|
|
localRet = createTopic(pData, newTopicName, &tmpTopic);
|
|
|
|
if(localRet != RS_RET_OK) {
|
|
errmsg.LogError(0, localRet, "Could not open dynamic topic '%s' [state %d] - discarding message", newTopicName, localRet);
|
|
ABORT_FINALIZE(localRet);
|
|
}
|
|
|
|
if((pCache[iFirstFree]->pName = ustrdup(newTopicName)) == NULL) {
|
|
d_free_topic(&tmpTopic);
|
|
ABORT_FINALIZE(RS_RET_OUT_OF_MEMORY);
|
|
}
|
|
pCache[iFirstFree]->pTopic = tmpTopic;
|
|
pCache[iFirstFree]->clkTickAccessed = getClockTopicAccess();
|
|
entry = pCache[iFirstFree];
|
|
pData->iCurrElt = iFirstFree;
|
|
DBGPRINTF("Added new entry %d for topic cache, topic '%s'.\n", iFirstFree, newTopicName);
|
|
|
|
finalize_it:
|
|
if (iRet == RS_RET_OK && entry != NULL) {
|
|
pthread_rwlock_rdlock(&entry->lock);
|
|
}
|
|
pthread_mutex_unlock(&pData->mutDynCache);
|
|
if (iRet == RS_RET_OK && entry != NULL) {
|
|
*topic = entry->pTopic;
|
|
*lock = &entry->lock;
|
|
} else {
|
|
*topic = NULL;
|
|
*lock = NULL;
|
|
}
|
|
RETiRet;
|
|
}
|
|
|
|
/* write data error request/replies to separate error file
|
|
* Note: we open the file but never close it before exit. If it
|
|
* needs to be closed, HUP must be sent.
|
|
*/
|
|
static rsRetVal
|
|
writeDataError(instanceData *const pData,
|
|
const char *const __restrict__ data,
|
|
const size_t lenData,
|
|
const int kafkaErr)
|
|
{
|
|
int bLocked = 0;
|
|
struct json_object *json = NULL;
|
|
DEFiRet;
|
|
|
|
if(pData->errorFile == NULL) {
|
|
FINALIZE;
|
|
}
|
|
|
|
json = json_object_new_object();
|
|
if(json == NULL) {
|
|
ABORT_FINALIZE(RS_RET_ERR);
|
|
}
|
|
struct json_object *jval;
|
|
jval = json_object_new_int(kafkaErr);
|
|
json_object_object_add(json, "errcode", jval);
|
|
jval = json_object_new_string(rd_kafka_err2str(kafkaErr));
|
|
json_object_object_add(json, "errmsg", jval);
|
|
jval = json_object_new_string_len(data, lenData);
|
|
json_object_object_add(json, "data", jval);
|
|
|
|
struct iovec iov[2];
|
|
iov[0].iov_base = (void*) json_object_get_string(json);
|
|
iov[0].iov_len = strlen(iov[0].iov_base);
|
|
iov[1].iov_base = "\n";
|
|
iov[1].iov_len = 1;
|
|
|
|
/* we must protect the file write do operations due to other wrks & HUP */
|
|
pthread_mutex_lock(&pData->mutErrFile);
|
|
bLocked = 1;
|
|
if(pData->fdErrFile == -1) {
|
|
pData->fdErrFile = open((char*)pData->errorFile,
|
|
O_WRONLY|O_CREAT|O_APPEND|O_LARGEFILE|O_CLOEXEC,
|
|
S_IRUSR|S_IWUSR|S_IRGRP|S_IWGRP);
|
|
if(pData->fdErrFile == -1) {
|
|
char errStr[1024];
|
|
rs_strerror_r(errno, errStr, sizeof(errStr));
|
|
DBGPRINTF("omkafka: error opening error file: %s\n", errStr);
|
|
ABORT_FINALIZE(RS_RET_ERR);
|
|
}
|
|
}
|
|
|
|
/* Note: we do not do real error-handling on the err file, as this
|
|
* complicates things way to much.
|
|
*/
|
|
const ssize_t nwritten = writev(pData->fdErrFile, iov, sizeof(iov)/sizeof(struct iovec));
|
|
if(nwritten != (ssize_t) iov[0].iov_len + 1) {
|
|
DBGPRINTF("omkafka: error %d writing error file, write returns %lld\n",
|
|
errno, (long long) nwritten);
|
|
}
|
|
|
|
finalize_it:
|
|
if(bLocked)
|
|
pthread_mutex_unlock(&pData->mutErrFile);
|
|
if(json != NULL)
|
|
json_object_put(json);
|
|
RETiRet;
|
|
}
|
|
|
|
static void
|
|
deliveryCallback(rd_kafka_t __attribute__((unused)) *rk,
|
|
void *payload, size_t len,
|
|
int error_code,
|
|
void *opaque, void __attribute__((unused)) *msg_opaque)
|
|
{
|
|
instanceData *const pData = (instanceData *) opaque;
|
|
if(error_code != 0)
|
|
writeDataError(pData, (char*) payload, len, error_code);
|
|
}
|
|
|
|
static void
|
|
kafkaLogger(const rd_kafka_t __attribute__((unused)) *rk, int level,
|
|
const char *fac, const char *buf)
|
|
{
|
|
DBGPRINTF("omkafka: kafka log message [%d,%s]: %s\n",
|
|
level, fac, buf);
|
|
}
|
|
|
|
/* should be called with write(rkLock) */
|
|
static inline void
|
|
do_rd_kafka_destroy(instanceData *const __restrict pData)
|
|
{
|
|
if (pData->rk == NULL) {
|
|
DBGPRINTF("omkafka: can't close, handle wasn't open\n");
|
|
} else {
|
|
int queuedCount = rd_kafka_outq_len(pData->rk);
|
|
DBGPRINTF("omkafka: closing - items left in outqueue: %d\n", queuedCount);
|
|
|
|
struct timespec tOut;
|
|
timeoutComp(&tOut, pData->closeTimeout);
|
|
|
|
while (timeoutVal(&tOut) > 0) {
|
|
queuedCount = rd_kafka_outq_len(pData->rk);
|
|
if (queuedCount > 0) {
|
|
rd_kafka_poll(pData->rk, 10);
|
|
} else {
|
|
break;
|
|
}
|
|
}
|
|
if (queuedCount > 0) {
|
|
DBGPRINTF("omkafka: queue-drain for close timed-out took too long, "
|
|
"items left in outqueue: %d\n",
|
|
rd_kafka_outq_len(pData->rk));
|
|
}
|
|
if (pData->dynaTopic) {
|
|
dynaTopicFreeCacheEntries(pData);
|
|
} else {
|
|
closeTopic(pData);
|
|
}
|
|
rd_kafka_destroy(pData->rk);
|
|
pData->rk = NULL;
|
|
}
|
|
}
|
|
|
|
/* should be called with write(rkLock) */
|
|
static void
|
|
closeKafka(instanceData *const __restrict__ pData)
|
|
{
|
|
if(pData->bIsOpen) {
|
|
do_rd_kafka_destroy(pData);
|
|
pData->bIsOpen = 0;
|
|
}
|
|
}
|
|
|
|
static void
|
|
errorCallback(rd_kafka_t __attribute__((unused)) *rk,
|
|
int __attribute__((unused)) err,
|
|
const char *reason,
|
|
void __attribute__((unused)) *opaque)
|
|
{
|
|
errmsg.LogError(0, RS_RET_KAFKA_ERROR,
|
|
"omkafka: kafka message %s", reason);
|
|
}
|
|
|
|
|
|
|
|
#if 0 /* the stock librdkafka version in Ubuntu 14.04 LTS does NOT support metadata :-( */
|
|
/* Note: this is a skeleton, with some code missing--> add it when it is actually implemented. */
|
|
static int
|
|
getConfiguredPartitions()
|
|
{
|
|
struct rd_kafka_metadata *pMetadata;
|
|
if(rd_kafka_metadata(pData->rk, 0, rkt, &pMetadata, 8)
|
|
== RD_KAFKA_RESP_ERR_NO_ERROR) {
|
|
dbgprintf("omkafka: topic '%s' has %d partitions\n",
|
|
pData->topic, pMetadata->topics[0]->partition_cnt);
|
|
rd_kafka_metadata_destroy(pMetadata);
|
|
} else {
|
|
dbgprintf("omkafka: error reading metadata\n");
|
|
// TODO: handle this gracefull **when** we actually need
|
|
// the metadata -- or remove completely. 2014-12-12 rgerhards
|
|
}
|
|
}
|
|
#endif
|
|
|
|
/* should be called with write(rkLock) */
|
|
static rsRetVal
|
|
openKafka(instanceData *const __restrict__ pData)
|
|
{
|
|
char errstr[MAX_ERRMSG];
|
|
int nBrokers = 0;
|
|
DEFiRet;
|
|
|
|
if(pData->bIsOpen)
|
|
FINALIZE;
|
|
|
|
pData->pTopic = NULL;
|
|
|
|
/* main conf */
|
|
rd_kafka_conf_t *const conf = rd_kafka_conf_new();
|
|
if(conf == NULL) {
|
|
errmsg.LogError(0, RS_RET_KAFKA_ERROR,
|
|
"omkafka: error creating kafka conf obj: %s\n",
|
|
rd_kafka_err2str(rd_kafka_errno2err(errno)));
|
|
ABORT_FINALIZE(RS_RET_KAFKA_ERROR);
|
|
}
|
|
for(int i = 0 ; i < pData->nConfParams ; ++i) {
|
|
if(rd_kafka_conf_set(conf,
|
|
pData->confParams[i].name,
|
|
pData->confParams[i].val,
|
|
errstr, sizeof(errstr))
|
|
!= RD_KAFKA_CONF_OK) {
|
|
if(pData->bReportErrs) {
|
|
errmsg.LogError(0, RS_RET_PARAM_ERROR, "error in kafka "
|
|
"parameter '%s=%s': %s",
|
|
pData->confParams[i].name,
|
|
pData->confParams[i].val, errstr);
|
|
}
|
|
ABORT_FINALIZE(RS_RET_PARAM_ERROR);
|
|
}
|
|
}
|
|
rd_kafka_conf_set_opaque(conf, (void *) pData);
|
|
rd_kafka_conf_set_dr_cb(conf, deliveryCallback);
|
|
rd_kafka_conf_set_error_cb(conf, errorCallback);
|
|
|
|
char kafkaErrMsg[1024];
|
|
pData->rk = rd_kafka_new(RD_KAFKA_PRODUCER, conf,
|
|
kafkaErrMsg, sizeof(kafkaErrMsg));
|
|
if(pData->rk == NULL) {
|
|
errmsg.LogError(0, RS_RET_KAFKA_ERROR,
|
|
"omkafka: error creating kafka handle: %s\n", kafkaErrMsg);
|
|
ABORT_FINALIZE(RS_RET_KAFKA_ERROR);
|
|
}
|
|
rd_kafka_set_logger(pData->rk, kafkaLogger);
|
|
if((nBrokers = rd_kafka_brokers_add(pData->rk, (char*)pData->brokers)) == 0) {
|
|
errmsg.LogError(0, RS_RET_KAFKA_NO_VALID_BROKERS,
|
|
"omkafka: no valid brokers specified: %s\n", pData->brokers);
|
|
ABORT_FINALIZE(RS_RET_KAFKA_NO_VALID_BROKERS);
|
|
}
|
|
|
|
pData->bIsOpen = 1;
|
|
finalize_it:
|
|
if(iRet == RS_RET_OK) {
|
|
pData->bReportErrs = 1;
|
|
} else {
|
|
pData->bReportErrs = 0;
|
|
if(pData->rk != NULL) {
|
|
do_rd_kafka_destroy(pData);
|
|
}
|
|
}
|
|
RETiRet;
|
|
}
|
|
|
|
static rsRetVal
|
|
setupKafkaHandle(instanceData *const __restrict__ pData, int recreate)
|
|
{
|
|
DEFiRet;
|
|
pthread_rwlock_wrlock(&pData->rkLock);
|
|
if (recreate) {
|
|
closeKafka(pData);
|
|
}
|
|
CHKiRet(openKafka(pData));
|
|
if (! pData->dynaTopic) {
|
|
if( pData->pTopic == NULL)
|
|
CHKiRet(prepareTopic(pData, pData->topic));
|
|
}
|
|
finalize_it:
|
|
if (iRet != RS_RET_OK) {
|
|
if (pData->rk != NULL) {
|
|
closeKafka(pData);
|
|
}
|
|
}
|
|
pthread_rwlock_unlock(&pData->rkLock);
|
|
RETiRet;
|
|
}
|
|
|
|
BEGINdoHUP
|
|
CODESTARTdoHUP
|
|
pthread_mutex_lock(&pData->mutErrFile);
|
|
if(pData->fdErrFile != -1) {
|
|
close(pData->fdErrFile);
|
|
pData->fdErrFile = -1;
|
|
}
|
|
pthread_mutex_unlock(&pData->mutErrFile);
|
|
CHKiRet(setupKafkaHandle(pData, 1));
|
|
finalize_it:
|
|
ENDdoHUP
|
|
|
|
BEGINcreateInstance
|
|
CODESTARTcreateInstance
|
|
pData->currPartition = 0;
|
|
pData->bIsOpen = 0;
|
|
pData->fdErrFile = -1;
|
|
pData->pTopic = NULL;
|
|
pData->bReportErrs = 1;
|
|
CHKiRet(pthread_mutex_init(&pData->mutErrFile, NULL));
|
|
CHKiRet(pthread_rwlock_init(&pData->rkLock, NULL));
|
|
CHKiRet(pthread_mutex_init(&pData->mutDynCache, NULL));
|
|
finalize_it:
|
|
ENDcreateInstance
|
|
|
|
|
|
BEGINcreateWrkrInstance
|
|
CODESTARTcreateWrkrInstance
|
|
ENDcreateWrkrInstance
|
|
|
|
|
|
BEGINisCompatibleWithFeature
|
|
CODESTARTisCompatibleWithFeature
|
|
ENDisCompatibleWithFeature
|
|
|
|
|
|
BEGINfreeInstance
|
|
CODESTARTfreeInstance
|
|
free(pData->errorFile);
|
|
free(pData->topic);
|
|
free(pData->brokers);
|
|
free(pData->tplName);
|
|
for(int i = 0 ; i < pData->nConfParams ; ++i) {
|
|
free((void*) pData->confParams[i].name);
|
|
free((void*) pData->confParams[i].val);
|
|
}
|
|
for(int i = 0 ; i < pData->nTopicConfParams ; ++i) {
|
|
free((void*) pData->topicConfParams[i].name);
|
|
free((void*) pData->topicConfParams[i].val);
|
|
}
|
|
if(pData->fdErrFile != -1)
|
|
close(pData->fdErrFile);
|
|
pthread_rwlock_wrlock(&pData->rkLock);
|
|
closeKafka(pData);
|
|
if(pData->dynaTopic && pData->dynCache != NULL) {
|
|
d_free(pData->dynCache);
|
|
pData->dynCache = NULL;
|
|
}
|
|
pthread_rwlock_unlock(&pData->rkLock);
|
|
pthread_rwlock_destroy(&pData->rkLock);
|
|
pthread_mutex_destroy(&pData->mutErrFile);
|
|
pthread_mutex_destroy(&pData->mutDynCache);
|
|
ENDfreeInstance
|
|
|
|
BEGINfreeWrkrInstance
|
|
CODESTARTfreeWrkrInstance
|
|
ENDfreeWrkrInstance
|
|
|
|
|
|
BEGINdbgPrintInstInfo
|
|
CODESTARTdbgPrintInstInfo
|
|
ENDdbgPrintInstInfo
|
|
|
|
|
|
BEGINtryResume
|
|
CODESTARTtryResume
|
|
/* Note: we do have an issue here: some errors unfortunately
|
|
* show up only when we actually produce data. As such, we cannot
|
|
* detect those kinds of problems in tryResume, what can yield
|
|
* in a false "RS_RET_OK" response in those cases. However, this
|
|
* problem is shared with other modules, and the rsyslog core engine
|
|
* has some mitigation against these kinds of problems. On the plus
|
|
* side, we can at least detect if something goes wrong during
|
|
* setupKafkaHandle(), and we save ourself any hassle in those cases.
|
|
* The current way is not ideal, but currently the best we can
|
|
* think of. If someone knows how to do a produce call without
|
|
* actually producing something (or otherwise check the health
|
|
* of the Kafka subsystem), please let us know on the rsyslog
|
|
* mailing list. -- rgerhards, 2014-12-14
|
|
*/
|
|
CHKiRet(setupKafkaHandle(pWrkrData->pData, 0));
|
|
|
|
finalize_it:
|
|
DBGPRINTF("omkafka: tryResume returned %d\n", iRet);
|
|
ENDtryResume
|
|
|
|
|
|
/* must be called with read(rkLock) */
|
|
static rsRetVal
|
|
writeKafka(instanceData *pData, uchar *msg, uchar *topic)
|
|
{
|
|
DEFiRet;
|
|
const int partition = getPartition(pData);
|
|
rd_kafka_topic_t *rkt = NULL;
|
|
pthread_rwlock_t *dynTopicLock = NULL;
|
|
int msg_enqueue_status = 0;
|
|
|
|
DBGPRINTF("omkafka: trying to send: key:'%s', msg:'%s'\n", pData->key, msg);
|
|
|
|
if(pData->dynaTopic) {
|
|
DBGPRINTF("omkafka: topic to insert to: %s\n", topic);
|
|
CHKiRet(prepareDynTopic(pData, topic, &rkt, &dynTopicLock));
|
|
} else {
|
|
rkt = pData->pTopic;
|
|
}
|
|
|
|
msg_enqueue_status = rd_kafka_produce(rkt, partition, RD_KAFKA_MSG_F_COPY,
|
|
msg, strlen((char*)msg), pData->key,
|
|
pData->key == NULL ? 0 : strlen((char*)pData->key),
|
|
NULL);
|
|
if(msg_enqueue_status == -1) {
|
|
errmsg.LogError(0, RS_RET_KAFKA_PRODUCE_ERR,
|
|
"omkafka: Failed to produce to topic '%s' "
|
|
"partition %d: %s\n",
|
|
rd_kafka_topic_name(rkt), partition,
|
|
rd_kafka_err2str(rd_kafka_errno2err(errno)));
|
|
}
|
|
const int callbacksCalled = rd_kafka_poll(pData->rk, 0); /* call callbacks */
|
|
if (pData->dynaTopic) {
|
|
pthread_rwlock_unlock(dynTopicLock);/* dynamic topic can't be used beyond this pt */
|
|
}
|
|
DBGPRINTF("omkafka: kafka outqueue length: %d, callbacks called %d\n",
|
|
rd_kafka_outq_len(pData->rk), callbacksCalled);
|
|
|
|
if (msg_enqueue_status == -1) {
|
|
STATSCOUNTER_INC(ctrKafkaFail, mutCtrKafkaFail);
|
|
ABORT_FINALIZE(RS_RET_KAFKA_PRODUCE_ERR);
|
|
/* ABORT_FINALIZE isn't absolutely necessary as of now,
|
|
because this is the last line anyway, but its useful to ensure
|
|
correctness in case we add more stuff below this line at some point*/
|
|
}
|
|
|
|
finalize_it:
|
|
DBGPRINTF("omkafka: writeKafka returned %d\n", iRet);
|
|
if(iRet != RS_RET_OK) {
|
|
iRet = RS_RET_SUSPENDED;
|
|
}
|
|
STATSCOUNTER_SETMAX_NOMUT(ctrQueueSize, rd_kafka_outq_len(pData->rk));
|
|
STATSCOUNTER_INC(ctrTopicSubmit, mutCtrTopicSubmit);
|
|
RETiRet;
|
|
}
|
|
|
|
|
|
BEGINdoAction
|
|
CODESTARTdoAction
|
|
instanceData *const pData = pWrkrData->pData;
|
|
if (! pData->bIsOpen)
|
|
CHKiRet(setupKafkaHandle(pData, 0));
|
|
|
|
pthread_rwlock_rdlock(&pData->rkLock);
|
|
|
|
/* support dynamic topic */
|
|
if(pData->dynaTopic)
|
|
iRet = writeKafka(pData, ppString[0], ppString[1]);
|
|
else
|
|
iRet = writeKafka(pData, ppString[0], pData->topic);
|
|
|
|
pthread_rwlock_unlock(&pData->rkLock);
|
|
finalize_it:
|
|
ENDdoAction
|
|
|
|
|
|
static inline void
|
|
setInstParamDefaults(instanceData *pData)
|
|
{
|
|
pData->topic = NULL;
|
|
pData->dynaTopic = 0;
|
|
pData->iDynaTopicCacheSize = 50;
|
|
pData->brokers = NULL;
|
|
pData->autoPartition = 0;
|
|
pData->fixedPartition = NO_FIXED_PARTITION;
|
|
pData->nPartitions = 1;
|
|
pData->nConfParams = 0;
|
|
pData->confParams = NULL;
|
|
pData->nTopicConfParams = 0;
|
|
pData->topicConfParams = NULL;
|
|
pData->errorFile = NULL;
|
|
pData->key = NULL;
|
|
pData->closeTimeout = 2000;
|
|
}
|
|
|
|
static rsRetVal
|
|
processKafkaParam(char *const param,
|
|
const char **const name,
|
|
const char **const paramval)
|
|
{
|
|
DEFiRet;
|
|
char *val = strstr(param, "=");
|
|
if(val == NULL) {
|
|
errmsg.LogError(0, RS_RET_PARAM_ERROR, "missing equal sign in "
|
|
"parameter '%s'", param);
|
|
ABORT_FINALIZE(RS_RET_PARAM_ERROR);
|
|
}
|
|
*val = '\0'; /* terminates name */
|
|
++val; /* now points to begin of value */
|
|
CHKmalloc(*name = strdup(param));
|
|
CHKmalloc(*paramval = strdup(val));
|
|
finalize_it:
|
|
RETiRet;
|
|
}
|
|
|
|
BEGINnewActInst
|
|
struct cnfparamvals *pvals;
|
|
int i;
|
|
int iNumTpls;
|
|
CODESTARTnewActInst
|
|
if((pvals = nvlstGetParams(lst, &actpblk, NULL)) == NULL) {
|
|
ABORT_FINALIZE(RS_RET_MISSING_CNFPARAMS);
|
|
}
|
|
|
|
CHKiRet(createInstance(&pData));
|
|
setInstParamDefaults(pData);
|
|
|
|
for(i = 0 ; i < actpblk.nParams ; ++i) {
|
|
if(!pvals[i].bUsed)
|
|
continue;
|
|
if(!strcmp(actpblk.descr[i].name, "topic")) {
|
|
pData->topic = (uchar*)es_str2cstr(pvals[i].val.d.estr, NULL);
|
|
} else if(!strcmp(actpblk.descr[i].name, "dynatopic")) {
|
|
pData->dynaTopic = pvals[i].val.d.n;
|
|
} else if(!strcmp(actpblk.descr[i].name, "dynatopic.cachesize")) {
|
|
pData->iDynaTopicCacheSize = pvals[i].val.d.n;
|
|
} else if(!strcmp(actpblk.descr[i].name, "closeTimeout")) {
|
|
pData->closeTimeout = pvals[i].val.d.n;
|
|
} else if(!strcmp(actpblk.descr[i].name, "partitions.auto")) {
|
|
pData->autoPartition = pvals[i].val.d.n;
|
|
} else if(!strcmp(actpblk.descr[i].name, "partitions.number")) {
|
|
pData->nPartitions = pvals[i].val.d.n;
|
|
} else if(!strcmp(actpblk.descr[i].name, "partitions.usefixed")) {
|
|
pData->fixedPartition = pvals[i].val.d.n;
|
|
} else if(!strcmp(actpblk.descr[i].name, "broker")) {
|
|
es_str_t *es = es_newStr(128);
|
|
int bNeedComma = 0;
|
|
for(int j = 0 ; j < pvals[i].val.d.ar->nmemb ; ++j) {
|
|
if(bNeedComma)
|
|
es_addChar(&es, ',');
|
|
es_addStr(&es, pvals[i].val.d.ar->arr[j]);
|
|
bNeedComma = 1;
|
|
}
|
|
pData->brokers = es_str2cstr(es, NULL);
|
|
es_deleteStr(es);
|
|
} else if(!strcmp(actpblk.descr[i].name, "confparam")) {
|
|
pData->nConfParams = pvals[i].val.d.ar->nmemb;
|
|
CHKmalloc(pData->confParams = malloc(sizeof(struct kafka_params) *
|
|
pvals[i].val.d.ar->nmemb ));
|
|
for(int j = 0 ; j < pvals[i].val.d.ar->nmemb ; ++j) {
|
|
char *cstr = es_str2cstr(pvals[i].val.d.ar->arr[j], NULL);
|
|
CHKiRet(processKafkaParam(cstr,
|
|
&pData->confParams[j].name,
|
|
&pData->confParams[j].val));
|
|
free(cstr);
|
|
}
|
|
} else if(!strcmp(actpblk.descr[i].name, "topicconfparam")) {
|
|
pData->nTopicConfParams = pvals[i].val.d.ar->nmemb;
|
|
CHKmalloc(pData->topicConfParams = malloc(sizeof(struct kafka_params) *
|
|
pvals[i].val.d.ar->nmemb ));
|
|
for(int j = 0 ; j < pvals[i].val.d.ar->nmemb ; ++j) {
|
|
char *cstr = es_str2cstr(pvals[i].val.d.ar->arr[j], NULL);
|
|
CHKiRet(processKafkaParam(cstr,
|
|
&pData->topicConfParams[j].name,
|
|
&pData->topicConfParams[j].val));
|
|
free(cstr);
|
|
}
|
|
} else if(!strcmp(actpblk.descr[i].name, "errorfile")) {
|
|
pData->errorFile = (uchar*)es_str2cstr(pvals[i].val.d.estr, NULL);
|
|
} else if(!strcmp(actpblk.descr[i].name, "key")) {
|
|
pData->key = (uchar*)es_str2cstr(pvals[i].val.d.estr, NULL);
|
|
} else if(!strcmp(actpblk.descr[i].name, "template")) {
|
|
pData->tplName = (uchar*)es_str2cstr(pvals[i].val.d.estr, NULL);
|
|
} else {
|
|
dbgprintf("omkafka: program error, non-handled param '%s'\n", actpblk.descr[i].name);
|
|
}
|
|
}
|
|
|
|
if(pData->brokers == NULL)
|
|
CHKmalloc(pData->brokers = strdup("localhost:9092"));
|
|
|
|
if(pData->dynaTopic && pData->topic == NULL) {
|
|
errmsg.LogError(0, RS_RET_CONFIG_ERROR,
|
|
"omkafka: requested dynamic topic, but no "
|
|
"name for topic template given - action definition invalid");
|
|
ABORT_FINALIZE(RS_RET_CONFIG_ERROR);
|
|
}
|
|
|
|
iNumTpls = 1;
|
|
if(pData->dynaTopic) ++iNumTpls;
|
|
CODE_STD_STRING_REQUESTnewActInst(iNumTpls);
|
|
CHKiRet(OMSRsetEntry(*ppOMSR, 0, (uchar*)strdup((pData->tplName == NULL) ?
|
|
"RSYSLOG_FileFormat" : (char*)pData->tplName),
|
|
OMSR_NO_RQD_TPL_OPTS));
|
|
if(pData->dynaTopic) {
|
|
CHKiRet(OMSRsetEntry(*ppOMSR, 1, ustrdup(pData->topic),
|
|
OMSR_NO_RQD_TPL_OPTS));
|
|
CHKmalloc(pData->dynCache = (dynaTopicCacheEntry**)
|
|
calloc(pData->iDynaTopicCacheSize, sizeof(dynaTopicCacheEntry*)));
|
|
pData->iCurrElt = -1;
|
|
}
|
|
pthread_mutex_lock(&closeTimeoutMut);
|
|
if (closeTimeout < pData->closeTimeout) {
|
|
closeTimeout = pData->closeTimeout;
|
|
}
|
|
pthread_mutex_unlock(&closeTimeoutMut);
|
|
|
|
CODE_STD_FINALIZERnewActInst
|
|
cnfparamvalsDestruct(pvals, &actpblk);
|
|
ENDnewActInst
|
|
|
|
BEGINparseSelectorAct
|
|
CODESTARTparseSelectorAct
|
|
CODE_STD_STRING_REQUESTparseSelectorAct(1)
|
|
/* first check if this config line is actually for us */
|
|
if(strncmp((char*) p, ":omkafka:", sizeof(":omkafka:") - 1)) {
|
|
ABORT_FINALIZE(RS_RET_CONFLINE_UNPROCESSED);
|
|
}
|
|
errmsg.LogError(0, RS_RET_LEGA_ACT_NOT_SUPPORTED,
|
|
"omkafka supports only RainerScript config format, use: "
|
|
"action(type=\"omkafka\" ...parameters...)");
|
|
ABORT_FINALIZE(RS_RET_LEGA_ACT_NOT_SUPPORTED);
|
|
CODE_STD_FINALIZERparseSelectorAct
|
|
ENDparseSelectorAct
|
|
|
|
|
|
BEGINmodExit
|
|
CODESTARTmodExit
|
|
statsobj.Destruct(&kafkaStats);
|
|
CHKiRet(objRelease(errmsg, CORE_COMPONENT));
|
|
CHKiRet(objRelease(statsobj, CORE_COMPONENT));
|
|
DESTROY_ATOMIC_HELPER_MUT(mutClock);
|
|
|
|
pthread_mutex_lock(&closeTimeoutMut);
|
|
int timeout = closeTimeout;
|
|
pthread_mutex_unlock(&closeTimeoutMut);
|
|
pthread_mutex_destroy(&closeTimeoutMut);
|
|
if (rd_kafka_wait_destroyed(timeout) != 0) {
|
|
DBGPRINTF("omkafka: couldn't close all resources gracefully. %d threads still remain.\n", rd_kafka_thread_cnt());
|
|
}
|
|
finalize_it:
|
|
ENDmodExit
|
|
|
|
|
|
BEGINqueryEtryPt
|
|
CODESTARTqueryEtryPt
|
|
CODEqueryEtryPt_STD_OMOD_QUERIES
|
|
CODEqueryEtryPt_STD_OMOD8_QUERIES
|
|
CODEqueryEtryPt_STD_CONF2_CNFNAME_QUERIES
|
|
CODEqueryEtryPt_STD_CONF2_OMOD_QUERIES
|
|
CODEqueryEtryPt_doHUP
|
|
ENDqueryEtryPt
|
|
|
|
|
|
BEGINmodInit()
|
|
CODESTARTmodInit
|
|
INITLegCnfVars
|
|
*ipIFVersProvided = CURR_MOD_IF_VERSION;
|
|
CODEmodInit_QueryRegCFSLineHdlr
|
|
CHKiRet(objUse(errmsg, CORE_COMPONENT));
|
|
CHKiRet(objUse(statsobj, CORE_COMPONENT));
|
|
|
|
INIT_ATOMIC_HELPER_MUT(mutClock);
|
|
|
|
DBGPRINTF("omkafka %s using librdkafka version %s\n",
|
|
VERSION, rd_kafka_version_str());
|
|
CHKiRet(statsobj.Construct(&kafkaStats));
|
|
CHKiRet(statsobj.SetName(kafkaStats, (uchar *)"omkafka"));
|
|
STATSCOUNTER_INIT(ctrTopicSubmit, mutCtrTopicSubmit);
|
|
CHKiRet(statsobj.AddCounter(kafkaStats, (uchar *)"submitted",
|
|
ctrType_IntCtr, CTR_FLAG_RESETTABLE, &ctrTopicSubmit));
|
|
ctrQueueSize = 0;
|
|
CHKiRet(statsobj.AddCounter(kafkaStats, (uchar *)"maxoutqsize",
|
|
ctrType_Int, CTR_FLAG_NONE, &ctrQueueSize));
|
|
STATSCOUNTER_INIT(ctrKafkaFail, mutCtrKafkaFail);
|
|
CHKiRet(statsobj.AddCounter(kafkaStats, (uchar *)"failures",
|
|
ctrType_IntCtr, CTR_FLAG_RESETTABLE, &ctrKafkaFail));
|
|
STATSCOUNTER_INIT(ctrCacheSkip, mutCtrCacheSkip);
|
|
CHKiRet(statsobj.AddCounter(kafkaStats, (uchar *)"topicdynacache.skipped",
|
|
ctrType_IntCtr, CTR_FLAG_RESETTABLE, &ctrCacheSkip));
|
|
STATSCOUNTER_INIT(ctrCacheMiss, mutCtrCacheMiss);
|
|
CHKiRet(statsobj.AddCounter(kafkaStats, (uchar *)"topicdynacache.miss",
|
|
ctrType_IntCtr, CTR_FLAG_RESETTABLE, &ctrCacheMiss));
|
|
STATSCOUNTER_INIT(ctrCacheEvict, mutCtrCacheEvict);
|
|
CHKiRet(statsobj.AddCounter(kafkaStats, (uchar *)"topicdynacache.evicted",
|
|
ctrType_IntCtr, CTR_FLAG_RESETTABLE, &ctrCacheEvict));
|
|
CHKiRet(statsobj.ConstructFinalize(kafkaStats));
|
|
CODEmodInit_QueryRegCFSLineHdlr
|
|
ENDmodInit
|