| 1 | /* |
| 2 | * librdkafka - The Apache Kafka C/C++ library |
| 3 | * |
| 4 | * Copyright (c) 2016 Magnus Edenhill |
| 5 | * All rights reserved. |
| 6 | * |
| 7 | * Redistribution and use in source and binary forms, with or without |
| 8 | * modification, are permitted provided that the following conditions are met: |
| 9 | * |
| 10 | * 1. Redistributions of source code must retain the above copyright notice, |
| 11 | * this list of conditions and the following disclaimer. |
| 12 | * 2. Redistributions in binary form must reproduce the above copyright notice, |
| 13 | * this list of conditions and the following disclaimer in the documentation |
| 14 | * and/or other materials provided with the distribution. |
| 15 | * |
| 16 | * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" |
| 17 | * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| 18 | * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE |
| 19 | * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE |
| 20 | * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| 21 | * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| 22 | * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
| 23 | * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
| 24 | * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
| 25 | * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
| 26 | * POSSIBILITY OF SUCH DAMAGE. |
| 27 | */ |
| 28 | |
| 29 | #include "rdkafka_int.h" |
| 30 | #include "rdkafka_offset.h" |
| 31 | #include "rdkafka_topic.h" |
| 32 | #include "rdkafka_interceptor.h" |
| 33 | |
| 34 | int RD_TLS rd_kafka_yield_thread = 0; |
| 35 | |
| 36 | void rd_kafka_yield (rd_kafka_t *rk) { |
| 37 | rd_kafka_yield_thread = 1; |
| 38 | } |
| 39 | |
| 40 | |
| 41 | /** |
| 42 | * @brief Check and reset yield flag. |
| 43 | * @returns rd_true if caller should yield, otherwise rd_false. |
| 44 | * @remarks rkq_lock MUST be held |
| 45 | */ |
| 46 | static RD_INLINE rd_bool_t rd_kafka_q_check_yield (rd_kafka_q_t *rkq) { |
| 47 | if (!(rkq->rkq_flags & RD_KAFKA_Q_F_YIELD)) |
| 48 | return rd_false; |
| 49 | |
| 50 | rkq->rkq_flags &= ~RD_KAFKA_Q_F_YIELD; |
| 51 | return rd_true; |
| 52 | } |
| 53 | /** |
| 54 | * Destroy a queue. refcnt must be at zero. |
| 55 | */ |
| 56 | void rd_kafka_q_destroy_final (rd_kafka_q_t *rkq) { |
| 57 | |
| 58 | mtx_lock(&rkq->rkq_lock); |
| 59 | if (unlikely(rkq->rkq_qio != NULL)) { |
| 60 | rd_free(rkq->rkq_qio); |
| 61 | rkq->rkq_qio = NULL; |
| 62 | } |
| 63 | /* Queue must have been disabled prior to final destruction, |
| 64 | * this is to catch the case where the queue owner/poll does not |
| 65 | * use rd_kafka_q_destroy_owner(). */ |
| 66 | rd_dassert(!(rkq->rkq_flags & RD_KAFKA_Q_F_READY)); |
| 67 | rd_kafka_q_disable0(rkq, 0/*no-lock*/); /* for the non-devel case */ |
| 68 | rd_kafka_q_fwd_set0(rkq, NULL, 0/*no-lock*/, 0 /*no-fwd-app*/); |
| 69 | rd_kafka_q_purge0(rkq, 0/*no-lock*/); |
| 70 | assert(!rkq->rkq_fwdq); |
| 71 | mtx_unlock(&rkq->rkq_lock); |
| 72 | mtx_destroy(&rkq->rkq_lock); |
| 73 | cnd_destroy(&rkq->rkq_cond); |
| 74 | |
| 75 | if (rkq->rkq_flags & RD_KAFKA_Q_F_ALLOCATED) |
| 76 | rd_free(rkq); |
| 77 | } |
| 78 | |
| 79 | |
| 80 | |
| 81 | /** |
| 82 | * Initialize a queue. |
| 83 | */ |
| 84 | void rd_kafka_q_init0 (rd_kafka_q_t *rkq, rd_kafka_t *rk, |
| 85 | const char *func, int line) { |
| 86 | rd_kafka_q_reset(rkq); |
| 87 | rkq->rkq_fwdq = NULL; |
| 88 | rkq->rkq_refcnt = 1; |
| 89 | rkq->rkq_flags = RD_KAFKA_Q_F_READY; |
| 90 | rkq->rkq_rk = rk; |
| 91 | rkq->rkq_qio = NULL; |
| 92 | rkq->rkq_serve = NULL; |
| 93 | rkq->rkq_opaque = NULL; |
| 94 | mtx_init(&rkq->rkq_lock, mtx_plain); |
| 95 | cnd_init(&rkq->rkq_cond); |
| 96 | #if ENABLE_DEVEL |
| 97 | rd_snprintf(rkq->rkq_name, sizeof(rkq->rkq_name), "%s:%d" , func, line); |
| 98 | #else |
| 99 | rkq->rkq_name = func; |
| 100 | #endif |
| 101 | } |
| 102 | |
| 103 | |
| 104 | /** |
| 105 | * Allocate a new queue and initialize it. |
| 106 | */ |
| 107 | rd_kafka_q_t *rd_kafka_q_new0 (rd_kafka_t *rk, const char *func, int line) { |
| 108 | rd_kafka_q_t *rkq = rd_malloc(sizeof(*rkq)); |
| 109 | rd_kafka_q_init(rkq, rk); |
| 110 | rkq->rkq_flags |= RD_KAFKA_Q_F_ALLOCATED; |
| 111 | #if ENABLE_DEVEL |
| 112 | rd_snprintf(rkq->rkq_name, sizeof(rkq->rkq_name), "%s:%d" , func, line); |
| 113 | #else |
| 114 | rkq->rkq_name = func; |
| 115 | #endif |
| 116 | return rkq; |
| 117 | } |
| 118 | |
| 119 | /** |
| 120 | * Set/clear forward queue. |
| 121 | * Queue forwarding enables message routing inside rdkafka. |
| 122 | * Typical use is to re-route all fetched messages for all partitions |
| 123 | * to one single queue. |
| 124 | * |
| 125 | * All access to rkq_fwdq are protected by rkq_lock. |
| 126 | */ |
| 127 | void rd_kafka_q_fwd_set0 (rd_kafka_q_t *srcq, rd_kafka_q_t *destq, |
| 128 | int do_lock, int fwd_app) { |
| 129 | |
| 130 | if (do_lock) |
| 131 | mtx_lock(&srcq->rkq_lock); |
| 132 | if (fwd_app) |
| 133 | srcq->rkq_flags |= RD_KAFKA_Q_F_FWD_APP; |
| 134 | if (srcq->rkq_fwdq) { |
| 135 | rd_kafka_q_destroy(srcq->rkq_fwdq); |
| 136 | srcq->rkq_fwdq = NULL; |
| 137 | } |
| 138 | if (destq) { |
| 139 | rd_kafka_q_keep(destq); |
| 140 | |
| 141 | /* If rkq has ops in queue, append them to fwdq's queue. |
| 142 | * This is an irreversible operation. */ |
| 143 | if (srcq->rkq_qlen > 0) { |
| 144 | rd_dassert(destq->rkq_flags & RD_KAFKA_Q_F_READY); |
| 145 | rd_kafka_q_concat(destq, srcq); |
| 146 | } |
| 147 | |
| 148 | srcq->rkq_fwdq = destq; |
| 149 | } |
| 150 | if (do_lock) |
| 151 | mtx_unlock(&srcq->rkq_lock); |
| 152 | } |
| 153 | |
| 154 | /** |
| 155 | * Purge all entries from a queue. |
| 156 | */ |
| 157 | int rd_kafka_q_purge0 (rd_kafka_q_t *rkq, int do_lock) { |
| 158 | rd_kafka_op_t *rko, *next; |
| 159 | TAILQ_HEAD(, rd_kafka_op_s) tmpq = TAILQ_HEAD_INITIALIZER(tmpq); |
| 160 | rd_kafka_q_t *fwdq; |
| 161 | int cnt = 0; |
| 162 | |
| 163 | if (do_lock) |
| 164 | mtx_lock(&rkq->rkq_lock); |
| 165 | |
| 166 | if ((fwdq = rd_kafka_q_fwd_get(rkq, 0))) { |
| 167 | if (do_lock) |
| 168 | mtx_unlock(&rkq->rkq_lock); |
| 169 | cnt = rd_kafka_q_purge(fwdq); |
| 170 | rd_kafka_q_destroy(fwdq); |
| 171 | return cnt; |
| 172 | } |
| 173 | |
| 174 | /* Move ops queue to tmpq to avoid lock-order issue |
| 175 | * by locks taken from rd_kafka_op_destroy(). */ |
| 176 | TAILQ_MOVE(&tmpq, &rkq->rkq_q, rko_link); |
| 177 | |
| 178 | /* Zero out queue */ |
| 179 | rd_kafka_q_reset(rkq); |
| 180 | |
| 181 | if (do_lock) |
| 182 | mtx_unlock(&rkq->rkq_lock); |
| 183 | |
| 184 | /* Destroy the ops */ |
| 185 | next = TAILQ_FIRST(&tmpq); |
| 186 | while ((rko = next)) { |
| 187 | next = TAILQ_NEXT(next, rko_link); |
| 188 | rd_kafka_op_destroy(rko); |
| 189 | cnt++; |
| 190 | } |
| 191 | |
| 192 | return cnt; |
| 193 | } |
| 194 | |
| 195 | |
| 196 | /** |
| 197 | * Purge all entries from a queue with a rktp version smaller than `version` |
| 198 | * This shaves off the head of the queue, up until the first rko with |
| 199 | * a non-matching rktp or version. |
| 200 | */ |
| 201 | void rd_kafka_q_purge_toppar_version (rd_kafka_q_t *rkq, |
| 202 | rd_kafka_toppar_t *rktp, int version) { |
| 203 | rd_kafka_op_t *rko, *next; |
| 204 | TAILQ_HEAD(, rd_kafka_op_s) tmpq = TAILQ_HEAD_INITIALIZER(tmpq); |
| 205 | int32_t cnt = 0; |
| 206 | int64_t size = 0; |
| 207 | rd_kafka_q_t *fwdq; |
| 208 | |
| 209 | mtx_lock(&rkq->rkq_lock); |
| 210 | |
| 211 | if ((fwdq = rd_kafka_q_fwd_get(rkq, 0))) { |
| 212 | mtx_unlock(&rkq->rkq_lock); |
| 213 | rd_kafka_q_purge_toppar_version(fwdq, rktp, version); |
| 214 | rd_kafka_q_destroy(fwdq); |
| 215 | return; |
| 216 | } |
| 217 | |
| 218 | /* Move ops to temporary queue and then destroy them from there |
| 219 | * without locks to avoid lock-ordering problems in op_destroy() */ |
| 220 | while ((rko = TAILQ_FIRST(&rkq->rkq_q)) && rko->rko_rktp && |
| 221 | rd_kafka_toppar_s2i(rko->rko_rktp) == rktp && |
| 222 | rko->rko_version < version) { |
| 223 | TAILQ_REMOVE(&rkq->rkq_q, rko, rko_link); |
| 224 | TAILQ_INSERT_TAIL(&tmpq, rko, rko_link); |
| 225 | cnt++; |
| 226 | size += rko->rko_len; |
| 227 | } |
| 228 | |
| 229 | |
| 230 | rkq->rkq_qlen -= cnt; |
| 231 | rkq->rkq_qsize -= size; |
| 232 | mtx_unlock(&rkq->rkq_lock); |
| 233 | |
| 234 | next = TAILQ_FIRST(&tmpq); |
| 235 | while ((rko = next)) { |
| 236 | next = TAILQ_NEXT(next, rko_link); |
| 237 | rd_kafka_op_destroy(rko); |
| 238 | } |
| 239 | } |
| 240 | |
| 241 | |
| 242 | /** |
| 243 | * Move 'cnt' entries from 'srcq' to 'dstq'. |
| 244 | * If 'cnt' == -1 all entries will be moved. |
| 245 | * Returns the number of entries moved. |
| 246 | */ |
| 247 | int rd_kafka_q_move_cnt (rd_kafka_q_t *dstq, rd_kafka_q_t *srcq, |
| 248 | int cnt, int do_locks) { |
| 249 | rd_kafka_op_t *rko; |
| 250 | int mcnt = 0; |
| 251 | |
| 252 | if (do_locks) { |
| 253 | mtx_lock(&srcq->rkq_lock); |
| 254 | mtx_lock(&dstq->rkq_lock); |
| 255 | } |
| 256 | |
| 257 | if (!dstq->rkq_fwdq && !srcq->rkq_fwdq) { |
| 258 | if (cnt > 0 && dstq->rkq_qlen == 0) |
| 259 | rd_kafka_q_io_event(dstq); |
| 260 | |
| 261 | /* Optimization, if 'cnt' is equal/larger than all |
| 262 | * items of 'srcq' we can move the entire queue. */ |
| 263 | if (cnt == -1 || |
| 264 | cnt >= (int)srcq->rkq_qlen) { |
| 265 | mcnt = srcq->rkq_qlen; |
| 266 | rd_kafka_q_concat0(dstq, srcq, 0/*no-lock*/); |
| 267 | } else { |
| 268 | while (mcnt < cnt && |
| 269 | (rko = TAILQ_FIRST(&srcq->rkq_q))) { |
| 270 | TAILQ_REMOVE(&srcq->rkq_q, rko, rko_link); |
| 271 | if (likely(!rko->rko_prio)) |
| 272 | TAILQ_INSERT_TAIL(&dstq->rkq_q, rko, |
| 273 | rko_link); |
| 274 | else |
| 275 | TAILQ_INSERT_SORTED( |
| 276 | &dstq->rkq_q, rko, |
| 277 | rd_kafka_op_t *, rko_link, |
| 278 | rd_kafka_op_cmp_prio); |
| 279 | |
| 280 | srcq->rkq_qlen--; |
| 281 | dstq->rkq_qlen++; |
| 282 | srcq->rkq_qsize -= rko->rko_len; |
| 283 | dstq->rkq_qsize += rko->rko_len; |
| 284 | mcnt++; |
| 285 | } |
| 286 | } |
| 287 | } else |
| 288 | mcnt = rd_kafka_q_move_cnt(dstq->rkq_fwdq ? dstq->rkq_fwdq:dstq, |
| 289 | srcq->rkq_fwdq ? srcq->rkq_fwdq:srcq, |
| 290 | cnt, do_locks); |
| 291 | |
| 292 | if (do_locks) { |
| 293 | mtx_unlock(&dstq->rkq_lock); |
| 294 | mtx_unlock(&srcq->rkq_lock); |
| 295 | } |
| 296 | |
| 297 | return mcnt; |
| 298 | } |
| 299 | |
| 300 | |
| 301 | /** |
| 302 | * Filters out outdated ops. |
| 303 | */ |
| 304 | static RD_INLINE rd_kafka_op_t *rd_kafka_op_filter (rd_kafka_q_t *rkq, |
| 305 | rd_kafka_op_t *rko, |
| 306 | int version) { |
| 307 | if (unlikely(!rko)) |
| 308 | return NULL; |
| 309 | |
| 310 | if (unlikely(rd_kafka_op_version_outdated(rko, version))) { |
| 311 | rd_kafka_q_deq0(rkq, rko); |
| 312 | rd_kafka_op_destroy(rko); |
| 313 | return NULL; |
| 314 | } |
| 315 | |
| 316 | return rko; |
| 317 | } |
| 318 | |
| 319 | |
| 320 | |
| 321 | /** |
| 322 | * Pop an op from a queue. |
| 323 | * |
| 324 | * Locality: any thread. |
| 325 | */ |
| 326 | |
| 327 | |
| 328 | /** |
| 329 | * Serve q like rd_kafka_q_serve() until an op is found that can be returned |
| 330 | * as an event to the application. |
| 331 | * |
| 332 | * @returns the first event:able op, or NULL on timeout. |
| 333 | * |
| 334 | * Locality: any thread |
| 335 | */ |
| 336 | rd_kafka_op_t *rd_kafka_q_pop_serve (rd_kafka_q_t *rkq, int timeout_ms, |
| 337 | int32_t version, |
| 338 | rd_kafka_q_cb_type_t cb_type, |
| 339 | rd_kafka_q_serve_cb_t *callback, |
| 340 | void *opaque) { |
| 341 | rd_kafka_op_t *rko; |
| 342 | rd_kafka_q_t *fwdq; |
| 343 | |
| 344 | rd_dassert(cb_type); |
| 345 | |
| 346 | mtx_lock(&rkq->rkq_lock); |
| 347 | |
| 348 | rd_kafka_yield_thread = 0; |
| 349 | if (!(fwdq = rd_kafka_q_fwd_get(rkq, 0))) { |
| 350 | struct timespec timeout_tspec; |
| 351 | |
| 352 | rd_timeout_init_timespec(&timeout_tspec, timeout_ms); |
| 353 | |
| 354 | while (1) { |
| 355 | rd_kafka_op_res_t res; |
| 356 | |
| 357 | /* Filter out outdated ops */ |
| 358 | retry: |
| 359 | while ((rko = TAILQ_FIRST(&rkq->rkq_q)) && |
| 360 | !(rko = rd_kafka_op_filter(rkq, rko, version))) |
| 361 | ; |
| 362 | |
| 363 | if (rko) { |
| 364 | /* Proper versioned op */ |
| 365 | rd_kafka_q_deq0(rkq, rko); |
| 366 | |
| 367 | /* Ops with callbacks are considered handled |
| 368 | * and we move on to the next op, if any. |
| 369 | * Ops w/o callbacks are returned immediately */ |
| 370 | res = rd_kafka_op_handle(rkq->rkq_rk, rkq, rko, |
| 371 | cb_type, opaque, |
| 372 | callback); |
| 373 | if (res == RD_KAFKA_OP_RES_HANDLED || |
| 374 | res == RD_KAFKA_OP_RES_KEEP) |
| 375 | goto retry; /* Next op */ |
| 376 | else if (unlikely(res == |
| 377 | RD_KAFKA_OP_RES_YIELD)) { |
| 378 | /* Callback yielded, unroll */ |
| 379 | mtx_unlock(&rkq->rkq_lock); |
| 380 | return NULL; |
| 381 | } else |
| 382 | break; /* Proper op, handle below. */ |
| 383 | } |
| 384 | |
| 385 | if (unlikely(rd_kafka_q_check_yield(rkq))) { |
| 386 | mtx_unlock(&rkq->rkq_lock); |
| 387 | return NULL; |
| 388 | } |
| 389 | |
| 390 | if (cnd_timedwait_abs(&rkq->rkq_cond, |
| 391 | &rkq->rkq_lock, |
| 392 | &timeout_tspec) != |
| 393 | thrd_success) { |
| 394 | mtx_unlock(&rkq->rkq_lock); |
| 395 | return NULL; |
| 396 | } |
| 397 | } |
| 398 | |
| 399 | mtx_unlock(&rkq->rkq_lock); |
| 400 | |
| 401 | } else { |
| 402 | /* Since the q_pop may block we need to release the parent |
| 403 | * queue's lock. */ |
| 404 | mtx_unlock(&rkq->rkq_lock); |
| 405 | rko = rd_kafka_q_pop_serve(fwdq, timeout_ms, version, |
| 406 | cb_type, callback, opaque); |
| 407 | rd_kafka_q_destroy(fwdq); |
| 408 | } |
| 409 | |
| 410 | |
| 411 | return rko; |
| 412 | } |
| 413 | |
| 414 | rd_kafka_op_t *rd_kafka_q_pop (rd_kafka_q_t *rkq, int timeout_ms, |
| 415 | int32_t version) { |
| 416 | return rd_kafka_q_pop_serve(rkq, timeout_ms, version, |
| 417 | RD_KAFKA_Q_CB_RETURN, |
| 418 | NULL, NULL); |
| 419 | } |
| 420 | |
| 421 | |
| 422 | /** |
| 423 | * Pop all available ops from a queue and call the provided |
| 424 | * callback for each op. |
| 425 | * `max_cnt` limits the number of ops served, 0 = no limit. |
| 426 | * |
| 427 | * Returns the number of ops served. |
| 428 | * |
| 429 | * Locality: any thread. |
| 430 | */ |
| 431 | int rd_kafka_q_serve (rd_kafka_q_t *rkq, int timeout_ms, |
| 432 | int max_cnt, rd_kafka_q_cb_type_t cb_type, |
| 433 | rd_kafka_q_serve_cb_t *callback, void *opaque) { |
| 434 | rd_kafka_t *rk = rkq->rkq_rk; |
| 435 | rd_kafka_op_t *rko; |
| 436 | rd_kafka_q_t localq; |
| 437 | rd_kafka_q_t *fwdq; |
| 438 | int cnt = 0; |
| 439 | struct timespec timeout_tspec; |
| 440 | |
| 441 | rd_dassert(cb_type); |
| 442 | |
| 443 | mtx_lock(&rkq->rkq_lock); |
| 444 | |
| 445 | rd_dassert(TAILQ_EMPTY(&rkq->rkq_q) || rkq->rkq_qlen > 0); |
| 446 | if ((fwdq = rd_kafka_q_fwd_get(rkq, 0))) { |
| 447 | int ret; |
| 448 | /* Since the q_pop may block we need to release the parent |
| 449 | * queue's lock. */ |
| 450 | mtx_unlock(&rkq->rkq_lock); |
| 451 | ret = rd_kafka_q_serve(fwdq, timeout_ms, max_cnt, |
| 452 | cb_type, callback, opaque); |
| 453 | rd_kafka_q_destroy(fwdq); |
| 454 | return ret; |
| 455 | } |
| 456 | |
| 457 | rd_timeout_init_timespec(&timeout_tspec, timeout_ms); |
| 458 | |
| 459 | /* Wait for op */ |
| 460 | while (!(rko = TAILQ_FIRST(&rkq->rkq_q)) && |
| 461 | !rd_kafka_q_check_yield(rkq) && |
| 462 | cnd_timedwait_abs(&rkq->rkq_cond, &rkq->rkq_lock, |
| 463 | &timeout_tspec) == thrd_success) |
| 464 | ; |
| 465 | |
| 466 | if (!rko) { |
| 467 | mtx_unlock(&rkq->rkq_lock); |
| 468 | return 0; |
| 469 | } |
| 470 | |
| 471 | /* Move the first `max_cnt` ops. */ |
| 472 | rd_kafka_q_init(&localq, rkq->rkq_rk); |
| 473 | rd_kafka_q_move_cnt(&localq, rkq, max_cnt == 0 ? -1/*all*/ : max_cnt, |
| 474 | 0/*no-locks*/); |
| 475 | |
| 476 | mtx_unlock(&rkq->rkq_lock); |
| 477 | |
| 478 | rd_kafka_yield_thread = 0; |
| 479 | |
| 480 | /* Call callback for each op */ |
| 481 | while ((rko = TAILQ_FIRST(&localq.rkq_q))) { |
| 482 | rd_kafka_op_res_t res; |
| 483 | |
| 484 | rd_kafka_q_deq0(&localq, rko); |
| 485 | res = rd_kafka_op_handle(rk, &localq, rko, cb_type, |
| 486 | opaque, callback); |
| 487 | /* op must have been handled */ |
| 488 | rd_kafka_assert(NULL, res != RD_KAFKA_OP_RES_PASS); |
| 489 | cnt++; |
| 490 | |
| 491 | if (unlikely(res == RD_KAFKA_OP_RES_YIELD || |
| 492 | rd_kafka_yield_thread)) { |
| 493 | /* Callback called rd_kafka_yield(), we must |
| 494 | * stop our callback dispatching and put the |
| 495 | * ops in localq back on the original queue head. */ |
| 496 | if (!TAILQ_EMPTY(&localq.rkq_q)) |
| 497 | rd_kafka_q_prepend(rkq, &localq); |
| 498 | break; |
| 499 | } |
| 500 | } |
| 501 | |
| 502 | rd_kafka_q_destroy_owner(&localq); |
| 503 | |
| 504 | return cnt; |
| 505 | } |
| 506 | |
| 507 | |
| 508 | |
| 509 | |
| 510 | |
| 511 | /** |
| 512 | * Populate 'rkmessages' array with messages from 'rkq'. |
| 513 | * If 'auto_commit' is set, each message's offset will be committed |
| 514 | * to the offset store for that toppar. |
| 515 | * |
| 516 | * Returns the number of messages added. |
| 517 | */ |
| 518 | |
| 519 | int rd_kafka_q_serve_rkmessages (rd_kafka_q_t *rkq, int timeout_ms, |
| 520 | rd_kafka_message_t **rkmessages, |
| 521 | size_t rkmessages_size) { |
| 522 | unsigned int cnt = 0; |
| 523 | TAILQ_HEAD(, rd_kafka_op_s) tmpq = TAILQ_HEAD_INITIALIZER(tmpq); |
| 524 | rd_kafka_op_t *rko, *next; |
| 525 | rd_kafka_t *rk = rkq->rkq_rk; |
| 526 | rd_kafka_q_t *fwdq; |
| 527 | struct timespec timeout_tspec; |
| 528 | |
| 529 | mtx_lock(&rkq->rkq_lock); |
| 530 | if ((fwdq = rd_kafka_q_fwd_get(rkq, 0))) { |
| 531 | /* Since the q_pop may block we need to release the parent |
| 532 | * queue's lock. */ |
| 533 | mtx_unlock(&rkq->rkq_lock); |
| 534 | cnt = rd_kafka_q_serve_rkmessages(fwdq, timeout_ms, |
| 535 | rkmessages, rkmessages_size); |
| 536 | rd_kafka_q_destroy(fwdq); |
| 537 | return cnt; |
| 538 | } |
| 539 | mtx_unlock(&rkq->rkq_lock); |
| 540 | |
| 541 | if (timeout_ms) |
| 542 | rd_kafka_app_poll_blocking(rk); |
| 543 | |
| 544 | rd_timeout_init_timespec(&timeout_tspec, timeout_ms); |
| 545 | |
| 546 | rd_kafka_yield_thread = 0; |
| 547 | while (cnt < rkmessages_size) { |
| 548 | rd_kafka_op_res_t res; |
| 549 | |
| 550 | mtx_lock(&rkq->rkq_lock); |
| 551 | |
| 552 | while (!(rko = TAILQ_FIRST(&rkq->rkq_q)) && |
| 553 | !rd_kafka_q_check_yield(rkq) && |
| 554 | cnd_timedwait_abs(&rkq->rkq_cond, &rkq->rkq_lock, |
| 555 | &timeout_tspec) == thrd_success) |
| 556 | ; |
| 557 | |
| 558 | if (!rko) { |
| 559 | mtx_unlock(&rkq->rkq_lock); |
| 560 | break; /* Timed out */ |
| 561 | } |
| 562 | |
| 563 | rd_kafka_q_deq0(rkq, rko); |
| 564 | |
| 565 | mtx_unlock(&rkq->rkq_lock); |
| 566 | |
| 567 | if (rd_kafka_op_version_outdated(rko, 0)) { |
| 568 | /* Outdated op, put on discard queue */ |
| 569 | TAILQ_INSERT_TAIL(&tmpq, rko, rko_link); |
| 570 | continue; |
| 571 | } |
| 572 | |
| 573 | /* Serve non-FETCH callbacks */ |
| 574 | res = rd_kafka_poll_cb(rk, rkq, rko, |
| 575 | RD_KAFKA_Q_CB_RETURN, NULL); |
| 576 | if (res == RD_KAFKA_OP_RES_KEEP || |
| 577 | res == RD_KAFKA_OP_RES_HANDLED) { |
| 578 | /* Callback served, rko is destroyed (if HANDLED). */ |
| 579 | continue; |
| 580 | } else if (unlikely(res == RD_KAFKA_OP_RES_YIELD || |
| 581 | rd_kafka_yield_thread)) { |
| 582 | /* Yield. */ |
| 583 | break; |
| 584 | } |
| 585 | rd_dassert(res == RD_KAFKA_OP_RES_PASS); |
| 586 | |
| 587 | /* Auto-commit offset, if enabled. */ |
| 588 | if (!rko->rko_err && rko->rko_type == RD_KAFKA_OP_FETCH) { |
| 589 | rd_kafka_toppar_t *rktp; |
| 590 | rktp = rd_kafka_toppar_s2i(rko->rko_rktp); |
| 591 | rd_kafka_toppar_lock(rktp); |
| 592 | rktp->rktp_app_offset = rko->rko_u.fetch.rkm.rkm_offset+1; |
| 593 | if (rktp->rktp_cgrp && |
| 594 | rk->rk_conf.enable_auto_offset_store) |
| 595 | rd_kafka_offset_store0(rktp, |
| 596 | rktp->rktp_app_offset, |
| 597 | 0/* no lock */); |
| 598 | rd_kafka_toppar_unlock(rktp); |
| 599 | } |
| 600 | |
| 601 | /* Get rkmessage from rko and append to array. */ |
| 602 | rkmessages[cnt++] = rd_kafka_message_get(rko); |
| 603 | } |
| 604 | |
| 605 | /* Discard non-desired and already handled ops */ |
| 606 | next = TAILQ_FIRST(&tmpq); |
| 607 | while (next) { |
| 608 | rko = next; |
| 609 | next = TAILQ_NEXT(next, rko_link); |
| 610 | rd_kafka_op_destroy(rko); |
| 611 | } |
| 612 | |
| 613 | rd_kafka_app_polled(rk); |
| 614 | |
| 615 | return cnt; |
| 616 | } |
| 617 | |
| 618 | |
| 619 | |
| 620 | void rd_kafka_queue_destroy (rd_kafka_queue_t *rkqu) { |
| 621 | if (rkqu->rkqu_is_owner) |
| 622 | rd_kafka_q_destroy_owner(rkqu->rkqu_q); |
| 623 | else |
| 624 | rd_kafka_q_destroy(rkqu->rkqu_q); |
| 625 | rd_free(rkqu); |
| 626 | } |
| 627 | |
| 628 | rd_kafka_queue_t *rd_kafka_queue_new0 (rd_kafka_t *rk, rd_kafka_q_t *rkq) { |
| 629 | rd_kafka_queue_t *rkqu; |
| 630 | |
| 631 | rkqu = rd_calloc(1, sizeof(*rkqu)); |
| 632 | |
| 633 | rkqu->rkqu_q = rkq; |
| 634 | rd_kafka_q_keep(rkq); |
| 635 | |
| 636 | rkqu->rkqu_rk = rk; |
| 637 | |
| 638 | return rkqu; |
| 639 | } |
| 640 | |
| 641 | |
| 642 | rd_kafka_queue_t *rd_kafka_queue_new (rd_kafka_t *rk) { |
| 643 | rd_kafka_q_t *rkq; |
| 644 | rd_kafka_queue_t *rkqu; |
| 645 | |
| 646 | rkq = rd_kafka_q_new(rk); |
| 647 | rkqu = rd_kafka_queue_new0(rk, rkq); |
| 648 | rd_kafka_q_destroy(rkq); /* Loose refcount from q_new, one is held |
| 649 | * by queue_new0 */ |
| 650 | rkqu->rkqu_is_owner = 1; |
| 651 | return rkqu; |
| 652 | } |
| 653 | |
| 654 | |
| 655 | rd_kafka_queue_t *rd_kafka_queue_get_main (rd_kafka_t *rk) { |
| 656 | return rd_kafka_queue_new0(rk, rk->rk_rep); |
| 657 | } |
| 658 | |
| 659 | |
| 660 | rd_kafka_queue_t *rd_kafka_queue_get_consumer (rd_kafka_t *rk) { |
| 661 | if (!rk->rk_cgrp) |
| 662 | return NULL; |
| 663 | return rd_kafka_queue_new0(rk, rk->rk_cgrp->rkcg_q); |
| 664 | } |
| 665 | |
| 666 | rd_kafka_queue_t *rd_kafka_queue_get_partition (rd_kafka_t *rk, |
| 667 | const char *topic, |
| 668 | int32_t partition) { |
| 669 | shptr_rd_kafka_toppar_t *s_rktp; |
| 670 | rd_kafka_toppar_t *rktp; |
| 671 | rd_kafka_queue_t *result; |
| 672 | |
| 673 | if (rk->rk_type == RD_KAFKA_PRODUCER) |
| 674 | return NULL; |
| 675 | |
| 676 | s_rktp = rd_kafka_toppar_get2(rk, topic, |
| 677 | partition, |
| 678 | 0, /* no ua_on_miss */ |
| 679 | 1 /* create_on_miss */); |
| 680 | |
| 681 | if (!s_rktp) |
| 682 | return NULL; |
| 683 | |
| 684 | rktp = rd_kafka_toppar_s2i(s_rktp); |
| 685 | result = rd_kafka_queue_new0(rk, rktp->rktp_fetchq); |
| 686 | rd_kafka_toppar_destroy(s_rktp); |
| 687 | |
| 688 | return result; |
| 689 | } |
| 690 | |
| 691 | rd_kafka_queue_t *rd_kafka_queue_get_background (rd_kafka_t *rk) { |
| 692 | if (rk->rk_background.q) |
| 693 | return rd_kafka_queue_new0(rk, rk->rk_background.q); |
| 694 | else |
| 695 | return NULL; |
| 696 | } |
| 697 | |
| 698 | |
| 699 | rd_kafka_resp_err_t rd_kafka_set_log_queue (rd_kafka_t *rk, |
| 700 | rd_kafka_queue_t *rkqu) { |
| 701 | rd_kafka_q_t *rkq; |
| 702 | if (!rkqu) |
| 703 | rkq = rk->rk_rep; |
| 704 | else |
| 705 | rkq = rkqu->rkqu_q; |
| 706 | rd_kafka_q_fwd_set(rk->rk_logq, rkq); |
| 707 | return RD_KAFKA_RESP_ERR_NO_ERROR; |
| 708 | } |
| 709 | |
| 710 | void rd_kafka_queue_forward (rd_kafka_queue_t *src, rd_kafka_queue_t *dst) { |
| 711 | rd_kafka_q_fwd_set0(src->rkqu_q, dst ? dst->rkqu_q : NULL, |
| 712 | 1, /* do_lock */ |
| 713 | 1 /* fwd_app */); |
| 714 | } |
| 715 | |
| 716 | |
| 717 | size_t rd_kafka_queue_length (rd_kafka_queue_t *rkqu) { |
| 718 | return (size_t)rd_kafka_q_len(rkqu->rkqu_q); |
| 719 | } |
| 720 | |
| 721 | /** |
| 722 | * @brief Enable or disable(fd==-1) fd-based wake-ups for queue |
| 723 | */ |
| 724 | void rd_kafka_q_io_event_enable (rd_kafka_q_t *rkq, int fd, |
| 725 | const void *payload, size_t size) { |
| 726 | struct rd_kafka_q_io *qio = NULL; |
| 727 | |
| 728 | if (fd != -1) { |
| 729 | qio = rd_malloc(sizeof(*qio) + size); |
| 730 | qio->fd = fd; |
| 731 | qio->size = size; |
| 732 | qio->payload = (void *)(qio+1); |
| 733 | qio->event_cb = NULL; |
| 734 | qio->event_cb_opaque = NULL; |
| 735 | memcpy(qio->payload, payload, size); |
| 736 | } |
| 737 | |
| 738 | mtx_lock(&rkq->rkq_lock); |
| 739 | if (rkq->rkq_qio) { |
| 740 | rd_free(rkq->rkq_qio); |
| 741 | rkq->rkq_qio = NULL; |
| 742 | } |
| 743 | |
| 744 | if (fd != -1) { |
| 745 | rkq->rkq_qio = qio; |
| 746 | } |
| 747 | |
| 748 | mtx_unlock(&rkq->rkq_lock); |
| 749 | |
| 750 | } |
| 751 | |
| 752 | void rd_kafka_queue_io_event_enable (rd_kafka_queue_t *rkqu, int fd, |
| 753 | const void *payload, size_t size) { |
| 754 | rd_kafka_q_io_event_enable(rkqu->rkqu_q, fd, payload, size); |
| 755 | } |
| 756 | |
| 757 | |
| 758 | /** |
| 759 | * @brief Enable or disable(event_cb==NULL) callback-based wake-ups for queue |
| 760 | */ |
| 761 | void rd_kafka_q_cb_event_enable (rd_kafka_q_t *rkq, |
| 762 | void (*event_cb) (rd_kafka_t *rk, |
| 763 | void *opaque), |
| 764 | void *opaque) { |
| 765 | struct rd_kafka_q_io *qio = NULL; |
| 766 | |
| 767 | if (event_cb) { |
| 768 | qio = rd_malloc(sizeof(*qio)); |
| 769 | qio->fd = -1; |
| 770 | qio->size = 0; |
| 771 | qio->payload = NULL; |
| 772 | qio->event_cb = event_cb; |
| 773 | qio->event_cb_opaque = opaque; |
| 774 | } |
| 775 | |
| 776 | mtx_lock(&rkq->rkq_lock); |
| 777 | if (rkq->rkq_qio) { |
| 778 | rd_free(rkq->rkq_qio); |
| 779 | rkq->rkq_qio = NULL; |
| 780 | } |
| 781 | |
| 782 | if (event_cb) { |
| 783 | rkq->rkq_qio = qio; |
| 784 | } |
| 785 | |
| 786 | mtx_unlock(&rkq->rkq_lock); |
| 787 | |
| 788 | } |
| 789 | |
| 790 | void rd_kafka_queue_cb_event_enable (rd_kafka_queue_t *rkqu, |
| 791 | void (*event_cb) (rd_kafka_t *rk, |
| 792 | void *opaque), |
| 793 | void *opaque) { |
| 794 | rd_kafka_q_cb_event_enable (rkqu->rkqu_q, event_cb, opaque); |
| 795 | } |
| 796 | |
| 797 | |
| 798 | /** |
| 799 | * Helper: wait for single op on 'rkq', and return its error, |
| 800 | * or .._TIMED_OUT on timeout. |
| 801 | */ |
| 802 | rd_kafka_resp_err_t rd_kafka_q_wait_result (rd_kafka_q_t *rkq, int timeout_ms) { |
| 803 | rd_kafka_op_t *rko; |
| 804 | rd_kafka_resp_err_t err; |
| 805 | |
| 806 | rko = rd_kafka_q_pop(rkq, timeout_ms, 0); |
| 807 | if (!rko) |
| 808 | err = RD_KAFKA_RESP_ERR__TIMED_OUT; |
| 809 | else { |
| 810 | err = rko->rko_err; |
| 811 | rd_kafka_op_destroy(rko); |
| 812 | } |
| 813 | |
| 814 | return err; |
| 815 | } |
| 816 | |
| 817 | |
| 818 | /** |
| 819 | * Apply \p callback on each op in queue. |
| 820 | * If the callback wishes to remove the rko it must do so using |
| 821 | * using rd_kafka_op_deq0(). |
| 822 | * |
| 823 | * @returns the sum of \p callback() return values. |
| 824 | * @remark rkq will be locked, callers should take care not to |
| 825 | * interact with \p rkq through other means from the callback to avoid |
| 826 | * deadlocks. |
| 827 | */ |
| 828 | int rd_kafka_q_apply (rd_kafka_q_t *rkq, |
| 829 | int (*callback) (rd_kafka_q_t *rkq, rd_kafka_op_t *rko, |
| 830 | void *opaque), |
| 831 | void *opaque) { |
| 832 | rd_kafka_op_t *rko, *next; |
| 833 | rd_kafka_q_t *fwdq; |
| 834 | int cnt = 0; |
| 835 | |
| 836 | mtx_lock(&rkq->rkq_lock); |
| 837 | if ((fwdq = rd_kafka_q_fwd_get(rkq, 0))) { |
| 838 | mtx_unlock(&rkq->rkq_lock); |
| 839 | cnt = rd_kafka_q_apply(fwdq, callback, opaque); |
| 840 | rd_kafka_q_destroy(fwdq); |
| 841 | return cnt; |
| 842 | } |
| 843 | |
| 844 | next = TAILQ_FIRST(&rkq->rkq_q); |
| 845 | while ((rko = next)) { |
| 846 | next = TAILQ_NEXT(next, rko_link); |
| 847 | cnt += callback(rkq, rko, opaque); |
| 848 | } |
| 849 | mtx_unlock(&rkq->rkq_lock); |
| 850 | |
| 851 | return cnt; |
| 852 | } |
| 853 | |
| 854 | /** |
| 855 | * @brief Convert relative to absolute offsets and also purge any messages |
| 856 | * that are older than \p min_offset. |
| 857 | * @remark Error ops with ERR__NOT_IMPLEMENTED will not be purged since |
| 858 | * they are used to indicate unknnown compression codecs and compressed |
| 859 | * messagesets may have a starting offset lower than what we requested. |
| 860 | * @remark \p rkq locking is not performed (caller's responsibility) |
| 861 | * @remark Must NOT be used on fwdq. |
| 862 | */ |
| 863 | void rd_kafka_q_fix_offsets (rd_kafka_q_t *rkq, int64_t min_offset, |
| 864 | int64_t base_offset) { |
| 865 | rd_kafka_op_t *rko, *next; |
| 866 | int adj_len = 0; |
| 867 | int64_t adj_size = 0; |
| 868 | |
| 869 | rd_kafka_assert(NULL, !rkq->rkq_fwdq); |
| 870 | |
| 871 | next = TAILQ_FIRST(&rkq->rkq_q); |
| 872 | while ((rko = next)) { |
| 873 | next = TAILQ_NEXT(next, rko_link); |
| 874 | |
| 875 | if (unlikely(rko->rko_type != RD_KAFKA_OP_FETCH)) |
| 876 | continue; |
| 877 | |
| 878 | rko->rko_u.fetch.rkm.rkm_offset += base_offset; |
| 879 | |
| 880 | if (rko->rko_u.fetch.rkm.rkm_offset < min_offset && |
| 881 | rko->rko_err != RD_KAFKA_RESP_ERR__NOT_IMPLEMENTED) { |
| 882 | adj_len++; |
| 883 | adj_size += rko->rko_len; |
| 884 | TAILQ_REMOVE(&rkq->rkq_q, rko, rko_link); |
| 885 | rd_kafka_op_destroy(rko); |
| 886 | continue; |
| 887 | } |
| 888 | } |
| 889 | |
| 890 | |
| 891 | rkq->rkq_qlen -= adj_len; |
| 892 | rkq->rkq_qsize -= adj_size; |
| 893 | } |
| 894 | |
| 895 | |
| 896 | /** |
| 897 | * @brief Print information and contents of queue |
| 898 | */ |
| 899 | void rd_kafka_q_dump (FILE *fp, rd_kafka_q_t *rkq) { |
| 900 | mtx_lock(&rkq->rkq_lock); |
| 901 | fprintf(fp, "Queue %p \"%s\" (refcnt %d, flags 0x%x, %d ops, " |
| 902 | "%" PRId64" bytes)\n" , |
| 903 | rkq, rkq->rkq_name, rkq->rkq_refcnt, rkq->rkq_flags, |
| 904 | rkq->rkq_qlen, rkq->rkq_qsize); |
| 905 | |
| 906 | if (rkq->rkq_qio) |
| 907 | fprintf(fp, " QIO fd %d\n" , rkq->rkq_qio->fd); |
| 908 | if (rkq->rkq_serve) |
| 909 | fprintf(fp, " Serve callback %p, opaque %p\n" , |
| 910 | rkq->rkq_serve, rkq->rkq_opaque); |
| 911 | |
| 912 | if (rkq->rkq_fwdq) { |
| 913 | fprintf(fp, " Forwarded ->\n" ); |
| 914 | rd_kafka_q_dump(fp, rkq->rkq_fwdq); |
| 915 | } else { |
| 916 | rd_kafka_op_t *rko; |
| 917 | |
| 918 | if (!TAILQ_EMPTY(&rkq->rkq_q)) |
| 919 | fprintf(fp, " Queued ops:\n" ); |
| 920 | TAILQ_FOREACH(rko, &rkq->rkq_q, rko_link) { |
| 921 | fprintf(fp, " %p %s (v%" PRId32", flags 0x%x, " |
| 922 | "prio %d, len %" PRId32", source %s, " |
| 923 | "replyq %p)\n" , |
| 924 | rko, rd_kafka_op2str(rko->rko_type), |
| 925 | rko->rko_version, rko->rko_flags, |
| 926 | rko->rko_prio, rko->rko_len, |
| 927 | #if ENABLE_DEVEL |
| 928 | rko->rko_source |
| 929 | #else |
| 930 | "-" |
| 931 | #endif |
| 932 | , |
| 933 | rko->rko_replyq.q |
| 934 | ); |
| 935 | } |
| 936 | } |
| 937 | |
| 938 | mtx_unlock(&rkq->rkq_lock); |
| 939 | } |
| 940 | |
| 941 | |
| 942 | void rd_kafka_enq_once_trigger_destroy (void *ptr) { |
| 943 | rd_kafka_enq_once_t *eonce = ptr; |
| 944 | |
| 945 | rd_kafka_enq_once_trigger(eonce, RD_KAFKA_RESP_ERR__DESTROY, "destroy" ); |
| 946 | } |
| 947 | |