1 | /* |
2 | * exchange.c |
3 | * |
4 | * Copyright (C) 2016 Aerospike, Inc. |
5 | * |
6 | * Portions may be licensed to Aerospike, Inc. under one or more contributor |
7 | * license agreements. |
8 | * |
9 | * This program is free software: you can redistribute it and/or modify it under |
10 | * the terms of the GNU Affero General Public License as published by the Free |
11 | * Software Foundation, either version 3 of the License, or (at your option) any |
12 | * later version. |
13 | * |
14 | * This program is distributed in the hope that it will be useful, but WITHOUT |
15 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS |
16 | * FOR A PARTICULAR PURPOSE. See the GNU Affero General Public License for more |
17 | * details. |
18 | * |
19 | * You should have received a copy of the GNU Affero General Public License |
20 | * along with this program. If not, see http://www.gnu.org/licenses/ |
21 | */ |
22 | |
23 | #include "fabric/exchange.h" |
24 | |
25 | #include <errno.h> |
26 | #include <pthread.h> |
27 | #include <unistd.h> |
28 | #include <sys/param.h> // For MAX() and MIN(). |
29 | |
30 | #include "citrusleaf/alloc.h" |
31 | #include "citrusleaf/cf_atomic.h" |
32 | #include "citrusleaf/cf_clock.h" |
33 | #include "citrusleaf/cf_queue.h" |
34 | |
35 | #include "cf_thread.h" |
36 | #include "dynbuf.h" |
37 | #include "fault.h" |
38 | #include "shash.h" |
39 | #include "socket.h" |
40 | |
41 | #include "base/cfg.h" |
42 | #include "base/datamodel.h" |
43 | #include "base/stats.h" |
44 | #include "fabric/fabric.h" |
45 | #include "fabric/hb.h" |
46 | #include "fabric/partition_balance.h" |
47 | #include "storage/storage.h" |
48 | |
49 | /* |
50 | * Overview |
51 | * ======== |
52 | * Cluster data exchange state machine. Exchanges per namespace partition |
53 | * version exchange for now, after evey cluster change. |
54 | * |
55 | * State transition diagram |
56 | * ======================== |
57 | * The exchange state transition diagram responds to three events |
58 | * 1. Incoming message |
59 | * 2. Timer event |
60 | * 3. Clustering module's cluster change event. |
61 | * |
62 | * There are four states |
63 | * 1. Rest - the exchange is complete with all exchanged data committed. |
64 | * 2. Exchanging - the cluster has changed since the last commit and new data |
65 | * exchange is in progress. |
66 | * 3. Ready to commit - this node has send its exchange data to all cluster |
67 | * members, received corresponding acks and also exchange data from all cluster |
68 | * members. |
69 | * 4. Orphaned - this node is an orphan. After a timeout blocks client |
70 | * transactions. |
71 | * |
72 | * Exchange starts by being in the orphaned state. |
73 | * |
74 | * Code organization |
75 | * ================= |
76 | * |
77 | * There are different sections for each state. Each state has a dispatcher |
78 | * which delegates the event handing to a state specific function. |
79 | * |
80 | * Locks |
81 | * ===== |
82 | * 1. g_exchanage_lock - protected the exchange state machine. |
83 | * 2. g_exchange_info_lock - prevents update of exchanged data for a round while |
84 | * exchange is in progress |
85 | * 3. g_exchange_commited_cluster_lock - a higher |
86 | * granularity, committed cluster r/w lock,used to allow read access to |
87 | * committed cluster data even while exchange is busy and holding on to the |
88 | * global exchange state machine lock. |
89 | * 4. g_external_event_publisher_lock - ensure order of external events |
90 | * published is maintained. |
91 | * |
92 | * Lock order |
93 | * ========== |
94 | * Locks to be obtained in the order mentioned above and relinquished in the |
95 | * reverse order. |
96 | */ |
97 | |
98 | /* |
99 | * ---------------------------------------------------------------------------- |
100 | * Constants |
101 | * ---------------------------------------------------------------------------- |
102 | */ |
103 | |
104 | /** |
105 | * Exchange protocol version information. |
106 | */ |
107 | #define AS_EXCHANGE_PROTOCOL_IDENTIFIER 1 |
108 | |
109 | /** |
110 | * A soft limit for the maximum cluster size. Meant to be optimize hash and list |
111 | * data structures and not as a limit on the number of nodes. |
112 | */ |
113 | #define AS_EXCHANGE_CLUSTER_MAX_SIZE_SOFT 200 |
114 | |
115 | /** |
116 | * A soft limit for the maximum number of unique vinfo's in a namespace. Meant |
117 | * to be optimize hash and list data structures and not as a limit on the number |
118 | * of vinfos processed. |
119 | */ |
120 | #define AS_EXCHANGE_UNIQUE_VINFO_MAX_SIZE_SOFT 200 |
121 | |
122 | /** |
123 | * Average number of partitions for a version information. Used as initial |
124 | * allocation size for every unique vinfo, hence a smaller value. |
125 | */ |
126 | #define AS_EXCHANGE_VINFO_NUM_PIDS_AVG 1024 |
127 | |
128 | /** |
129 | * Maximum event listeners. |
130 | */ |
131 | #define AS_EXTERNAL_EVENT_LISTENER_MAX 7 |
132 | |
133 | /* |
134 | * ---------------------------------------------------------------------------- |
135 | * Exchange data format for namespaces payload |
136 | * ---------------------------------------------------------------------------- |
137 | */ |
138 | |
139 | /** |
140 | * Partition data exchanged for each unique vinfo for a namespace. |
141 | */ |
142 | typedef struct as_exchange_vinfo_payload_s |
143 | { |
144 | /** |
145 | * The partition vinfo. |
146 | */ |
147 | as_partition_version vinfo; |
148 | |
149 | /** |
150 | * Count of partitions having this vinfo. |
151 | */ |
152 | uint32_t num_pids; |
153 | |
154 | /** |
155 | * Partition having this vinfo. |
156 | */ |
157 | uint16_t pids[]; |
158 | }__attribute__((__packed__)) as_exchange_vinfo_payload; |
159 | |
160 | /** |
161 | * Information exchanged for a single namespace. |
162 | */ |
163 | typedef struct as_exchange_ns_vinfos_payload_s |
164 | { |
165 | /** |
166 | * Count of version infos. |
167 | */ |
168 | uint32_t num_vinfos; |
169 | |
170 | /** |
171 | * Parition version information for each unique version. |
172 | */ |
173 | as_exchange_vinfo_payload vinfos[]; |
174 | }__attribute__((__packed__)) as_exchange_ns_vinfos_payload; |
175 | |
176 | /** |
177 | * Received data stored per node, per namespace, before actual commit. |
178 | */ |
179 | typedef struct as_exchange_node_namespace_data_s |
180 | { |
181 | /** |
182 | * Mapped local namespace. |
183 | */ |
184 | as_namespace* local_namespace; |
185 | |
186 | /** |
187 | * Partition versions for this namespace. This field is reused across |
188 | * exchange rounds and may not be null even if the local namespace is null. |
189 | */ |
190 | as_exchange_ns_vinfos_payload* partition_versions; |
191 | |
192 | /** |
193 | * Sender's rack id. |
194 | */ |
195 | uint32_t rack_id; |
196 | |
197 | /** |
198 | * Sender's roster generation. |
199 | */ |
200 | uint32_t roster_generation; |
201 | |
202 | /** |
203 | * Sender's roster count. |
204 | */ |
205 | uint32_t roster_count; |
206 | |
207 | /** |
208 | * Sending node's roster for this namespace. |
209 | */ |
210 | cf_node* roster; |
211 | |
212 | /** |
213 | * Sending node's roster rack-ids for this namespace. |
214 | */ |
215 | cf_node* roster_rack_ids; |
216 | |
217 | /** |
218 | * Sender's eventual regime for this namespace. |
219 | */ |
220 | uint32_t eventual_regime; |
221 | |
222 | /** |
223 | * Sender's rebalance regime for this namespace. |
224 | */ |
225 | uint32_t rebalance_regime; |
226 | |
227 | /** |
228 | * Sender's rebalance flags for this namespace. |
229 | */ |
230 | uint32_t rebalance_flags; |
231 | } as_exchange_node_namespace_data; |
232 | |
233 | /** |
234 | * Exchanged data for a single node. |
235 | */ |
236 | typedef struct as_exchange_node_data_s |
237 | { |
238 | /** |
239 | * Used by exchange listeners during upgrades for compatibility purposes. |
240 | */ |
241 | uint32_t compatibility_id; |
242 | |
243 | /** |
244 | * Number of sender's namespaces that have a matching local namespace. |
245 | */ |
246 | uint32_t num_namespaces; |
247 | |
248 | /** |
249 | * Data for sender's namespaces having a matching local namespace. |
250 | */ |
251 | as_exchange_node_namespace_data namespace_data[AS_NAMESPACE_SZ]; |
252 | } as_exchange_node_data; |
253 | |
254 | /* |
255 | * ---------------------------------------------------------------------------- |
256 | * Exchange internal data structures |
257 | * ---------------------------------------------------------------------------- |
258 | */ |
259 | |
260 | /** |
261 | * Exchange subsystem status. |
262 | */ |
263 | typedef enum |
264 | { |
265 | AS_EXCHANGE_SYS_STATE_UNINITIALIZED, |
266 | AS_EXCHANGE_SYS_STATE_RUNNING, |
267 | AS_EXCHANGE_SYS_STATE_SHUTTING_DOWN, |
268 | AS_EXCHANGE_SYS_STATE_STOPPED |
269 | } as_exchange_sys_state; |
270 | |
271 | /** |
272 | * Exchange message types. |
273 | */ |
274 | typedef enum |
275 | { |
276 | /** |
277 | * Exchange data for one node. |
278 | */ |
279 | AS_EXCHANGE_MSG_TYPE_DATA, |
280 | |
281 | /** |
282 | * Ack on receipt of exchanged data. |
283 | */ |
284 | AS_EXCHANGE_MSG_TYPE_DATA_ACK, |
285 | |
286 | /** |
287 | * Not used. |
288 | */ |
289 | AS_EXCHANGE_MSG_TYPE_DATA_NACK, |
290 | |
291 | /** |
292 | * The source is ready to commit exchanged information. |
293 | */ |
294 | AS_EXCHANGE_MSG_TYPE_READY_TO_COMMIT, |
295 | |
296 | /** |
297 | * Message from the principal asking all nodes to commit the exchanged |
298 | * information. |
299 | */ |
300 | AS_EXCHANGE_MSG_TYPE_COMMIT, |
301 | |
302 | /** |
303 | * Sentinel value for exchange message types. |
304 | */ |
305 | AS_EXCHANGE_MSG_TYPE_SENTINEL |
306 | } as_exchange_msg_type; |
307 | |
308 | /** |
309 | * Internal exchange event type. |
310 | */ |
311 | typedef enum |
312 | { |
313 | /** |
314 | * Cluster change event. |
315 | */ |
316 | AS_EXCHANGE_EVENT_CLUSTER_CHANGE, |
317 | |
318 | /** |
319 | * Timer event. |
320 | */ |
321 | AS_EXCHANGE_EVENT_TIMER, |
322 | |
323 | /** |
324 | * Incoming message event. |
325 | */ |
326 | AS_EXCHANGE_EVENT_MSG, |
327 | } as_exchange_event_type; |
328 | |
329 | /** |
330 | * Internal exchange event. |
331 | */ |
332 | typedef struct as_exchange_event_s |
333 | { |
334 | /** |
335 | * The type of the event. |
336 | */ |
337 | as_exchange_event_type type; |
338 | |
339 | /** |
340 | * Message for incoming message events. |
341 | */ |
342 | msg* msg; |
343 | |
344 | /** |
345 | * Source for incoming message events. |
346 | */ |
347 | cf_node msg_source; |
348 | |
349 | /** |
350 | * Clustering event instance for clustering events. |
351 | */ |
352 | as_clustering_event* clustering_event; |
353 | } as_exchange_event; |
354 | |
355 | /** |
356 | * Exchange subsystem state in the state transition diagram. |
357 | */ |
358 | typedef enum as_exchange_state_s |
359 | { |
360 | /** |
361 | * Exchange subsystem is at rest will all data exchanged synchronized and |
362 | * committed. |
363 | */ |
364 | AS_EXCHANGE_STATE_REST, |
365 | |
366 | /** |
367 | * Data exchange is in progress. |
368 | */ |
369 | AS_EXCHANGE_STATE_EXCHANGING, |
370 | |
371 | /** |
372 | * Data exchange is complete and this node is ready to commit data. |
373 | */ |
374 | AS_EXCHANGE_STATE_READY_TO_COMMIT, |
375 | |
376 | /** |
377 | * Self node is orphaned. |
378 | */ |
379 | AS_EXCHANGE_STATE_ORPHANED |
380 | } as_exchange_state; |
381 | |
382 | /** |
383 | * State for a single node in the succession list. |
384 | */ |
385 | typedef struct as_exchange_node_state_s |
386 | { |
387 | /** |
388 | * Inidicates if peer node has acknowledged send from self. |
389 | */ |
390 | bool send_acked; |
391 | |
392 | /** |
393 | * Inidicates if self node has received data from this peer. |
394 | */ |
395 | bool received; |
396 | |
397 | /** |
398 | * Inidicates if this peer node is ready to commit. Only relevant and used |
399 | * by the current principal. |
400 | */ |
401 | bool is_ready_to_commit; |
402 | |
403 | /** |
404 | * Exchange data received from this peer node. Member variables may be heap |
405 | * allocated and hence should be freed carefully while discarding this |
406 | * structure instance. |
407 | */ |
408 | as_exchange_node_data* data; |
409 | } as_exchange_node_state; |
410 | |
411 | /** |
412 | * State maintained by the exchange subsystem. |
413 | */ |
414 | typedef struct as_exchange_s |
415 | { |
416 | /** |
417 | * Exchange subsystem status. |
418 | */ |
419 | as_exchange_sys_state sys_state; |
420 | |
421 | /** |
422 | * Exchange state in the state transition diagram. |
423 | */ |
424 | as_exchange_state state; |
425 | |
426 | /** |
427 | * Time when this node's exchange data was sent out. |
428 | */ |
429 | cf_clock send_ts; |
430 | |
431 | /** |
432 | * Time when this node's ready to commit was sent out. |
433 | */ |
434 | cf_clock ready_to_commit_send_ts; |
435 | |
436 | /** |
437 | * Thread id of the timer event generator. |
438 | */ |
439 | pthread_t timer_tid; |
440 | |
441 | /** |
442 | * Nodes that are not yet ready to commit. |
443 | */ |
444 | cf_vector ready_to_commit_pending_nodes; |
445 | |
446 | /** |
447 | * Current cluster key. |
448 | */ |
449 | as_cluster_key cluster_key; |
450 | |
451 | /** |
452 | * Exchange's copy of the succession list. |
453 | */ |
454 | cf_vector succession_list; |
455 | |
456 | /** |
457 | * The principal node in current succession list. Always the first node. |
458 | */ |
459 | cf_node principal; |
460 | |
461 | /** |
462 | * Used by exchange listeners during upgrades for compatibility purposes. |
463 | */ |
464 | uint32_t compatibility_ids[AS_CLUSTER_SZ]; |
465 | |
466 | /** |
467 | * Used by exchange listeners during upgrades for compatibility purposes. |
468 | */ |
469 | uint32_t min_compatibility_id; |
470 | |
471 | /** |
472 | * Committed cluster generation. |
473 | */ |
474 | uint64_t committed_cluster_generation; |
475 | |
476 | /** |
477 | * Last committed cluster key. |
478 | */ |
479 | as_cluster_key committed_cluster_key; |
480 | |
481 | /** |
482 | * Last committed cluster size - size of the succession list. |
483 | */ |
484 | uint32_t committed_cluster_size; |
485 | |
486 | /** |
487 | * Last committed exchange's succession list. |
488 | */ |
489 | cf_vector committed_succession_list; |
490 | |
491 | /** |
492 | * The principal node in the committed succession list. Always the first |
493 | * node. |
494 | */ |
495 | cf_node committed_principal; |
496 | |
497 | /** |
498 | * The time this node entered orphan state. |
499 | */ |
500 | cf_clock orphan_state_start_time; |
501 | |
502 | /** |
503 | * Indicates if transactions have already been blocked in the orphan state. |
504 | */ |
505 | bool orphan_state_are_transactions_blocked; |
506 | |
507 | /** |
508 | * Will have an as_exchange_node_state entry for every node in the |
509 | * succession list. |
510 | */ |
511 | cf_shash* nodeid_to_node_state; |
512 | |
513 | /** |
514 | * Self node's partition version payload for current round. |
515 | */ |
516 | cf_dyn_buf self_data_dyn_buf[AS_NAMESPACE_SZ]; |
517 | |
518 | /** |
519 | * This node's exchange data fabric message to send for current round. |
520 | */ |
521 | msg* data_msg; |
522 | } as_exchange; |
523 | |
524 | /** |
525 | * Internal storage for external event listeners. |
526 | */ |
527 | typedef struct as_exchange_event_listener_s |
528 | { |
529 | /** |
530 | * The listener's calback function. |
531 | */ |
532 | as_exchange_cluster_changed_cb event_callback; |
533 | |
534 | /** |
535 | * The listeners user data object passed back as is to the callback |
536 | * function. |
537 | */ |
538 | void* udata; |
539 | } as_exchange_event_listener; |
540 | |
541 | /** |
542 | * External event publisher state. |
543 | */ |
544 | typedef struct as_exchange_external_event_publisher_s |
545 | { |
546 | /** |
547 | * State of the external event publisher. |
548 | */ |
549 | as_exchange_sys_state sys_state; |
550 | |
551 | /** |
552 | * Inidicates if there is an event to publish. |
553 | */ |
554 | bool event_queued; |
555 | |
556 | /** |
557 | * The pending event to publish. |
558 | */ |
559 | as_exchange_cluster_changed_event to_publish; |
560 | |
561 | /** |
562 | * The static succession list published with the message. |
563 | */ |
564 | cf_vector published_succession_list; |
565 | |
566 | /** |
567 | * Conditional variable to signal a pending event. |
568 | */ |
569 | pthread_cond_t is_pending; |
570 | |
571 | /** |
572 | * Thread id of the publisher thread. |
573 | */ |
574 | pthread_t event_publisher_tid; |
575 | |
576 | /** |
577 | * Mutex to protect the conditional variable. |
578 | */ |
579 | pthread_mutex_t is_pending_mutex; |
580 | |
581 | /** |
582 | * External event listeners. |
583 | */ |
584 | as_exchange_event_listener event_listeners[AS_EXTERNAL_EVENT_LISTENER_MAX]; |
585 | |
586 | /** |
587 | * Event listener count. |
588 | */ |
589 | uint32_t event_listener_count; |
590 | } as_exchange_external_event_publisher; |
591 | |
592 | /* |
593 | * ---------------------------------------------------------------------------- |
594 | * Externs |
595 | * ---------------------------------------------------------------------------- |
596 | */ |
597 | void |
598 | as_skew_monitor_update(); |
599 | |
600 | /* |
601 | * ---------------------------------------------------------------------------- |
602 | * Globals |
603 | * ---------------------------------------------------------------------------- |
604 | */ |
605 | |
606 | /** |
607 | * Singleton exchange state all initialized to zero. |
608 | */ |
609 | static as_exchange g_exchange = { 0 }; |
610 | |
611 | /** |
612 | * Rebalance flags. |
613 | */ |
614 | typedef enum |
615 | { |
616 | AS_EXCHANGE_REBALANCE_FLAG_UNIFORM = 0x01, |
617 | AS_EXCHANGE_REBALANCE_FLAG_QUIESCE = 0x02 |
618 | } as_exchange_rebalance_flags; |
619 | |
620 | /** |
621 | * The fields in the exchange message. Should never change the order or elements |
622 | * in between. |
623 | */ |
624 | typedef enum |
625 | { |
626 | AS_EXCHANGE_MSG_ID, |
627 | AS_EXCHANGE_MSG_TYPE, |
628 | AS_EXCHANGE_MSG_CLUSTER_KEY, |
629 | AS_EXCHANGE_MSG_NAMESPACES, |
630 | AS_EXCHANGE_MSG_NS_PARTITION_VERSIONS, |
631 | AS_EXCHANGE_MSG_NS_RACK_IDS, |
632 | AS_EXCHANGE_MSG_NS_ROSTER_GENERATIONS, |
633 | AS_EXCHANGE_MSG_NS_ROSTERS, |
634 | AS_EXCHANGE_MSG_NS_ROSTERS_RACK_IDS, |
635 | AS_EXCHANGE_MSG_NS_EVENTUAL_REGIMES, |
636 | AS_EXCHANGE_MSG_NS_REBALANCE_REGIMES, |
637 | AS_EXCHANGE_MSG_NS_REBALANCE_FLAGS, |
638 | AS_EXCHANGE_MSG_COMPATIBILITY_ID, |
639 | |
640 | NUM_EXCHANGE_MSG_FIELDS |
641 | } as_exchange_msg_fields; |
642 | |
643 | /** |
644 | * Exchange message template. |
645 | */ |
646 | static const msg_template exchange_msg_template[] = { |
647 | { AS_EXCHANGE_MSG_ID, M_FT_UINT32 }, |
648 | { AS_EXCHANGE_MSG_TYPE, M_FT_UINT32 }, |
649 | { AS_EXCHANGE_MSG_CLUSTER_KEY, M_FT_UINT64 }, |
650 | { AS_EXCHANGE_MSG_NAMESPACES, M_FT_MSGPACK }, |
651 | { AS_EXCHANGE_MSG_NS_PARTITION_VERSIONS, M_FT_MSGPACK }, |
652 | { AS_EXCHANGE_MSG_NS_RACK_IDS, M_FT_MSGPACK }, |
653 | { AS_EXCHANGE_MSG_NS_ROSTER_GENERATIONS, M_FT_MSGPACK }, |
654 | { AS_EXCHANGE_MSG_NS_ROSTERS, M_FT_MSGPACK }, |
655 | { AS_EXCHANGE_MSG_NS_ROSTERS_RACK_IDS, M_FT_MSGPACK }, |
656 | { AS_EXCHANGE_MSG_NS_EVENTUAL_REGIMES, M_FT_MSGPACK }, |
657 | { AS_EXCHANGE_MSG_NS_REBALANCE_REGIMES, M_FT_MSGPACK }, |
658 | { AS_EXCHANGE_MSG_NS_REBALANCE_FLAGS, M_FT_MSGPACK }, |
659 | { AS_EXCHANGE_MSG_COMPATIBILITY_ID, M_FT_UINT32 } |
660 | }; |
661 | |
662 | COMPILER_ASSERT(sizeof(exchange_msg_template) / sizeof(msg_template) == |
663 | NUM_EXCHANGE_MSG_FIELDS); |
664 | |
665 | /** |
666 | * Global lock to set or get exchanged info from other threads. |
667 | */ |
668 | pthread_mutex_t g_exchanged_info_lock = PTHREAD_MUTEX_INITIALIZER; |
669 | |
670 | /** |
671 | * Global lock to serialize all reads and writes to the exchange state. |
672 | */ |
673 | pthread_mutex_t g_exchange_lock = PTHREAD_RECURSIVE_MUTEX_INITIALIZER_NP; |
674 | |
675 | /** |
676 | * Acquire a lock on the exchange subsystem. |
677 | */ |
678 | #define EXCHANGE_LOCK() \ |
679 | ({ \ |
680 | pthread_mutex_lock (&g_exchange_lock); \ |
681 | LOCK_DEBUG("locked in %s", __FUNCTION__); \ |
682 | }) |
683 | |
684 | /** |
685 | * Relinquish the lock on the exchange subsystem. |
686 | */ |
687 | #define EXCHANGE_UNLOCK() \ |
688 | ({ \ |
689 | pthread_mutex_unlock (&g_exchange_lock); \ |
690 | LOCK_DEBUG("unLocked in %s", __FUNCTION__); \ |
691 | }) |
692 | |
693 | /** |
694 | * Global lock to set or get committed exchange cluster. This is a lower |
695 | * granularity lock to allow read access to committed cluster even while |
696 | * exchange is busy for example rebalancing under the exchange lock. |
697 | */ |
698 | pthread_rwlock_t g_exchange_commited_cluster_lock = PTHREAD_RWLOCK_INITIALIZER; |
699 | |
700 | /** |
701 | * Acquire a read lock on the committed exchange cluster. |
702 | */ |
703 | #define EXCHANGE_COMMITTED_CLUSTER_RLOCK() \ |
704 | ({ \ |
705 | pthread_rwlock_rdlock (&g_exchange_commited_cluster_lock); \ |
706 | LOCK_DEBUG("committed data locked in %s", __FUNCTION__); \ |
707 | }) |
708 | |
709 | /** |
710 | * Acquire a write lock on the committed exchange cluster. |
711 | */ |
712 | #define EXCHANGE_COMMITTED_CLUSTER_WLOCK() \ |
713 | ({ \ |
714 | pthread_rwlock_wrlock (&g_exchange_commited_cluster_lock); \ |
715 | LOCK_DEBUG("committed data locked in %s", __FUNCTION__); \ |
716 | }) |
717 | |
718 | /** |
719 | * Relinquish the lock on the committed exchange cluster. |
720 | */ |
721 | #define EXCHANGE_COMMITTED_CLUSTER_UNLOCK() \ |
722 | ({ \ |
723 | pthread_rwlock_unlock (&g_exchange_commited_cluster_lock); \ |
724 | LOCK_DEBUG("committed data unLocked in %s", __FUNCTION__); \ |
725 | }) |
726 | |
727 | /** |
728 | * Singleton external events publisher. |
729 | */ |
730 | static as_exchange_external_event_publisher g_external_event_publisher; |
731 | |
732 | /** |
733 | * The fat lock for all clustering events listener changes. |
734 | */ |
735 | static pthread_mutex_t g_external_event_publisher_lock = |
736 | PTHREAD_RECURSIVE_MUTEX_INITIALIZER_NP; |
737 | |
738 | /** |
739 | * Acquire a lock on the event publisher. |
740 | */ |
741 | #define EXTERNAL_EVENT_PUBLISHER_LOCK() \ |
742 | ({ \ |
743 | pthread_mutex_lock (&g_external_event_publisher_lock); \ |
744 | LOCK_DEBUG("publisher locked in %s", __FUNCTION__); \ |
745 | }) |
746 | |
747 | /** |
748 | * Relinquish the lock on the external event publisher. |
749 | */ |
750 | #define EXTERNAL_EVENT_PUBLISHER_UNLOCK() \ |
751 | ({ \ |
752 | pthread_mutex_unlock (&g_external_event_publisher_lock); \ |
753 | LOCK_DEBUG("publisher unLocked in %s", __FUNCTION__); \ |
754 | }) |
755 | |
756 | /* |
757 | * ---------------------------------------------------------------------------- |
758 | * Logging macros. |
759 | * ---------------------------------------------------------------------------- |
760 | */ |
761 | |
762 | /** |
763 | * Used to limit potentially long log lines. Includes space for NULL terminator. |
764 | */ |
765 | #define CRASH(format, ...) cf_crash(AS_EXCHANGE, format, ##__VA_ARGS__) |
766 | #define WARNING(format, ...) cf_warning(AS_EXCHANGE, format, ##__VA_ARGS__) |
767 | #define INFO(format, ...) cf_info(AS_EXCHANGE, format, ##__VA_ARGS__) |
768 | #define DEBUG(format, ...) cf_debug(AS_EXCHANGE, format, ##__VA_ARGS__) |
769 | #define DETAIL(format, ...) cf_detail(AS_EXCHANGE, format, ##__VA_ARGS__) |
770 | #define LOG(severity, format, ...) \ |
771 | ({ \ |
772 | switch (severity) { \ |
773 | case CF_CRITICAL: \ |
774 | CRASH(format, ##__VA_ARGS__); \ |
775 | break; \ |
776 | case CF_WARNING: \ |
777 | WARNING(format, ##__VA_ARGS__); \ |
778 | break; \ |
779 | case CF_INFO: \ |
780 | INFO(format, ##__VA_ARGS__); \ |
781 | break; \ |
782 | case CF_DEBUG: \ |
783 | DEBUG(format, ##__VA_ARGS__); \ |
784 | break; \ |
785 | case CF_DETAIL: \ |
786 | DETAIL(format, ##__VA_ARGS__); \ |
787 | break; \ |
788 | default: \ |
789 | break; \ |
790 | } \ |
791 | }) |
792 | |
793 | /** |
794 | * Size of the (per-namespace) self payload dynamic buffer. |
795 | */ |
796 | #define AS_EXCHANGE_SELF_DYN_BUF_SIZE() (AS_EXCHANGE_UNIQUE_VINFO_MAX_SIZE_SOFT \ |
797 | * ((AS_EXCHANGE_VINFO_NUM_PIDS_AVG * sizeof(uint16_t)) \ |
798 | + sizeof(as_partition_version))) |
799 | |
800 | /** |
801 | * Scratch size for exchange messages. |
802 | * TODO: Compute this properly. |
803 | */ |
804 | #define AS_EXCHANGE_MSG_SCRATCH_SIZE 2048 |
805 | |
806 | #ifdef LOCK_DEBUG_ENABLED |
807 | #define LOCK_DEBUG(format, ...) DEBUG(format, ##__VA_ARGS__) |
808 | #else |
809 | #define LOCK_DEBUG(format, ...) |
810 | #endif |
811 | |
812 | /** |
813 | * Timer event generation interval. |
814 | */ |
815 | #define EXCHANGE_TIMER_TICK_INTERVAL() (75) |
816 | |
817 | /** |
818 | * Minimum timeout interval for sent exchange data. |
819 | */ |
820 | #define EXCHANGE_SEND_MIN_TIMEOUT() (MAX(75, as_hb_tx_interval_get() / 2)) |
821 | |
822 | /** |
823 | * Maximum timeout interval for sent exchange data. |
824 | */ |
825 | #define EXCHANGE_SEND_MAX_TIMEOUT() (30000) |
826 | |
827 | /** |
828 | * Timeout for receiving commit message after transitioning to ready to commit. |
829 | */ |
830 | #define EXCHANGE_READY_TO_COMMIT_TIMEOUT() (EXCHANGE_SEND_MIN_TIMEOUT()) |
831 | |
832 | /** |
833 | * Send timeout is a step function with this value as the interval for each |
834 | * step. |
835 | */ |
836 | #define EXCHANGE_SEND_STEP_INTERVAL() \ |
837 | (MAX(EXCHANGE_SEND_MIN_TIMEOUT(), as_hb_tx_interval_get())) |
838 | |
839 | /** |
840 | * Check if exchange is initialized. |
841 | */ |
842 | #define EXCHANGE_IS_INITIALIZED() \ |
843 | ({ \ |
844 | EXCHANGE_LOCK(); \ |
845 | bool initialized = (g_exchange.sys_state \ |
846 | != AS_EXCHANGE_SYS_STATE_UNINITIALIZED); \ |
847 | EXCHANGE_UNLOCK(); \ |
848 | initialized; \ |
849 | }) |
850 | |
851 | /** |
852 | * * Check if exchange is running. |
853 | */ |
854 | #define EXCHANGE_IS_RUNNING() \ |
855 | ({ \ |
856 | EXCHANGE_LOCK(); \ |
857 | bool running = (EXCHANGE_IS_INITIALIZED() \ |
858 | && g_exchange.sys_state == AS_EXCHANGE_SYS_STATE_RUNNING); \ |
859 | EXCHANGE_UNLOCK(); \ |
860 | running; \ |
861 | }) |
862 | |
863 | /** |
864 | * Create temporary stack variables. |
865 | */ |
866 | #define TOKEN_PASTE(x, y) x##y |
867 | #define STACK_VAR(x, y) TOKEN_PASTE(x, y) |
868 | |
869 | /** |
870 | * Convert a vector to a stack allocated array. |
871 | */ |
872 | #define cf_vector_to_stack_array(vector_p, nodes_array_p, num_nodes_p) \ |
873 | ({ \ |
874 | *num_nodes_p = cf_vector_size(vector_p); \ |
875 | if (*num_nodes_p > 0) { \ |
876 | *nodes_array_p = alloca(sizeof(cf_node) * (*num_nodes_p)); \ |
877 | for (int i = 0; i < *num_nodes_p; i++) { \ |
878 | cf_vector_get(vector_p, i, &(*nodes_array_p)[i]); \ |
879 | } \ |
880 | } \ |
881 | else { \ |
882 | *nodes_array_p = NULL; \ |
883 | } \ |
884 | }) |
885 | |
886 | /** |
887 | * Create and initialize a lockless stack allocated vector to initially sized to |
888 | * store cluster node number of elements. |
889 | */ |
890 | #define cf_vector_stack_create(value_type) \ |
891 | ({ \ |
892 | cf_vector * STACK_VAR(vector, __LINE__) = (cf_vector*)alloca( \ |
893 | sizeof(cf_vector)); \ |
894 | size_t buffer_size = AS_EXCHANGE_CLUSTER_MAX_SIZE_SOFT \ |
895 | * sizeof(value_type); \ |
896 | void* STACK_VAR(buff, __LINE__) = alloca(buffer_size); cf_vector_init_smalloc( \ |
897 | STACK_VAR(vector, __LINE__), sizeof(value_type), \ |
898 | (uint8_t*)STACK_VAR(buff, __LINE__), buffer_size, \ |
899 | VECTOR_FLAG_INITZERO); \ |
900 | STACK_VAR(vector, __LINE__); \ |
901 | }) |
902 | |
903 | /* |
904 | * ---------------------------------------------------------------------------- |
905 | * Vector functions to be moved to cf_vector |
906 | * ---------------------------------------------------------------------------- |
907 | */ |
908 | |
909 | /** |
910 | * Convert a vector to an array. |
911 | * FIXME: return pointer to the internal vector storage. |
912 | */ |
913 | static cf_node* |
914 | vector_to_array(cf_vector* vector) |
915 | { |
916 | return (cf_node*)vector->vector; |
917 | } |
918 | |
919 | /** |
920 | * Clear / delete all entries in a vector. |
921 | */ |
922 | static void |
923 | vector_clear(cf_vector* vector) |
924 | { |
925 | cf_vector_delete_range(vector, 0, cf_vector_size(vector)); |
926 | } |
927 | |
928 | /** |
929 | * Find the index of an element in the vector. Equality is based on mem compare. |
930 | * |
931 | * @param vector the source vector. |
932 | * @param element the element to find. |
933 | * @return the index if the element is found, -1 otherwise. |
934 | */ |
935 | static int |
936 | vector_find(cf_vector* vector, const void* element) |
937 | { |
938 | int element_count = cf_vector_size(vector); |
939 | size_t value_len = VECTOR_ELEM_SZ(vector); |
940 | for (int i = 0; i < element_count; i++) { |
941 | // No null check required since we are iterating under a lock and within |
942 | // vector bounds. |
943 | void* src_element = cf_vector_getp(vector, i); |
944 | if (src_element) { |
945 | if (memcmp(element, src_element, value_len) == 0) { |
946 | return i; |
947 | } |
948 | } |
949 | } |
950 | return -1; |
951 | } |
952 | |
953 | /** |
954 | * Copy all elements form the source vector to the destination vector to the |
955 | * destination vector. Assumes the source and destination vector are not being |
956 | * modified while the copy operation is in progress. |
957 | * |
958 | * @param dest the destination vector. |
959 | * @param src the source vector. |
960 | * @return the number of elements copied. |
961 | */ |
962 | static int |
963 | vector_copy(cf_vector* dest, cf_vector* src) |
964 | { |
965 | int element_count = cf_vector_size(src); |
966 | int copied_count = 0; |
967 | for (int i = 0; i < element_count; i++) { |
968 | // No null check required since we are iterating under a lock and within |
969 | // vector bounds. |
970 | void* src_element = cf_vector_getp(src, i); |
971 | if (src_element) { |
972 | cf_vector_append(dest, src_element); |
973 | copied_count++; |
974 | } |
975 | } |
976 | return copied_count; |
977 | } |
978 | |
979 | /** |
980 | * Generate a hash code for a blob using Jenkins hash function. |
981 | */ |
982 | static uint32_t |
983 | exchange_blob_hash(const uint8_t* value, size_t value_size) |
984 | { |
985 | uint32_t hash = 0; |
986 | for (int i = 0; i < value_size; ++i) { |
987 | hash += value[i]; |
988 | hash += (hash << 10); |
989 | hash ^= (hash >> 6); |
990 | } |
991 | hash += (hash << 3); |
992 | hash ^= (hash >> 11); |
993 | hash += (hash << 15); |
994 | |
995 | return hash; |
996 | } |
997 | |
998 | /** |
999 | * Generate a hash code for a mesh node key. |
1000 | */ |
1001 | static uint32_t |
1002 | exchange_vinfo_shash(const void* value) |
1003 | { |
1004 | return exchange_blob_hash((const uint8_t*)value, |
1005 | sizeof(as_partition_version)); |
1006 | } |
1007 | |
1008 | /* |
1009 | * ---------------------------------------------------------------------------- |
1010 | * Clustering external event publisher |
1011 | * ---------------------------------------------------------------------------- |
1012 | */ |
1013 | |
1014 | /** |
1015 | * * Check if event publisher is running. |
1016 | */ |
1017 | static bool |
1018 | exchange_external_event_publisher_is_running() |
1019 | { |
1020 | EXTERNAL_EVENT_PUBLISHER_LOCK(); |
1021 | bool running = g_external_event_publisher.sys_state |
1022 | == AS_EXCHANGE_SYS_STATE_RUNNING; |
1023 | EXTERNAL_EVENT_PUBLISHER_UNLOCK(); |
1024 | return running; |
1025 | } |
1026 | |
1027 | /** |
1028 | * Initialize the event publisher. |
1029 | */ |
1030 | static void |
1031 | exchange_external_event_publisher_init() |
1032 | { |
1033 | EXTERNAL_EVENT_PUBLISHER_LOCK(); |
1034 | memset(&g_external_event_publisher, 0, sizeof(g_external_event_publisher)); |
1035 | cf_vector_init(&g_external_event_publisher.published_succession_list, |
1036 | sizeof(cf_node), |
1037 | AS_EXCHANGE_CLUSTER_MAX_SIZE_SOFT, VECTOR_FLAG_INITZERO); |
1038 | |
1039 | pthread_mutex_init(&g_external_event_publisher.is_pending_mutex, NULL); |
1040 | pthread_cond_init(&g_external_event_publisher.is_pending, NULL); |
1041 | EXTERNAL_EVENT_PUBLISHER_UNLOCK(); |
1042 | } |
1043 | |
1044 | /** |
1045 | * Register a clustering event listener. |
1046 | */ |
1047 | static void |
1048 | exchange_external_event_listener_register( |
1049 | as_exchange_cluster_changed_cb event_callback, void* udata) |
1050 | { |
1051 | EXTERNAL_EVENT_PUBLISHER_LOCK(); |
1052 | |
1053 | if (g_external_event_publisher.event_listener_count |
1054 | >= AS_EXTERNAL_EVENT_LISTENER_MAX) { |
1055 | CRASH("cannot register more than %d event listeners" , |
1056 | AS_EXTERNAL_EVENT_LISTENER_MAX); |
1057 | } |
1058 | |
1059 | g_external_event_publisher.event_listeners[g_external_event_publisher.event_listener_count].event_callback = |
1060 | event_callback; |
1061 | g_external_event_publisher.event_listeners[g_external_event_publisher.event_listener_count].udata = |
1062 | udata; |
1063 | g_external_event_publisher.event_listener_count++; |
1064 | |
1065 | EXTERNAL_EVENT_PUBLISHER_UNLOCK(); |
1066 | } |
1067 | |
1068 | /** |
1069 | * Wakeup the publisher thread. |
1070 | */ |
1071 | static void |
1072 | exchange_external_event_publisher_thr_wakeup() |
1073 | { |
1074 | pthread_mutex_lock(&g_external_event_publisher.is_pending_mutex); |
1075 | pthread_cond_signal(&g_external_event_publisher.is_pending); |
1076 | pthread_mutex_unlock(&g_external_event_publisher.is_pending_mutex); |
1077 | } |
1078 | |
1079 | /** |
1080 | * Queue up and external event to publish. |
1081 | */ |
1082 | static void |
1083 | exchange_external_event_queue(as_exchange_cluster_changed_event* event) |
1084 | { |
1085 | EXTERNAL_EVENT_PUBLISHER_LOCK(); |
1086 | memcpy(&g_external_event_publisher.to_publish, event, |
1087 | sizeof(g_external_event_publisher.to_publish)); |
1088 | |
1089 | vector_clear(&g_external_event_publisher.published_succession_list); |
1090 | if (event->succession) { |
1091 | // Use the static list for the published event, so that the input event |
1092 | // object can be destroyed irrespective of when the it is published. |
1093 | for (int i = 0; i < event->cluster_size; i++) { |
1094 | cf_vector_append( |
1095 | &g_external_event_publisher.published_succession_list, |
1096 | &event->succession[i]); |
1097 | } |
1098 | g_external_event_publisher.to_publish.succession = vector_to_array( |
1099 | &g_external_event_publisher.published_succession_list); |
1100 | |
1101 | } |
1102 | else { |
1103 | g_external_event_publisher.to_publish.succession = NULL; |
1104 | } |
1105 | |
1106 | g_external_event_publisher.event_queued = true; |
1107 | |
1108 | EXTERNAL_EVENT_PUBLISHER_UNLOCK(); |
1109 | |
1110 | // Wake up the publisher thread. |
1111 | exchange_external_event_publisher_thr_wakeup(); |
1112 | } |
1113 | |
1114 | /** |
1115 | * Publish external events if any are pending. |
1116 | */ |
1117 | static void |
1118 | exchange_external_events_publish() |
1119 | { |
1120 | EXTERNAL_EVENT_PUBLISHER_LOCK(); |
1121 | |
1122 | if (g_external_event_publisher.event_queued) { |
1123 | g_external_event_publisher.event_queued = false; |
1124 | for (uint32_t i = 0; |
1125 | i < g_external_event_publisher.event_listener_count; i++) { |
1126 | (g_external_event_publisher.event_listeners[i].event_callback)( |
1127 | &g_external_event_publisher.to_publish, |
1128 | g_external_event_publisher.event_listeners[i].udata); |
1129 | } |
1130 | } |
1131 | EXTERNAL_EVENT_PUBLISHER_UNLOCK(); |
1132 | } |
1133 | |
1134 | /** |
1135 | * External event publisher thread. |
1136 | */ |
1137 | static void* |
1138 | exchange_external_event_publisher_thr(void* arg) |
1139 | { |
1140 | pthread_mutex_lock(&g_external_event_publisher.is_pending_mutex); |
1141 | |
1142 | while (true) { |
1143 | pthread_cond_wait(&g_external_event_publisher.is_pending, |
1144 | &g_external_event_publisher.is_pending_mutex); |
1145 | if (exchange_external_event_publisher_is_running()) { |
1146 | exchange_external_events_publish(); |
1147 | } |
1148 | else { |
1149 | // Publisher stopped, exit the tread. |
1150 | break; |
1151 | } |
1152 | } |
1153 | |
1154 | return NULL; |
1155 | } |
1156 | |
1157 | /** |
1158 | * Start the event publisher. |
1159 | */ |
1160 | static void |
1161 | exchange_external_event_publisher_start() |
1162 | { |
1163 | EXTERNAL_EVENT_PUBLISHER_LOCK(); |
1164 | g_external_event_publisher.sys_state = AS_EXCHANGE_SYS_STATE_RUNNING; |
1165 | g_external_event_publisher.event_publisher_tid = |
1166 | cf_thread_create_joinable(exchange_external_event_publisher_thr, |
1167 | NULL); |
1168 | EXTERNAL_EVENT_PUBLISHER_UNLOCK(); |
1169 | } |
1170 | |
1171 | /** |
1172 | * Stop the event publisher. |
1173 | */ |
1174 | static void |
1175 | external_event_publisher_stop() |
1176 | { |
1177 | EXTERNAL_EVENT_PUBLISHER_LOCK(); |
1178 | g_external_event_publisher.sys_state = AS_EXCHANGE_SYS_STATE_SHUTTING_DOWN; |
1179 | EXTERNAL_EVENT_PUBLISHER_UNLOCK(); |
1180 | |
1181 | exchange_external_event_publisher_thr_wakeup(); |
1182 | cf_thread_join(g_external_event_publisher.event_publisher_tid); |
1183 | |
1184 | EXTERNAL_EVENT_PUBLISHER_LOCK(); |
1185 | g_external_event_publisher.sys_state = AS_EXCHANGE_SYS_STATE_STOPPED; |
1186 | g_external_event_publisher.event_queued = false; |
1187 | EXTERNAL_EVENT_PUBLISHER_UNLOCK(); |
1188 | } |
1189 | |
1190 | /* |
1191 | * ---------------------------------------------------------------------------- |
1192 | * Node state related |
1193 | * ---------------------------------------------------------------------------- |
1194 | */ |
1195 | |
1196 | /** |
1197 | * Initialize node state. |
1198 | */ |
1199 | static void |
1200 | exchange_node_state_init(as_exchange_node_state* node_state) |
1201 | { |
1202 | memset(node_state, 0, sizeof(*node_state)); |
1203 | |
1204 | node_state->data = cf_calloc(1, sizeof(as_exchange_node_data)); |
1205 | } |
1206 | |
1207 | /** |
1208 | * Reset node state. |
1209 | */ |
1210 | static void |
1211 | exchange_node_state_reset(as_exchange_node_state* node_state) |
1212 | { |
1213 | node_state->send_acked = false; |
1214 | node_state->received = false; |
1215 | node_state->is_ready_to_commit = false; |
1216 | |
1217 | node_state->data->num_namespaces = 0; |
1218 | for (int i = 0; i < AS_NAMESPACE_SZ; i++) { |
1219 | node_state->data->namespace_data[i].local_namespace = NULL; |
1220 | } |
1221 | } |
1222 | |
1223 | /** |
1224 | * Destroy node state. |
1225 | */ |
1226 | static void |
1227 | exchange_node_state_destroy(as_exchange_node_state* node_state) |
1228 | { |
1229 | for (int i = 0; i < AS_NAMESPACE_SZ; i++) { |
1230 | if (node_state->data->namespace_data[i].partition_versions) { |
1231 | cf_free(node_state->data->namespace_data[i].partition_versions); |
1232 | } |
1233 | |
1234 | if (node_state->data->namespace_data[i].roster) { |
1235 | cf_free(node_state->data->namespace_data[i].roster); |
1236 | } |
1237 | |
1238 | if (node_state->data->namespace_data[i].roster_rack_ids) { |
1239 | cf_free(node_state->data->namespace_data[i].roster_rack_ids); |
1240 | } |
1241 | } |
1242 | |
1243 | cf_free(node_state->data); |
1244 | } |
1245 | |
1246 | /** |
1247 | * Reduce function to match node -> node state hash to the succession list. |
1248 | * Should always be invoked under a lock over the main hash. |
1249 | */ |
1250 | static int |
1251 | exchange_node_states_reset_reduce(const void* key, void* data, void* udata) |
1252 | { |
1253 | const cf_node* node = (const cf_node*)key; |
1254 | as_exchange_node_state* node_state = (as_exchange_node_state*)data; |
1255 | |
1256 | int node_index = vector_find(&g_exchange.succession_list, node); |
1257 | if (node_index < 0) { |
1258 | // Node not in succession list |
1259 | exchange_node_state_destroy(node_state); |
1260 | return CF_SHASH_REDUCE_DELETE; |
1261 | } |
1262 | |
1263 | exchange_node_state_reset(node_state); |
1264 | return CF_SHASH_OK; |
1265 | } |
1266 | |
1267 | /** |
1268 | * Adjust the nodeid_to_node_state hash to have an entry for every node in the |
1269 | * succession list with state reset for a new round of exchange. Removes entries |
1270 | * not in the succession list. |
1271 | */ |
1272 | static void |
1273 | exchange_node_states_reset() |
1274 | { |
1275 | EXCHANGE_LOCK(); |
1276 | |
1277 | // Fix existing entries by reseting entries in succession and removing |
1278 | // entries not in succession list. |
1279 | cf_shash_reduce(g_exchange.nodeid_to_node_state, |
1280 | exchange_node_states_reset_reduce, NULL); |
1281 | |
1282 | // Add missing entries. |
1283 | int succession_length = cf_vector_size(&g_exchange.succession_list); |
1284 | |
1285 | as_exchange_node_state temp_state; |
1286 | for (int i = 0; i < succession_length; i++) { |
1287 | cf_node nodeid; |
1288 | |
1289 | cf_vector_get(&g_exchange.succession_list, i, &nodeid); |
1290 | if (cf_shash_get(g_exchange.nodeid_to_node_state, &nodeid, &temp_state) |
1291 | == CF_SHASH_ERR_NOT_FOUND) { |
1292 | exchange_node_state_init(&temp_state); |
1293 | |
1294 | cf_shash_put(g_exchange.nodeid_to_node_state, &nodeid, &temp_state); |
1295 | } |
1296 | } |
1297 | |
1298 | EXCHANGE_UNLOCK(); |
1299 | } |
1300 | |
1301 | /** |
1302 | * Reduce function to find nodes that had not acked self node's exchange data. |
1303 | */ |
1304 | static int |
1305 | exchange_nodes_find_send_unacked_reduce(const void* key, void* data, |
1306 | void* udata) |
1307 | { |
1308 | const cf_node* node = (const cf_node*)key; |
1309 | as_exchange_node_state* node_state = (as_exchange_node_state*)data; |
1310 | cf_vector* unacked = (cf_vector*)udata; |
1311 | |
1312 | if (!node_state->send_acked) { |
1313 | cf_vector_append(unacked, node); |
1314 | } |
1315 | return CF_SHASH_OK; |
1316 | } |
1317 | |
1318 | /** |
1319 | * Find nodes that have not acked self node's exchange data. |
1320 | */ |
1321 | static void |
1322 | exchange_nodes_find_send_unacked(cf_vector* unacked) |
1323 | { |
1324 | cf_shash_reduce(g_exchange.nodeid_to_node_state, |
1325 | exchange_nodes_find_send_unacked_reduce, unacked); |
1326 | } |
1327 | |
1328 | /** |
1329 | * Reduce function to find peer nodes from whom self node has not received |
1330 | * exchange data. |
1331 | */ |
1332 | static int |
1333 | exchange_nodes_find_not_received_reduce(const void* key, void* data, |
1334 | void* udata) |
1335 | { |
1336 | const cf_node* node = (const cf_node*)key; |
1337 | as_exchange_node_state* node_state = (as_exchange_node_state*)data; |
1338 | cf_vector* not_received = (cf_vector*)udata; |
1339 | |
1340 | if (!node_state->received) { |
1341 | cf_vector_append(not_received, node); |
1342 | } |
1343 | return CF_SHASH_OK; |
1344 | } |
1345 | |
1346 | /** |
1347 | * Find peer nodes from whom self node has not received exchange data. |
1348 | */ |
1349 | static void |
1350 | exchange_nodes_find_not_received(cf_vector* not_received) |
1351 | { |
1352 | cf_shash_reduce(g_exchange.nodeid_to_node_state, |
1353 | exchange_nodes_find_not_received_reduce, not_received); |
1354 | } |
1355 | |
1356 | /** |
1357 | * Reduce function to find peer nodes that are not ready to commit. |
1358 | */ |
1359 | static int |
1360 | exchange_nodes_find_not_ready_to_commit_reduce(const void* key, void* data, |
1361 | void* udata) |
1362 | { |
1363 | const cf_node* node = (const cf_node*)key; |
1364 | as_exchange_node_state* node_state = (as_exchange_node_state*)data; |
1365 | cf_vector* not_ready_to_commit = (cf_vector*)udata; |
1366 | |
1367 | if (!node_state->is_ready_to_commit) { |
1368 | cf_vector_append(not_ready_to_commit, node); |
1369 | } |
1370 | return CF_SHASH_OK; |
1371 | } |
1372 | |
1373 | /** |
1374 | * Find peer nodes that are not ready to commit. |
1375 | */ |
1376 | static void |
1377 | exchange_nodes_find_not_ready_to_commit(cf_vector* not_ready_to_commit) |
1378 | { |
1379 | cf_shash_reduce(g_exchange.nodeid_to_node_state, |
1380 | exchange_nodes_find_not_ready_to_commit_reduce, |
1381 | not_ready_to_commit); |
1382 | } |
1383 | |
1384 | /** |
1385 | * Update the node state for a node. |
1386 | */ |
1387 | static void |
1388 | exchange_node_state_update(cf_node nodeid, as_exchange_node_state* node_state) |
1389 | { |
1390 | cf_shash_put(g_exchange.nodeid_to_node_state, &nodeid, node_state); |
1391 | } |
1392 | |
1393 | /** |
1394 | * Get state of a node from the hash. If not found crash because this entry |
1395 | * should be present in the hash. |
1396 | */ |
1397 | static void |
1398 | exchange_node_state_get_safe(cf_node nodeid, as_exchange_node_state* node_state) |
1399 | { |
1400 | if (cf_shash_get(g_exchange.nodeid_to_node_state, &nodeid, node_state) |
1401 | == CF_SHASH_ERR_NOT_FOUND) { |
1402 | CRASH( |
1403 | "node entry for node %" PRIx64" missing from node state hash" , nodeid); |
1404 | } |
1405 | } |
1406 | |
1407 | /* |
1408 | * ---------------------------------------------------------------------------- |
1409 | * Message related |
1410 | * ---------------------------------------------------------------------------- |
1411 | */ |
1412 | |
1413 | /** |
1414 | * Fill compulsary fields in a message common to all message types. |
1415 | */ |
1416 | static void |
1417 | exchange_msg_src_fill(msg* msg, as_exchange_msg_type type) |
1418 | { |
1419 | EXCHANGE_LOCK(); |
1420 | msg_set_uint32(msg, AS_EXCHANGE_MSG_ID, AS_EXCHANGE_PROTOCOL_IDENTIFIER); |
1421 | msg_set_uint64(msg, AS_EXCHANGE_MSG_CLUSTER_KEY, g_exchange.cluster_key); |
1422 | msg_set_uint32(msg, AS_EXCHANGE_MSG_TYPE, type); |
1423 | EXCHANGE_UNLOCK(); |
1424 | } |
1425 | |
1426 | /** |
1427 | * Get the msg buffer from a pool and fill in all compulsory fields. |
1428 | * @return the msg buff with compulsory fields filled in. |
1429 | */ |
1430 | static msg* |
1431 | exchange_msg_get(as_exchange_msg_type type) |
1432 | { |
1433 | msg* msg = as_fabric_msg_get(M_TYPE_EXCHANGE); |
1434 | exchange_msg_src_fill(msg, type); |
1435 | return msg; |
1436 | } |
1437 | |
1438 | /** |
1439 | * Return the message buffer back to the pool. |
1440 | */ |
1441 | static void |
1442 | exchange_msg_return(msg* msg) |
1443 | { |
1444 | as_fabric_msg_put(msg); |
1445 | } |
1446 | |
1447 | /** |
1448 | * Get message id. |
1449 | */ |
1450 | static int |
1451 | exchange_msg_id_get(msg* msg, uint32_t* msg_id) |
1452 | { |
1453 | if (msg_get_uint32(msg, AS_EXCHANGE_MSG_ID, msg_id) != 0) { |
1454 | return -1; |
1455 | } |
1456 | return 0; |
1457 | } |
1458 | |
1459 | /** |
1460 | * Get message type. |
1461 | */ |
1462 | static int |
1463 | exchange_msg_type_get(msg* msg, as_exchange_msg_type* msg_type) |
1464 | { |
1465 | if (msg_get_uint32(msg, AS_EXCHANGE_MSG_TYPE, msg_type) != 0) { |
1466 | return -1; |
1467 | } |
1468 | return 0; |
1469 | } |
1470 | |
1471 | /** |
1472 | * Get message cluster key. |
1473 | */ |
1474 | static int |
1475 | exchange_msg_cluster_key_get(msg* msg, as_cluster_key* cluster_key) |
1476 | { |
1477 | if (msg_get_uint64(msg, AS_EXCHANGE_MSG_CLUSTER_KEY, cluster_key) != 0) { |
1478 | return -1; |
1479 | } |
1480 | return 0; |
1481 | } |
1482 | |
1483 | /** |
1484 | * Set data payload for a message. |
1485 | */ |
1486 | static void |
1487 | exchange_msg_data_payload_set(msg* msg) |
1488 | { |
1489 | uint32_t ns_count = g_config.n_namespaces; |
1490 | |
1491 | cf_vector_define(namespace_list, sizeof(msg_buf_ele), ns_count, 0); |
1492 | cf_vector_define(partition_versions, sizeof(msg_buf_ele), ns_count, 0); |
1493 | uint32_t rack_ids[ns_count]; |
1494 | |
1495 | bool have_roster = false; |
1496 | bool have_roster_rack_ids = false; |
1497 | uint32_t roster_generations[ns_count]; |
1498 | cf_vector_define(rosters, sizeof(msg_buf_ele), ns_count, 0); |
1499 | cf_vector_define(rosters_rack_ids, sizeof(msg_buf_ele), ns_count, 0); |
1500 | |
1501 | bool have_regimes = false; |
1502 | uint32_t eventual_regimes[ns_count]; |
1503 | uint32_t rebalance_regimes[ns_count]; |
1504 | |
1505 | bool have_rebalance_flags = false; |
1506 | uint32_t rebalance_flags[ns_count]; |
1507 | |
1508 | memset(rebalance_flags, 0, sizeof(rebalance_flags)); |
1509 | |
1510 | msg_set_uint32(msg, AS_EXCHANGE_MSG_COMPATIBILITY_ID, |
1511 | AS_EXCHANGE_COMPATIBILITY_ID); |
1512 | |
1513 | for (uint32_t ns_ix = 0; ns_ix < ns_count; ns_ix++) { |
1514 | as_namespace* ns = g_config.namespaces[ns_ix]; |
1515 | |
1516 | msg_buf_ele ns_ele = { |
1517 | .sz = (uint32_t)strlen(ns->name), |
1518 | .ptr = (uint8_t*)ns->name |
1519 | }; |
1520 | |
1521 | msg_buf_ele pv_ele = { |
1522 | .sz = (uint32_t)g_exchange.self_data_dyn_buf[ns_ix].used_sz, |
1523 | .ptr = g_exchange.self_data_dyn_buf[ns_ix].buf |
1524 | }; |
1525 | |
1526 | msg_buf_ele rn_ele = { |
1527 | .sz = (uint32_t)(ns->smd_roster_count * sizeof(cf_node)), |
1528 | .ptr = (uint8_t*)ns->smd_roster |
1529 | }; |
1530 | |
1531 | msg_buf_ele rri_ele = { |
1532 | .sz = (uint32_t)(ns->smd_roster_count * sizeof(uint32_t)), |
1533 | .ptr = (uint8_t*)ns->smd_roster_rack_ids |
1534 | }; |
1535 | |
1536 | cf_vector_append(&namespace_list, &ns_ele); |
1537 | cf_vector_append(&partition_versions, &pv_ele); |
1538 | rack_ids[ns_ix] = ns->rack_id; |
1539 | |
1540 | if (ns->smd_roster_generation != 0) { |
1541 | have_roster = true; |
1542 | |
1543 | if (! have_roster_rack_ids) { |
1544 | for (uint32_t n = 0; n < ns->smd_roster_count; n++) { |
1545 | if (ns->smd_roster_rack_ids[n] != 0) { |
1546 | have_roster_rack_ids = true; |
1547 | break; |
1548 | } |
1549 | } |
1550 | } |
1551 | } |
1552 | |
1553 | roster_generations[ns_ix] = ns->smd_roster_generation; |
1554 | cf_vector_append(&rosters, &rn_ele); |
1555 | cf_vector_append(&rosters_rack_ids, &rri_ele); |
1556 | |
1557 | eventual_regimes[ns_ix] = ns->eventual_regime; |
1558 | rebalance_regimes[ns_ix] = ns->rebalance_regime; |
1559 | |
1560 | if (eventual_regimes[ns_ix] != 0 || rebalance_regimes[ns_ix] != 0) { |
1561 | have_regimes = true; |
1562 | } |
1563 | |
1564 | if (ns->cfg_prefer_uniform_balance) { |
1565 | rebalance_flags[ns_ix] |= AS_EXCHANGE_REBALANCE_FLAG_UNIFORM; |
1566 | } |
1567 | |
1568 | if (ns->pending_quiesce) { |
1569 | rebalance_flags[ns_ix] |= AS_EXCHANGE_REBALANCE_FLAG_QUIESCE; |
1570 | } |
1571 | |
1572 | if (rebalance_flags[ns_ix] != 0) { |
1573 | have_rebalance_flags = true; |
1574 | } |
1575 | } |
1576 | |
1577 | msg_msgpack_list_set_buf(msg, AS_EXCHANGE_MSG_NAMESPACES, &namespace_list); |
1578 | msg_msgpack_list_set_buf(msg, AS_EXCHANGE_MSG_NS_PARTITION_VERSIONS, |
1579 | &partition_versions); |
1580 | msg_msgpack_list_set_uint32(msg, AS_EXCHANGE_MSG_NS_RACK_IDS, rack_ids, |
1581 | ns_count); |
1582 | |
1583 | if (have_roster) { |
1584 | msg_msgpack_list_set_uint32(msg, AS_EXCHANGE_MSG_NS_ROSTER_GENERATIONS, |
1585 | roster_generations, ns_count); |
1586 | msg_msgpack_list_set_buf(msg, AS_EXCHANGE_MSG_NS_ROSTERS, &rosters); |
1587 | |
1588 | if (have_roster_rack_ids) { |
1589 | msg_msgpack_list_set_buf(msg, AS_EXCHANGE_MSG_NS_ROSTERS_RACK_IDS, |
1590 | &rosters_rack_ids); |
1591 | } |
1592 | } |
1593 | |
1594 | if (have_regimes) { |
1595 | msg_msgpack_list_set_uint32(msg, AS_EXCHANGE_MSG_NS_EVENTUAL_REGIMES, |
1596 | eventual_regimes, ns_count); |
1597 | msg_msgpack_list_set_uint32(msg, AS_EXCHANGE_MSG_NS_REBALANCE_REGIMES, |
1598 | rebalance_regimes, ns_count); |
1599 | } |
1600 | |
1601 | if (have_rebalance_flags) { |
1602 | msg_msgpack_list_set_uint32(msg, AS_EXCHANGE_MSG_NS_REBALANCE_FLAGS, |
1603 | rebalance_flags, ns_count); |
1604 | } |
1605 | } |
1606 | |
1607 | /** |
1608 | * Check sanity of an incoming message. If this check passes the message is |
1609 | * guaranteed to have valid protocol identifier, valid type and valid matching |
1610 | * cluster key with source node being a part of the cluster. |
1611 | * @return 0 if the message in valid, -1 if the message is invalid and should be |
1612 | * ignored. |
1613 | */ |
1614 | static bool |
1615 | exchange_msg_is_sane(cf_node source, msg* msg) |
1616 | { |
1617 | uint32_t id = 0; |
1618 | if (exchange_msg_id_get(msg, &id) != 0|| |
1619 | id != AS_EXCHANGE_PROTOCOL_IDENTIFIER) { |
1620 | DEBUG( |
1621 | "received exchange message with mismatching identifier - expected %u but was %u" , |
1622 | AS_EXCHANGE_PROTOCOL_IDENTIFIER, id); |
1623 | return false; |
1624 | } |
1625 | |
1626 | as_exchange_msg_type msg_type = 0; |
1627 | |
1628 | if (exchange_msg_type_get(msg, &msg_type) != 0 |
1629 | || msg_type >= AS_EXCHANGE_MSG_TYPE_SENTINEL) { |
1630 | WARNING("received exchange message with invalid message type %u" , |
1631 | msg_type); |
1632 | return false; |
1633 | } |
1634 | |
1635 | EXCHANGE_LOCK(); |
1636 | as_cluster_key current_cluster_key = g_exchange.cluster_key; |
1637 | bool is_in_cluster = vector_find(&g_exchange.succession_list, &source) >= 0; |
1638 | EXCHANGE_UNLOCK(); |
1639 | |
1640 | if (!is_in_cluster) { |
1641 | DEBUG("received exchange message from node %" PRIx64" not in cluster" , |
1642 | source); |
1643 | return false; |
1644 | } |
1645 | |
1646 | as_cluster_key incoming_cluster_key = 0; |
1647 | if (exchange_msg_cluster_key_get(msg, &incoming_cluster_key) != 0 |
1648 | || (current_cluster_key != incoming_cluster_key) |
1649 | || current_cluster_key == 0) { |
1650 | DEBUG("received exchange message with mismatching cluster key - expected %" PRIx64" but was %" PRIx64, |
1651 | current_cluster_key, incoming_cluster_key); |
1652 | return false; |
1653 | } |
1654 | |
1655 | return true; |
1656 | } |
1657 | |
1658 | /** |
1659 | * Send a message over fabric. |
1660 | * |
1661 | * @param msg the message to send. |
1662 | * @param dest the desination node. |
1663 | * @param error_msg the error message. |
1664 | */ |
1665 | static void |
1666 | exchange_msg_send(msg* msg, cf_node dest, char* error_msg) |
1667 | { |
1668 | if (as_fabric_send(dest, msg, AS_FABRIC_CHANNEL_CTRL)) { |
1669 | // Fabric will not return the message to the pool. Do it ourself. |
1670 | exchange_msg_return(msg); |
1671 | WARNING("%s (dest:%" PRIx64")" , error_msg, dest); |
1672 | } |
1673 | } |
1674 | |
1675 | /** |
1676 | * Send a message over to a list of destination nodes. |
1677 | * |
1678 | * @param msg the message to send. |
1679 | * @param dests the node list to send the message to. |
1680 | * @param num_dests the number of destination nodes. |
1681 | * @param error_msg the error message. |
1682 | */ |
1683 | static void |
1684 | exchange_msg_send_list(msg* msg, cf_node* dests, int num_dests, char* error_msg) |
1685 | { |
1686 | if (as_fabric_send_list(dests, num_dests, msg, AS_FABRIC_CHANNEL_CTRL) |
1687 | != 0) { |
1688 | // Fabric will not return the message to the pool. Do it ourself. |
1689 | exchange_msg_return(msg); |
1690 | as_clustering_log_cf_node_array(CF_WARNING, AS_EXCHANGE, error_msg, |
1691 | dests, num_dests); |
1692 | } |
1693 | } |
1694 | |
1695 | /** |
1696 | * Send a commit message to a destination node. |
1697 | * @param dest the destination node. |
1698 | */ |
1699 | static void |
1700 | exchange_commit_msg_send(cf_node dest) |
1701 | { |
1702 | msg* commit_msg = exchange_msg_get(AS_EXCHANGE_MSG_TYPE_COMMIT); |
1703 | DEBUG("sending commit message to node %" PRIx64, dest); |
1704 | exchange_msg_send(commit_msg, dest, "error sending commit message" ); |
1705 | } |
1706 | |
1707 | /** |
1708 | * Send a commit message to a list of destination nodes. |
1709 | * @param dests the destination nodes. |
1710 | * @param num_dests the number of destination nodes. |
1711 | */ |
1712 | static void |
1713 | exchange_commit_msg_send_all(cf_node* dests, int num_dests) |
1714 | { |
1715 | msg* commit_msg = exchange_msg_get(AS_EXCHANGE_MSG_TYPE_COMMIT); |
1716 | as_clustering_log_cf_node_array(CF_DEBUG, AS_EXCHANGE, |
1717 | "sending commit message to nodes:" , dests, num_dests); |
1718 | exchange_msg_send_list(commit_msg, dests, num_dests, |
1719 | "error sending commit message" ); |
1720 | } |
1721 | |
1722 | /** |
1723 | * Send ready to commit message to the principal. |
1724 | */ |
1725 | static void |
1726 | exchange_ready_to_commit_msg_send() |
1727 | { |
1728 | EXCHANGE_LOCK(); |
1729 | g_exchange.ready_to_commit_send_ts = cf_getms(); |
1730 | cf_node principal = g_exchange.principal; |
1731 | EXCHANGE_UNLOCK(); |
1732 | |
1733 | msg* ready_to_commit_msg = exchange_msg_get( |
1734 | AS_EXCHANGE_MSG_TYPE_READY_TO_COMMIT); |
1735 | DEBUG("sending ready to commit message to node %" PRIx64, principal); |
1736 | exchange_msg_send(ready_to_commit_msg, principal, |
1737 | "error sending ready to commit message" ); |
1738 | } |
1739 | |
1740 | /** |
1741 | * Send exchange data to all nodes that have not acked the send. |
1742 | */ |
1743 | static void |
1744 | exchange_data_msg_send_pending_ack() |
1745 | { |
1746 | EXCHANGE_LOCK(); |
1747 | g_exchange.send_ts = cf_getms(); |
1748 | |
1749 | cf_node* unacked_nodes; |
1750 | int num_unacked_nodes; |
1751 | cf_vector* unacked_nodes_vector = cf_vector_stack_create(cf_node); |
1752 | |
1753 | exchange_nodes_find_send_unacked(unacked_nodes_vector); |
1754 | cf_vector_to_stack_array(unacked_nodes_vector, &unacked_nodes, |
1755 | &num_unacked_nodes); |
1756 | |
1757 | cf_vector_destroy(unacked_nodes_vector); |
1758 | |
1759 | if (!num_unacked_nodes) { |
1760 | goto Exit; |
1761 | } |
1762 | |
1763 | // FIXME - temporary assert, until we're sure. |
1764 | cf_assert(g_exchange.data_msg != NULL, AS_EXCHANGE, "payload not built" ); |
1765 | |
1766 | as_clustering_log_cf_node_array(CF_DEBUG, AS_EXCHANGE, |
1767 | "sending exchange data to nodes:" , unacked_nodes, |
1768 | num_unacked_nodes); |
1769 | |
1770 | msg_incr_ref(g_exchange.data_msg); |
1771 | |
1772 | exchange_msg_send_list(g_exchange.data_msg, unacked_nodes, |
1773 | num_unacked_nodes, "error sending exchange data" ); |
1774 | Exit: |
1775 | EXCHANGE_UNLOCK(); |
1776 | } |
1777 | |
1778 | /** |
1779 | * Send a commit message to a destination node. |
1780 | * @param dest the destination node. |
1781 | */ |
1782 | static void |
1783 | exchange_data_ack_msg_send(cf_node dest) |
1784 | { |
1785 | msg* ack_msg = exchange_msg_get(AS_EXCHANGE_MSG_TYPE_DATA_ACK); |
1786 | DEBUG("sending data ack message to node %" PRIx64, dest); |
1787 | exchange_msg_send(ack_msg, dest, "error sending data ack message" ); |
1788 | } |
1789 | |
1790 | /* |
1791 | * ---------------------------------------------------------------------------- |
1792 | * Data payload related |
1793 | * ---------------------------------------------------------------------------- |
1794 | */ |
1795 | |
1796 | /** |
1797 | * Add a pid to the namespace hash for the input vinfo. |
1798 | */ |
1799 | static void |
1800 | exchange_namespace_hash_pid_add(cf_shash* ns_hash, as_partition_version* vinfo, |
1801 | uint16_t pid) |
1802 | { |
1803 | if (as_partition_version_is_null(vinfo)) { |
1804 | // Ignore NULL vinfos. |
1805 | return; |
1806 | } |
1807 | |
1808 | cf_vector* pid_vector; |
1809 | |
1810 | // Append the hash. |
1811 | if (cf_shash_get(ns_hash, vinfo, &pid_vector) != CF_SHASH_OK) { |
1812 | // We are seeing this vinfo for the first time. |
1813 | pid_vector = cf_vector_create(sizeof(uint16_t), |
1814 | AS_EXCHANGE_VINFO_NUM_PIDS_AVG, 0); |
1815 | cf_shash_put(ns_hash, vinfo, &pid_vector); |
1816 | } |
1817 | |
1818 | cf_vector_append(pid_vector, &pid); |
1819 | } |
1820 | |
1821 | /** |
1822 | * Destroy the pid vector for each vinfo. |
1823 | */ |
1824 | static int |
1825 | exchange_namespace_hash_destroy_reduce(const void* key, void* data, void* udata) |
1826 | { |
1827 | cf_vector* pid_vector = *(cf_vector**)data; |
1828 | cf_vector_destroy(pid_vector); |
1829 | return CF_SHASH_REDUCE_DELETE; |
1830 | } |
1831 | |
1832 | /** |
1833 | * Serialize each vinfo and accumulated pids to the input buffer. |
1834 | */ |
1835 | static int |
1836 | exchange_namespace_hash_serialize_reduce(const void* key, void* data, |
1837 | void* udata) |
1838 | { |
1839 | const as_partition_version* vinfo = (const as_partition_version*)key; |
1840 | cf_vector* pid_vector = *(cf_vector**)data; |
1841 | cf_dyn_buf* dyn_buf = (cf_dyn_buf*)udata; |
1842 | |
1843 | // Append the vinfo. |
1844 | cf_dyn_buf_append_buf(dyn_buf, (uint8_t*)vinfo, sizeof(*vinfo)); |
1845 | |
1846 | // Append the count of pids. |
1847 | uint32_t num_pids = cf_vector_size(pid_vector); |
1848 | cf_dyn_buf_append_buf(dyn_buf, (uint8_t*)&num_pids, sizeof(num_pids)); |
1849 | |
1850 | // Append each pid. |
1851 | for (int i = 0; i < num_pids; i++) { |
1852 | uint16_t* pid = cf_vector_getp(pid_vector, i); |
1853 | cf_dyn_buf_append_buf(dyn_buf, (uint8_t*)pid, sizeof(*pid)); |
1854 | } |
1855 | |
1856 | return CF_SHASH_OK; |
1857 | } |
1858 | |
1859 | /** |
1860 | * Append namespace payload, in as_exchange_namespace_payload format, for a |
1861 | * namespace to the dynamic buffer. |
1862 | * |
1863 | * @param ns the namespace. |
1864 | * @param dyn_buf the dynamic buffer. |
1865 | */ |
1866 | static void |
1867 | exchange_data_namespace_payload_add(as_namespace* ns, cf_dyn_buf* dyn_buf) |
1868 | { |
1869 | // A hash from each unique non null vinfo to a vector of partition ids |
1870 | // having the vinfo. |
1871 | cf_shash* ns_hash = cf_shash_create(exchange_vinfo_shash, |
1872 | sizeof(as_partition_version), sizeof(cf_vector*), |
1873 | AS_EXCHANGE_UNIQUE_VINFO_MAX_SIZE_SOFT, 0); |
1874 | |
1875 | as_partition* partitions = ns->partitions; |
1876 | |
1877 | // Populate the hash with one entry for each vinfo |
1878 | for (int i = 0; i < AS_PARTITIONS; i++) { |
1879 | as_partition_version* current_vinfo = &partitions[i].version; |
1880 | exchange_namespace_hash_pid_add(ns_hash, current_vinfo, i); |
1881 | } |
1882 | |
1883 | // We are ready to populate the dyn buffer with this ns's data. |
1884 | DEBUG("namespace %s has %d unique vinfos" , ns->name, |
1885 | cf_shash_get_size(ns_hash)); |
1886 | |
1887 | // Append the vinfo count. |
1888 | uint32_t num_vinfos = cf_shash_get_size(ns_hash); |
1889 | cf_dyn_buf_append_buf(dyn_buf, (uint8_t*)&num_vinfos, sizeof(num_vinfos)); |
1890 | |
1891 | // Append vinfos and partitions. |
1892 | cf_shash_reduce(ns_hash, exchange_namespace_hash_serialize_reduce, dyn_buf); |
1893 | |
1894 | // Destroy the intermediate hash and the pid vectors. |
1895 | cf_shash_reduce(ns_hash, exchange_namespace_hash_destroy_reduce, NULL); |
1896 | |
1897 | cf_shash_destroy(ns_hash); |
1898 | } |
1899 | |
1900 | /** |
1901 | * Prepare the exchanged data payloads for current exchange round. |
1902 | */ |
1903 | static void |
1904 | exchange_data_payload_prepare() |
1905 | { |
1906 | EXCHANGE_LOCK(); |
1907 | |
1908 | // Block / abort migrations and freeze the partition version infos. |
1909 | as_partition_balance_disallow_migrations(); |
1910 | as_partition_balance_synchronize_migrations(); |
1911 | |
1912 | // Ensure ns->smd_roster is synchronized exchanged partition versions. |
1913 | pthread_mutex_lock(&g_exchanged_info_lock); |
1914 | |
1915 | for (uint32_t ns_ix = 0; ns_ix < g_config.n_namespaces; ns_ix++) { |
1916 | as_namespace* ns = g_config.namespaces[ns_ix]; |
1917 | |
1918 | // May change flags on partition versions we're about to exchange. |
1919 | as_partition_balance_protect_roster_set(ns); |
1920 | |
1921 | // Append payload for each namespace. |
1922 | |
1923 | // TODO - add API to reset dynbuf? |
1924 | g_exchange.self_data_dyn_buf[ns_ix].used_sz = 0; |
1925 | |
1926 | exchange_data_namespace_payload_add(ns, |
1927 | &g_exchange.self_data_dyn_buf[ns_ix]); |
1928 | } |
1929 | |
1930 | g_exchange.data_msg = exchange_msg_get(AS_EXCHANGE_MSG_TYPE_DATA); |
1931 | exchange_msg_data_payload_set(g_exchange.data_msg); |
1932 | |
1933 | pthread_mutex_unlock(&g_exchanged_info_lock); |
1934 | |
1935 | EXCHANGE_UNLOCK(); |
1936 | } |
1937 | |
1938 | /** |
1939 | * Indicates if the per-namespace fields in an incoming data message are valid. |
1940 | * |
1941 | * @return number of namespaces. |
1942 | */ |
1943 | static uint32_t |
1944 | exchange_data_msg_get_num_namespaces(as_exchange_event* msg_event) |
1945 | { |
1946 | uint32_t num_namespaces_sent = 0; |
1947 | uint32_t num_namespace_elements_sent = 0; |
1948 | |
1949 | if (!msg_msgpack_list_get_count(msg_event->msg, |
1950 | AS_EXCHANGE_MSG_NAMESPACES, &num_namespaces_sent) |
1951 | || num_namespaces_sent > AS_NAMESPACE_SZ) { |
1952 | WARNING("received invalid namespaces from node %" PRIx64, |
1953 | msg_event->msg_source); |
1954 | return 0; |
1955 | } |
1956 | |
1957 | if (!msg_msgpack_list_get_count(msg_event->msg, |
1958 | AS_EXCHANGE_MSG_NS_PARTITION_VERSIONS, &num_namespace_elements_sent) |
1959 | || num_namespaces_sent != num_namespace_elements_sent) { |
1960 | WARNING("received invalid partition versions from node %" PRIx64, |
1961 | msg_event->msg_source); |
1962 | return 0; |
1963 | } |
1964 | |
1965 | if (!msg_msgpack_list_get_count(msg_event->msg, |
1966 | AS_EXCHANGE_MSG_NS_RACK_IDS, &num_namespace_elements_sent) |
1967 | || num_namespaces_sent != num_namespace_elements_sent) { |
1968 | WARNING("received invalid cluster groups from node %" PRIx64, |
1969 | msg_event->msg_source); |
1970 | return 0; |
1971 | } |
1972 | |
1973 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_ROSTER_GENERATIONS) |
1974 | && (!msg_msgpack_list_get_count(msg_event->msg, |
1975 | AS_EXCHANGE_MSG_NS_ROSTER_GENERATIONS, |
1976 | &num_namespace_elements_sent) |
1977 | || num_namespaces_sent != num_namespace_elements_sent)) { |
1978 | WARNING("received invalid roster generations from node %" PRIx64, |
1979 | msg_event->msg_source); |
1980 | return 0; |
1981 | } |
1982 | |
1983 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_ROSTERS) |
1984 | && (!msg_msgpack_list_get_count(msg_event->msg, |
1985 | AS_EXCHANGE_MSG_NS_ROSTERS, |
1986 | &num_namespace_elements_sent) |
1987 | || num_namespaces_sent != num_namespace_elements_sent)) { |
1988 | WARNING("received invalid rosters from node %" PRIx64, |
1989 | msg_event->msg_source); |
1990 | return 0; |
1991 | } |
1992 | |
1993 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_ROSTERS_RACK_IDS) |
1994 | && (!msg_msgpack_list_get_count(msg_event->msg, |
1995 | AS_EXCHANGE_MSG_NS_ROSTERS_RACK_IDS, |
1996 | &num_namespace_elements_sent) |
1997 | || num_namespaces_sent != num_namespace_elements_sent)) { |
1998 | WARNING("received invalid rosters-rack-ids from node %" PRIx64, |
1999 | msg_event->msg_source); |
2000 | return 0; |
2001 | } |
2002 | |
2003 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_EVENTUAL_REGIMES) |
2004 | && (!msg_msgpack_list_get_count(msg_event->msg, |
2005 | AS_EXCHANGE_MSG_NS_EVENTUAL_REGIMES, |
2006 | &num_namespace_elements_sent) |
2007 | || num_namespaces_sent != num_namespace_elements_sent)) { |
2008 | WARNING("received invalid eventual regimes from node %" PRIx64, |
2009 | msg_event->msg_source); |
2010 | return 0; |
2011 | } |
2012 | |
2013 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_REBALANCE_REGIMES) |
2014 | && (!msg_msgpack_list_get_count(msg_event->msg, |
2015 | AS_EXCHANGE_MSG_NS_REBALANCE_REGIMES, |
2016 | &num_namespace_elements_sent) |
2017 | || num_namespaces_sent != num_namespace_elements_sent)) { |
2018 | WARNING("received invalid rebalance regimes from node %" PRIx64, |
2019 | msg_event->msg_source); |
2020 | return 0; |
2021 | } |
2022 | |
2023 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_REBALANCE_FLAGS) |
2024 | && (!msg_msgpack_list_get_count(msg_event->msg, |
2025 | AS_EXCHANGE_MSG_NS_REBALANCE_FLAGS, |
2026 | &num_namespace_elements_sent) |
2027 | || num_namespaces_sent != num_namespace_elements_sent)) { |
2028 | WARNING("received invalid rebalance flags from node %" PRIx64, |
2029 | msg_event->msg_source); |
2030 | return 0; |
2031 | } |
2032 | |
2033 | return num_namespaces_sent; |
2034 | } |
2035 | |
2036 | /** |
2037 | * Basic validation for incoming namespace payload. |
2038 | * Validates that |
2039 | * 1. Number of vinfos < AS_PARTITIONS. |
2040 | * 2. Each partition is between 0 and AS_PARTITIONS. |
2041 | * 3. Namespaces payload does not exceed payload_end_ptr. |
2042 | * |
2043 | * @param ns_payload pointer to start of the namespace payload. |
2044 | * @param ns_payload_size the size of the input namespace payload. |
2045 | * @return true if this is a valid payload. |
2046 | */ |
2047 | static bool |
2048 | exchange_namespace_payload_is_valid(as_exchange_ns_vinfos_payload* ns_payload, |
2049 | uint32_t ns_payload_size) |
2050 | { |
2051 | // Pointer past the last byte in the payload. |
2052 | uint8_t* payload_end_ptr = (uint8_t*)ns_payload + ns_payload_size; |
2053 | |
2054 | if ((uint8_t*)ns_payload->vinfos > payload_end_ptr) { |
2055 | return false; |
2056 | } |
2057 | |
2058 | if (ns_payload->num_vinfos > AS_PARTITIONS) { |
2059 | return false; |
2060 | } |
2061 | |
2062 | uint8_t* read_ptr = (uint8_t*)ns_payload->vinfos; |
2063 | |
2064 | for (uint32_t i = 0; i < ns_payload->num_vinfos; i++) { |
2065 | if (read_ptr >= payload_end_ptr) { |
2066 | return false; |
2067 | } |
2068 | |
2069 | as_exchange_vinfo_payload* vinfo_payload = |
2070 | (as_exchange_vinfo_payload*)read_ptr; |
2071 | |
2072 | if ((uint8_t*)vinfo_payload->pids > payload_end_ptr) { |
2073 | return false; |
2074 | } |
2075 | |
2076 | if (vinfo_payload->num_pids > AS_PARTITIONS) { |
2077 | return false; |
2078 | } |
2079 | |
2080 | size_t pids_size = vinfo_payload->num_pids * sizeof(uint16_t); |
2081 | |
2082 | if ((uint8_t*)vinfo_payload->pids + pids_size > payload_end_ptr) { |
2083 | return false; |
2084 | } |
2085 | |
2086 | for (uint32_t j = 0; j < vinfo_payload->num_pids; j++) { |
2087 | if (vinfo_payload->pids[j] >= AS_PARTITIONS) { |
2088 | return false; |
2089 | } |
2090 | } |
2091 | |
2092 | read_ptr += sizeof(as_exchange_vinfo_payload) + pids_size; |
2093 | } |
2094 | |
2095 | if (read_ptr != payload_end_ptr) { |
2096 | // There are unaccounted for extra bytes in the payload. |
2097 | return false; |
2098 | } |
2099 | |
2100 | return true; |
2101 | } |
2102 | |
2103 | /* |
2104 | * ---------------------------------------------------------------------------- |
2105 | * Common across all states |
2106 | * ---------------------------------------------------------------------------- |
2107 | */ |
2108 | |
2109 | /** |
2110 | * Update committed exchange cluster data. |
2111 | * @param cluster_key the new cluster key. |
2112 | * @param succession the new succession. Can be NULL, for orphan state. |
2113 | */ |
2114 | static void |
2115 | exchange_commited_cluster_update(as_cluster_key cluster_key, |
2116 | cf_vector* succession) |
2117 | { |
2118 | EXCHANGE_COMMITTED_CLUSTER_WLOCK(); |
2119 | g_exchange.committed_cluster_key = cluster_key; |
2120 | vector_clear(&g_exchange.committed_succession_list); |
2121 | |
2122 | if (succession && cf_vector_size(succession) > 0) { |
2123 | vector_copy(&g_exchange.committed_succession_list, |
2124 | &g_exchange.succession_list); |
2125 | g_exchange.committed_cluster_size = cf_vector_size(succession); |
2126 | cf_vector_get(succession, 0, &g_exchange.committed_principal); |
2127 | } |
2128 | else { |
2129 | g_exchange.committed_cluster_size = 0; |
2130 | g_exchange.committed_principal = 0; |
2131 | } |
2132 | |
2133 | g_exchange.committed_cluster_generation++; |
2134 | EXCHANGE_COMMITTED_CLUSTER_UNLOCK(); |
2135 | } |
2136 | |
2137 | /** |
2138 | * Indicates if self node is the cluster principal. |
2139 | */ |
2140 | static bool |
2141 | exchange_self_is_principal() |
2142 | { |
2143 | EXCHANGE_LOCK(); |
2144 | bool is_principal = (g_config.self_node == g_exchange.principal); |
2145 | EXCHANGE_UNLOCK(); |
2146 | return is_principal; |
2147 | } |
2148 | |
2149 | /** |
2150 | * Dump exchange state. |
2151 | */ |
2152 | static void |
2153 | exchange_dump(cf_fault_severity severity, bool verbose) |
2154 | { |
2155 | EXCHANGE_LOCK(); |
2156 | cf_vector* node_vector = cf_vector_stack_create(cf_node); |
2157 | |
2158 | char* state_str = "" ; |
2159 | switch (g_exchange.state) { |
2160 | case AS_EXCHANGE_STATE_REST: |
2161 | state_str = "rest" ; |
2162 | break; |
2163 | case AS_EXCHANGE_STATE_EXCHANGING: |
2164 | state_str = "exchanging" ; |
2165 | break; |
2166 | case AS_EXCHANGE_STATE_READY_TO_COMMIT: |
2167 | state_str = "ready to commit" ; |
2168 | break; |
2169 | case AS_EXCHANGE_STATE_ORPHANED: |
2170 | state_str = "orphaned" ; |
2171 | break; |
2172 | } |
2173 | |
2174 | LOG(severity, "EXG: state: %s" , state_str); |
2175 | |
2176 | if (g_exchange.state == AS_EXCHANGE_STATE_ORPHANED) { |
2177 | LOG(severity, "EXG: client transactions blocked: %s" , |
2178 | g_exchange.orphan_state_are_transactions_blocked ? |
2179 | "true" : "false" ); |
2180 | LOG(severity, "EXG: orphan since: %" PRIu64"(millis)" , |
2181 | cf_getms() - g_exchange.orphan_state_start_time); |
2182 | } |
2183 | else { |
2184 | LOG(severity, "EXG: cluster key: %" PRIx64, g_exchange.cluster_key); |
2185 | as_clustering_log_cf_node_vector(severity, AS_EXCHANGE, |
2186 | "EXG: succession:" , &g_exchange.succession_list); |
2187 | |
2188 | if (verbose) { |
2189 | vector_clear(node_vector); |
2190 | exchange_nodes_find_send_unacked(node_vector); |
2191 | as_clustering_log_cf_node_vector(severity, AS_EXCHANGE, |
2192 | "EXG: send pending:" , node_vector); |
2193 | |
2194 | vector_clear(node_vector); |
2195 | exchange_nodes_find_not_received(node_vector); |
2196 | as_clustering_log_cf_node_vector(severity, AS_EXCHANGE, |
2197 | "EXG: receive pending:" , node_vector); |
2198 | |
2199 | if (exchange_self_is_principal()) { |
2200 | vector_clear(node_vector); |
2201 | exchange_nodes_find_not_ready_to_commit(node_vector); |
2202 | as_clustering_log_cf_node_vector(severity, AS_EXCHANGE, |
2203 | "EXG: ready to commit pending:" , node_vector); |
2204 | } |
2205 | } |
2206 | } |
2207 | |
2208 | cf_vector_destroy(node_vector); |
2209 | EXCHANGE_UNLOCK(); |
2210 | } |
2211 | |
2212 | /** |
2213 | * Reset state for new round of exchange, while reusing as mush heap allocated |
2214 | * space for exchanged data. |
2215 | * @param new_succession_list new succession list. Can be NULL for orphaned |
2216 | * state. |
2217 | * @param new_cluster_key 0 for orphaned state. |
2218 | */ |
2219 | static void |
2220 | exchange_reset_for_new_round(cf_vector* new_succession_list, |
2221 | as_cluster_key new_cluster_key) |
2222 | { |
2223 | EXCHANGE_LOCK(); |
2224 | vector_clear(&g_exchange.succession_list); |
2225 | g_exchange.principal = 0; |
2226 | |
2227 | if (new_succession_list && cf_vector_size(new_succession_list) > 0) { |
2228 | vector_copy(&g_exchange.succession_list, new_succession_list); |
2229 | // Set the principal node. |
2230 | cf_vector_get(&g_exchange.succession_list, 0, &g_exchange.principal); |
2231 | } |
2232 | |
2233 | // Reset accumulated node states. |
2234 | exchange_node_states_reset(); |
2235 | |
2236 | g_exchange.cluster_key = new_cluster_key; |
2237 | |
2238 | if (g_exchange.data_msg) { |
2239 | as_fabric_msg_put(g_exchange.data_msg); |
2240 | g_exchange.data_msg = NULL; |
2241 | } |
2242 | |
2243 | EXCHANGE_UNLOCK(); |
2244 | } |
2245 | |
2246 | /** |
2247 | * Commit exchange state to reflect self node being an orphan. |
2248 | */ |
2249 | static void |
2250 | exchange_orphan_commit() |
2251 | { |
2252 | EXCHANGE_LOCK(); |
2253 | exchange_commited_cluster_update(0, NULL); |
2254 | WARNING("blocking client transactions in orphan state!" ); |
2255 | as_partition_balance_revert_to_orphan(); |
2256 | g_exchange.orphan_state_are_transactions_blocked = true; |
2257 | EXCHANGE_UNLOCK(); |
2258 | } |
2259 | |
2260 | /** |
2261 | * Receive an orphaned event and abort current round. |
2262 | */ |
2263 | static void |
2264 | exchange_orphaned_handle(as_clustering_event* orphaned_event) |
2265 | { |
2266 | DEBUG("got orphaned event" ); |
2267 | |
2268 | EXCHANGE_LOCK(); |
2269 | |
2270 | if (g_exchange.state != AS_EXCHANGE_STATE_REST |
2271 | && g_exchange.state != AS_EXCHANGE_STATE_ORPHANED) { |
2272 | INFO("aborting partition exchange with cluster key %" PRIx64, |
2273 | g_exchange.cluster_key); |
2274 | } |
2275 | |
2276 | g_exchange.state = AS_EXCHANGE_STATE_ORPHANED; |
2277 | exchange_reset_for_new_round(NULL, 0); |
2278 | |
2279 | // Stop ongoing migrations if any. |
2280 | as_partition_balance_disallow_migrations(); |
2281 | as_partition_balance_synchronize_migrations(); |
2282 | |
2283 | // Update the time this node got into orphan state. |
2284 | g_exchange.orphan_state_start_time = cf_getms(); |
2285 | |
2286 | // Potentially temporary orphan state. We will timeout and commit orphan |
2287 | // state if this persists for long. |
2288 | g_exchange.orphan_state_are_transactions_blocked = false; |
2289 | |
2290 | EXCHANGE_UNLOCK(); |
2291 | } |
2292 | |
2293 | /** |
2294 | * Receive a cluster change event and start a new data exchange round. |
2295 | */ |
2296 | static void |
2297 | exchange_cluster_change_handle(as_clustering_event* clustering_event) |
2298 | { |
2299 | EXCHANGE_LOCK(); |
2300 | |
2301 | DEBUG("got cluster change event" ); |
2302 | |
2303 | if (g_exchange.state != AS_EXCHANGE_STATE_REST |
2304 | && g_exchange.state != AS_EXCHANGE_STATE_ORPHANED) { |
2305 | INFO("aborting partition exchange with cluster key %" PRIx64, |
2306 | g_exchange.cluster_key); |
2307 | } |
2308 | |
2309 | exchange_reset_for_new_round(clustering_event->succession_list, |
2310 | clustering_event->cluster_key); |
2311 | |
2312 | g_exchange.state = AS_EXCHANGE_STATE_EXCHANGING; |
2313 | |
2314 | INFO("data exchange started with cluster key %" PRIx64, |
2315 | g_exchange.cluster_key); |
2316 | |
2317 | // Prepare the data payload. |
2318 | exchange_data_payload_prepare(); |
2319 | |
2320 | EXCHANGE_UNLOCK(); |
2321 | |
2322 | exchange_data_msg_send_pending_ack(); |
2323 | } |
2324 | |
2325 | /** |
2326 | * Handle a cluster change event. |
2327 | * @param cluster_change_event the cluster change event. |
2328 | */ |
2329 | static void |
2330 | exchange_clustering_event_handle(as_exchange_event* exchange_clustering_event) |
2331 | { |
2332 | as_clustering_event* clustering_event = |
2333 | exchange_clustering_event->clustering_event; |
2334 | |
2335 | switch (clustering_event->type) { |
2336 | case AS_CLUSTERING_ORPHANED: |
2337 | exchange_orphaned_handle(clustering_event); |
2338 | break; |
2339 | case AS_CLUSTERING_CLUSTER_CHANGED: |
2340 | exchange_cluster_change_handle(clustering_event); |
2341 | break; |
2342 | } |
2343 | } |
2344 | |
2345 | /* |
2346 | * ---------------------------------------------------------------------------- |
2347 | * Orphan state event handling |
2348 | * ---------------------------------------------------------------------------- |
2349 | */ |
2350 | |
2351 | /** |
2352 | * The wait time in orphan state after which client transactions and transaction |
2353 | * related interactions (e.g. valid partition map publishing) should be blocked. |
2354 | */ |
2355 | static uint32_t |
2356 | exchange_orphan_transaction_block_timeout() |
2357 | { |
2358 | return (uint32_t)as_clustering_quantum_interval() |
2359 | * AS_EXCHANGE_REVERT_ORPHAN_INTERVALS; |
2360 | } |
2361 | |
2362 | /** |
2363 | * Handle the timer event and if we have been an orphan for too long, block |
2364 | * client transactions. |
2365 | */ |
2366 | static void |
2367 | exchange_orphan_timer_event_handle() |
2368 | { |
2369 | uint32_t timeout = exchange_orphan_transaction_block_timeout(); |
2370 | EXCHANGE_LOCK(); |
2371 | if (!g_exchange.orphan_state_are_transactions_blocked |
2372 | && g_exchange.orphan_state_start_time + timeout < cf_getms()) { |
2373 | exchange_orphan_commit(); |
2374 | } |
2375 | EXCHANGE_UNLOCK(); |
2376 | } |
2377 | |
2378 | /** |
2379 | * Event processing in the orphan state. |
2380 | */ |
2381 | static void |
2382 | exchange_orphan_event_handle(as_exchange_event* event) |
2383 | { |
2384 | switch (event->type) { |
2385 | case AS_EXCHANGE_EVENT_CLUSTER_CHANGE: |
2386 | exchange_clustering_event_handle(event); |
2387 | break; |
2388 | case AS_EXCHANGE_EVENT_TIMER: |
2389 | exchange_orphan_timer_event_handle(); |
2390 | break; |
2391 | default: |
2392 | break; |
2393 | } |
2394 | } |
2395 | |
2396 | /* |
2397 | * ---------------------------------------------------------------------------- |
2398 | * Rest state event handling |
2399 | * ---------------------------------------------------------------------------- |
2400 | */ |
2401 | |
2402 | /** |
2403 | * Process a message event when in rest state. |
2404 | */ |
2405 | static void |
2406 | exchange_rest_msg_event_handle(as_exchange_event* msg_event) |
2407 | { |
2408 | EXCHANGE_LOCK(); |
2409 | |
2410 | if (!exchange_msg_is_sane(msg_event->msg_source, msg_event->msg)) { |
2411 | goto Exit; |
2412 | } |
2413 | |
2414 | as_exchange_msg_type msg_type; |
2415 | exchange_msg_type_get(msg_event->msg, &msg_type); |
2416 | |
2417 | if (exchange_self_is_principal() |
2418 | && msg_type == AS_EXCHANGE_MSG_TYPE_READY_TO_COMMIT) { |
2419 | // The commit message did not make it to the source node, hence it send |
2420 | // us the ready to commit message. Resend the commit message. |
2421 | DEBUG("received a ready to commit message from %" PRIx64, |
2422 | msg_event->msg_source); |
2423 | exchange_commit_msg_send(msg_event->msg_source); |
2424 | } |
2425 | else { |
2426 | DEBUG( |
2427 | "rest state received unexpected mesage of type %d from node %" PRIx64, |
2428 | msg_type, msg_event->msg_source); |
2429 | |
2430 | } |
2431 | |
2432 | Exit: |
2433 | |
2434 | EXCHANGE_UNLOCK(); |
2435 | } |
2436 | |
2437 | /** |
2438 | * Event processing in the rest state. |
2439 | */ |
2440 | static void |
2441 | exchange_rest_event_handle(as_exchange_event* event) |
2442 | { |
2443 | switch (event->type) { |
2444 | case AS_EXCHANGE_EVENT_CLUSTER_CHANGE: |
2445 | exchange_clustering_event_handle(event); |
2446 | break; |
2447 | case AS_EXCHANGE_EVENT_MSG: |
2448 | exchange_rest_msg_event_handle(event); |
2449 | break; |
2450 | default: |
2451 | break; |
2452 | } |
2453 | } |
2454 | |
2455 | /* |
2456 | * ---------------------------------------------------------------------------- |
2457 | * Exchanging state event handling |
2458 | * ---------------------------------------------------------------------------- |
2459 | */ |
2460 | |
2461 | /** |
2462 | * Commit namespace payload for a node. |
2463 | * Assumes the namespace vinfo and succession list have been zero set before. |
2464 | */ |
2465 | static void |
2466 | exchange_namespace_payload_pre_commit_for_node(cf_node node, |
2467 | as_exchange_node_namespace_data* namespace_data) |
2468 | { |
2469 | as_namespace* ns = namespace_data->local_namespace; |
2470 | |
2471 | uint32_t sl_ix = ns->cluster_size++; |
2472 | |
2473 | ns->succession[sl_ix] = node; |
2474 | |
2475 | as_exchange_ns_vinfos_payload* ns_payload = |
2476 | namespace_data->partition_versions; |
2477 | uint8_t* read_ptr = (uint8_t*)ns_payload->vinfos; |
2478 | |
2479 | for (int i = 0; i < ns_payload->num_vinfos; i++) { |
2480 | as_exchange_vinfo_payload* vinfo_payload = |
2481 | (as_exchange_vinfo_payload*)read_ptr; |
2482 | |
2483 | for (int j = 0; j < vinfo_payload->num_pids; j++) { |
2484 | memcpy(&ns->cluster_versions[sl_ix][vinfo_payload->pids[j]], |
2485 | &vinfo_payload->vinfo, sizeof(vinfo_payload->vinfo)); |
2486 | } |
2487 | |
2488 | read_ptr += sizeof(as_exchange_vinfo_payload) |
2489 | + vinfo_payload->num_pids * sizeof(uint16_t); |
2490 | } |
2491 | |
2492 | ns->rack_ids[sl_ix] = namespace_data->rack_id; |
2493 | |
2494 | if (namespace_data->roster_generation > ns->roster_generation) { |
2495 | ns->roster_generation = namespace_data->roster_generation; |
2496 | ns->roster_count = namespace_data->roster_count; |
2497 | |
2498 | memcpy(ns->roster, namespace_data->roster, |
2499 | ns->roster_count * sizeof(cf_node)); |
2500 | |
2501 | if (namespace_data->roster_rack_ids) { |
2502 | memcpy(ns->roster_rack_ids, namespace_data->roster_rack_ids, |
2503 | ns->roster_count * sizeof(uint32_t)); |
2504 | } |
2505 | else { |
2506 | memset(ns->roster_rack_ids, 0, ns->roster_count * sizeof(uint32_t)); |
2507 | } |
2508 | } |
2509 | |
2510 | if (ns->eventual_regime != 0 && |
2511 | namespace_data->eventual_regime > ns->eventual_regime) { |
2512 | ns->eventual_regime = namespace_data->eventual_regime; |
2513 | } |
2514 | |
2515 | ns->rebalance_regimes[sl_ix] = namespace_data->rebalance_regime; |
2516 | |
2517 | // Prefer uniform balance only if all nodes prefer it. |
2518 | if ((namespace_data->rebalance_flags & |
2519 | AS_EXCHANGE_REBALANCE_FLAG_UNIFORM) == 0) { |
2520 | ns->prefer_uniform_balance = false; |
2521 | } |
2522 | |
2523 | bool is_node_quiesced = (namespace_data->rebalance_flags & |
2524 | AS_EXCHANGE_REBALANCE_FLAG_QUIESCE) != 0; |
2525 | |
2526 | if (node == g_config.self_node) { |
2527 | ns->is_quiesced = is_node_quiesced; |
2528 | } |
2529 | |
2530 | ns->quiesced[sl_ix] = is_node_quiesced; |
2531 | } |
2532 | |
2533 | /** |
2534 | * Commit exchange data for a given node. |
2535 | */ |
2536 | static void |
2537 | exchange_data_pre_commit_for_node(cf_node node, uint32_t ix) |
2538 | { |
2539 | EXCHANGE_LOCK(); |
2540 | as_exchange_node_state node_state; |
2541 | exchange_node_state_get_safe(node, &node_state); |
2542 | |
2543 | g_exchange.compatibility_ids[ix] = node_state.data->compatibility_id; |
2544 | |
2545 | if (node_state.data->compatibility_id < g_exchange.min_compatibility_id) { |
2546 | g_exchange.min_compatibility_id = node_state.data->compatibility_id; |
2547 | } |
2548 | |
2549 | for (uint32_t i = 0; i < node_state.data->num_namespaces; i++) { |
2550 | exchange_namespace_payload_pre_commit_for_node(node, |
2551 | &node_state.data->namespace_data[i]); |
2552 | } |
2553 | |
2554 | EXCHANGE_UNLOCK(); |
2555 | } |
2556 | |
2557 | /** |
2558 | * Check that there's not a mixture of AP and CP nodes in any namespace. |
2559 | */ |
2560 | static bool |
2561 | exchange_data_pre_commit_ap_cp_check() |
2562 | { |
2563 | for (uint32_t i = 0; i < g_config.n_namespaces; i++) { |
2564 | as_namespace* ns = g_config.namespaces[i]; |
2565 | |
2566 | cf_node ap_node = (cf_node)0; |
2567 | cf_node cp_node = (cf_node)0; |
2568 | |
2569 | for (uint32_t n = 0; n < ns->cluster_size; n++) { |
2570 | if (ns->rebalance_regimes[n] == 0) { |
2571 | ap_node = ns->succession[n]; |
2572 | } |
2573 | else { |
2574 | cp_node = ns->succession[n]; |
2575 | } |
2576 | } |
2577 | |
2578 | if (ap_node != (cf_node)0 && cp_node != (cf_node)0) { |
2579 | WARNING("{%s} has mixture of AP and SC nodes - for example %lx is AP and %lx is SC" , |
2580 | ns->name, ap_node, cp_node); |
2581 | return false; |
2582 | } |
2583 | } |
2584 | |
2585 | return true; |
2586 | } |
2587 | |
2588 | /** |
2589 | * Pre commit namespace data anticipating a successful commit from the |
2590 | * principal. This pre commit is to ensure regime advances in cp mode to cover |
2591 | * the case where the principal commits exchange data but the commit to a |
2592 | * non-principal is lost. |
2593 | */ |
2594 | static bool |
2595 | exchange_exchanging_pre_commit() |
2596 | { |
2597 | EXCHANGE_LOCK(); |
2598 | pthread_mutex_lock(&g_exchanged_info_lock); |
2599 | |
2600 | memset(g_exchange.compatibility_ids, 0, |
2601 | sizeof(g_exchange.compatibility_ids)); |
2602 | g_exchange.min_compatibility_id = UINT32_MAX; |
2603 | |
2604 | // Reset exchange data for all namespaces. |
2605 | for (int i = 0; i < g_config.n_namespaces; i++) { |
2606 | as_namespace* ns = g_config.namespaces[i]; |
2607 | memset(ns->succession, 0, sizeof(ns->succession)); |
2608 | |
2609 | // Assuming zero to represent "null" partition. |
2610 | memset(ns->cluster_versions, 0, sizeof(ns->cluster_versions)); |
2611 | |
2612 | memset(ns->rack_ids, 0, sizeof(ns->rack_ids)); |
2613 | |
2614 | ns->is_quiesced = false; |
2615 | memset(ns->quiesced, 0, sizeof(ns->quiesced)); |
2616 | |
2617 | ns->roster_generation = 0; |
2618 | ns->roster_count = 0; |
2619 | memset(ns->roster, 0, sizeof(ns->roster)); |
2620 | memset(ns->roster_rack_ids, 0, sizeof(ns->roster_rack_ids)); |
2621 | |
2622 | // Note - not clearing ns->eventual_regime - prior non-0 value means CP. |
2623 | // Note - not clearing ns->rebalance_regime - it's not set here. |
2624 | memset(ns->rebalance_regimes, 0, sizeof(ns->rebalance_regimes)); |
2625 | |
2626 | ns->cluster_size = 0; |
2627 | |
2628 | // Any node that does not prefer uniform balance will set this false. |
2629 | ns->prefer_uniform_balance = true; |
2630 | } |
2631 | |
2632 | // Fill the namespace partition version info in succession list order. |
2633 | int num_nodes = cf_vector_size(&g_exchange.succession_list); |
2634 | for (int i = 0; i < num_nodes; i++) { |
2635 | cf_node node; |
2636 | cf_vector_get(&g_exchange.succession_list, i, &node); |
2637 | exchange_data_pre_commit_for_node(node, i); |
2638 | } |
2639 | |
2640 | // Collected all exchanged data - do final configuration consistency checks. |
2641 | if (!exchange_data_pre_commit_ap_cp_check()) { |
2642 | WARNING("abandoned exchange - fix configuration conflict" ); |
2643 | pthread_mutex_unlock(&g_exchanged_info_lock); |
2644 | EXCHANGE_UNLOCK(); |
2645 | return false; |
2646 | } |
2647 | |
2648 | for (int i = 0; i < g_config.n_namespaces; i++) { |
2649 | as_namespace* ns = g_config.namespaces[i]; |
2650 | |
2651 | if (ns->eventual_regime != 0) { |
2652 | ns->eventual_regime += 2; |
2653 | |
2654 | as_storage_save_regime(ns); |
2655 | |
2656 | INFO("{%s} eventual-regime %u ready" , ns->name, |
2657 | ns->eventual_regime); |
2658 | } |
2659 | } |
2660 | |
2661 | pthread_mutex_unlock(&g_exchanged_info_lock); |
2662 | EXCHANGE_UNLOCK(); |
2663 | |
2664 | return true; |
2665 | } |
2666 | |
2667 | /** |
2668 | * Check to see if all exchange data is sent and received. If so switch to |
2669 | * ready_to_commit state. |
2670 | */ |
2671 | static void |
2672 | exchange_exchanging_check_switch_ready_to_commit() |
2673 | { |
2674 | EXCHANGE_LOCK(); |
2675 | |
2676 | cf_vector* node_vector = cf_vector_stack_create(cf_node); |
2677 | |
2678 | if (g_exchange.state == AS_EXCHANGE_STATE_REST |
2679 | || g_exchange.cluster_key == 0) { |
2680 | goto Exit; |
2681 | } |
2682 | |
2683 | exchange_nodes_find_send_unacked(node_vector); |
2684 | if (cf_vector_size(node_vector) > 0) { |
2685 | // We still have unacked exchange send messages. |
2686 | goto Exit; |
2687 | } |
2688 | |
2689 | vector_clear(node_vector); |
2690 | exchange_nodes_find_not_received(node_vector); |
2691 | if (cf_vector_size(node_vector) > 0) { |
2692 | // We still haven't received exchange messages from all nodes in the |
2693 | // succession list. |
2694 | goto Exit; |
2695 | } |
2696 | |
2697 | if (!exchange_exchanging_pre_commit()) { |
2698 | // Pre-commit failed. We are not ready to commit. |
2699 | goto Exit; |
2700 | } |
2701 | |
2702 | g_exchange.state = AS_EXCHANGE_STATE_READY_TO_COMMIT; |
2703 | |
2704 | DEBUG("ready to commit exchange data for cluster key %" PRIx64, |
2705 | g_exchange.cluster_key); |
2706 | |
2707 | Exit: |
2708 | cf_vector_destroy(node_vector); |
2709 | |
2710 | if (g_exchange.state == AS_EXCHANGE_STATE_READY_TO_COMMIT) { |
2711 | exchange_ready_to_commit_msg_send(); |
2712 | } |
2713 | |
2714 | EXCHANGE_UNLOCK(); |
2715 | } |
2716 | |
2717 | /** |
2718 | * Handle incoming data message. |
2719 | * |
2720 | * Assumes the message has been checked for sanity. |
2721 | */ |
2722 | static void |
2723 | exchange_exchanging_data_msg_handle(as_exchange_event* msg_event) |
2724 | { |
2725 | EXCHANGE_LOCK(); |
2726 | |
2727 | DEBUG("received exchange data from node %" PRIx64, msg_event->msg_source); |
2728 | |
2729 | as_exchange_node_state node_state; |
2730 | exchange_node_state_get_safe(msg_event->msg_source, &node_state); |
2731 | |
2732 | if (!node_state.received) { |
2733 | node_state.data->compatibility_id = 0; |
2734 | msg_get_uint32(msg_event->msg, AS_EXCHANGE_MSG_COMPATIBILITY_ID, |
2735 | &node_state.data->compatibility_id); |
2736 | |
2737 | uint32_t num_namespaces_sent = exchange_data_msg_get_num_namespaces( |
2738 | msg_event); |
2739 | |
2740 | if (num_namespaces_sent == 0) { |
2741 | WARNING("ignoring invalid exchange data from node %" PRIx64, |
2742 | msg_event->msg_source); |
2743 | goto Exit; |
2744 | } |
2745 | |
2746 | cf_vector_define(namespace_list, sizeof(msg_buf_ele), |
2747 | num_namespaces_sent, 0); |
2748 | cf_vector_define(partition_versions, sizeof(msg_buf_ele), |
2749 | num_namespaces_sent, 0); |
2750 | uint32_t rack_ids[num_namespaces_sent]; |
2751 | |
2752 | uint32_t roster_generations[num_namespaces_sent]; |
2753 | cf_vector_define(rosters, sizeof(msg_buf_ele), num_namespaces_sent, 0); |
2754 | cf_vector_define(rosters_rack_ids, sizeof(msg_buf_ele), |
2755 | num_namespaces_sent, 0); |
2756 | |
2757 | memset(roster_generations, 0, sizeof(roster_generations)); |
2758 | |
2759 | uint32_t eventual_regimes[num_namespaces_sent]; |
2760 | uint32_t rebalance_regimes[num_namespaces_sent]; |
2761 | uint32_t rebalance_flags[num_namespaces_sent]; |
2762 | |
2763 | memset(eventual_regimes, 0, sizeof(eventual_regimes)); |
2764 | memset(rebalance_regimes, 0, sizeof(rebalance_regimes)); |
2765 | memset(rebalance_flags, 0, sizeof(rebalance_flags)); |
2766 | |
2767 | if (!msg_msgpack_list_get_buf_array_presized(msg_event->msg, |
2768 | AS_EXCHANGE_MSG_NAMESPACES, &namespace_list)) { |
2769 | WARNING("received invalid namespaces from node %" PRIx64, |
2770 | msg_event->msg_source); |
2771 | goto Exit; |
2772 | } |
2773 | |
2774 | if (!msg_msgpack_list_get_buf_array_presized(msg_event->msg, |
2775 | AS_EXCHANGE_MSG_NS_PARTITION_VERSIONS, &partition_versions)) { |
2776 | WARNING("received invalid partition versions from node %" PRIx64, |
2777 | msg_event->msg_source); |
2778 | goto Exit; |
2779 | } |
2780 | |
2781 | uint32_t num_rack_ids = num_namespaces_sent; |
2782 | |
2783 | if (!msg_msgpack_list_get_uint32_array(msg_event->msg, |
2784 | AS_EXCHANGE_MSG_NS_RACK_IDS, rack_ids, &num_rack_ids)) { |
2785 | WARNING("received invalid cluster groups from node %" PRIx64, |
2786 | msg_event->msg_source); |
2787 | goto Exit; |
2788 | } |
2789 | |
2790 | uint32_t num_roster_generations = num_namespaces_sent; |
2791 | |
2792 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_ROSTER_GENERATIONS) |
2793 | && !msg_msgpack_list_get_uint32_array(msg_event->msg, |
2794 | AS_EXCHANGE_MSG_NS_ROSTER_GENERATIONS, |
2795 | roster_generations, &num_roster_generations)) { |
2796 | WARNING("received invalid roster generations from node %" PRIx64, |
2797 | msg_event->msg_source); |
2798 | goto Exit; |
2799 | } |
2800 | |
2801 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_ROSTERS) |
2802 | && !msg_msgpack_list_get_buf_array_presized(msg_event->msg, |
2803 | AS_EXCHANGE_MSG_NS_ROSTERS, &rosters)) { |
2804 | WARNING("received invalid rosters from node %" PRIx64, |
2805 | msg_event->msg_source); |
2806 | goto Exit; |
2807 | } |
2808 | |
2809 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_ROSTERS_RACK_IDS) |
2810 | && !msg_msgpack_list_get_buf_array_presized(msg_event->msg, |
2811 | AS_EXCHANGE_MSG_NS_ROSTERS_RACK_IDS, |
2812 | &rosters_rack_ids)) { |
2813 | WARNING("received invalid rosters-rack-ids from node %" PRIx64, |
2814 | msg_event->msg_source); |
2815 | goto Exit; |
2816 | } |
2817 | |
2818 | uint32_t num_eventual_regimes = num_namespaces_sent; |
2819 | |
2820 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_EVENTUAL_REGIMES) |
2821 | && !msg_msgpack_list_get_uint32_array(msg_event->msg, |
2822 | AS_EXCHANGE_MSG_NS_EVENTUAL_REGIMES, eventual_regimes, |
2823 | &num_eventual_regimes)) { |
2824 | WARNING("received invalid eventual regimes from node %" PRIx64, |
2825 | msg_event->msg_source); |
2826 | goto Exit; |
2827 | } |
2828 | |
2829 | uint32_t num_rebalance_regimes = num_namespaces_sent; |
2830 | |
2831 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_REBALANCE_REGIMES) |
2832 | && !msg_msgpack_list_get_uint32_array(msg_event->msg, |
2833 | AS_EXCHANGE_MSG_NS_REBALANCE_REGIMES, rebalance_regimes, |
2834 | &num_rebalance_regimes)) { |
2835 | WARNING("received invalid rebalance regimes from node %" PRIx64, |
2836 | msg_event->msg_source); |
2837 | goto Exit; |
2838 | } |
2839 | |
2840 | uint32_t num_rebalance_flags = num_namespaces_sent; |
2841 | |
2842 | if (msg_is_set(msg_event->msg, AS_EXCHANGE_MSG_NS_REBALANCE_FLAGS) |
2843 | && !msg_msgpack_list_get_uint32_array(msg_event->msg, |
2844 | AS_EXCHANGE_MSG_NS_REBALANCE_FLAGS, rebalance_flags, |
2845 | &num_rebalance_flags)) { |
2846 | WARNING("received invalid rebalance flags from node %" PRIx64, |
2847 | msg_event->msg_source); |
2848 | goto Exit; |
2849 | } |
2850 | |
2851 | node_state.data->num_namespaces = 0; |
2852 | |
2853 | for (uint32_t i = 0; i < num_namespaces_sent; i++) { |
2854 | msg_buf_ele* namespace_name_element = cf_vector_getp( |
2855 | &namespace_list, i); |
2856 | |
2857 | // Find a match for the namespace. |
2858 | as_namespace* matching_namespace = as_namespace_get_bybuf( |
2859 | namespace_name_element->ptr, namespace_name_element->sz); |
2860 | |
2861 | if (!matching_namespace) { |
2862 | continue; |
2863 | } |
2864 | |
2865 | as_exchange_node_namespace_data* namespace_data = |
2866 | &node_state.data->namespace_data[node_state.data->num_namespaces]; |
2867 | node_state.data->num_namespaces++; |
2868 | |
2869 | namespace_data->local_namespace = matching_namespace; |
2870 | namespace_data->rack_id = rack_ids[i]; |
2871 | namespace_data->roster_generation = roster_generations[i]; |
2872 | namespace_data->eventual_regime = eventual_regimes[i]; |
2873 | namespace_data->rebalance_regime = rebalance_regimes[i]; |
2874 | namespace_data->rebalance_flags = rebalance_flags[i]; |
2875 | |
2876 | // Copy partition versions. |
2877 | msg_buf_ele* partition_versions_element = cf_vector_getp( |
2878 | &partition_versions, i); |
2879 | |
2880 | if (!exchange_namespace_payload_is_valid( |
2881 | (as_exchange_ns_vinfos_payload*)partition_versions_element->ptr, |
2882 | partition_versions_element->sz)) { |
2883 | WARNING( |
2884 | "received invalid partition versions for namespace %s from node %" PRIx64, |
2885 | matching_namespace->name, msg_event->msg_source); |
2886 | goto Exit; |
2887 | } |
2888 | |
2889 | namespace_data->partition_versions = cf_realloc( |
2890 | namespace_data->partition_versions, |
2891 | partition_versions_element->sz); |
2892 | |
2893 | memcpy(namespace_data->partition_versions, |
2894 | partition_versions_element->ptr, |
2895 | partition_versions_element->sz); |
2896 | |
2897 | // Copy rosters. |
2898 | // TODO - make this piece a utility function? |
2899 | if (namespace_data->roster_generation == 0) { |
2900 | namespace_data->roster_count = 0; |
2901 | } |
2902 | else { |
2903 | msg_buf_ele* roster_ele = cf_vector_getp(&rosters, i); |
2904 | |
2905 | namespace_data->roster_count = roster_ele->sz / sizeof(cf_node); |
2906 | |
2907 | if (namespace_data->roster_count == 0 |
2908 | || namespace_data->roster_count > AS_CLUSTER_SZ |
2909 | || roster_ele->sz % sizeof(cf_node) != 0) { |
2910 | WARNING( |
2911 | "received invalid roster for namespace %s from node %" PRIx64, |
2912 | matching_namespace->name, msg_event->msg_source); |
2913 | goto Exit; |
2914 | } |
2915 | |
2916 | namespace_data->roster = cf_realloc(namespace_data->roster, |
2917 | roster_ele->sz); |
2918 | |
2919 | memcpy(namespace_data->roster, roster_ele->ptr, roster_ele->sz); |
2920 | |
2921 | uint32_t rri_ele_sz = 0; |
2922 | |
2923 | if (cf_vector_size(&rosters_rack_ids) != 0) { |
2924 | msg_buf_ele* rri_ele = cf_vector_getp(&rosters_rack_ids, i); |
2925 | |
2926 | if (rri_ele->sz != 0) { |
2927 | rri_ele_sz = rri_ele->sz; |
2928 | |
2929 | if (rri_ele_sz |
2930 | != namespace_data->roster_count |
2931 | * sizeof(uint32_t)) { |
2932 | WARNING( |
2933 | "received invalid roster-rack-ids for namespace %s from node %" PRIx64, |
2934 | matching_namespace->name, |
2935 | msg_event->msg_source); |
2936 | goto Exit; |
2937 | } |
2938 | |
2939 | namespace_data->roster_rack_ids = cf_realloc( |
2940 | namespace_data->roster_rack_ids, rri_ele_sz); |
2941 | |
2942 | memcpy(namespace_data->roster_rack_ids, rri_ele->ptr, |
2943 | rri_ele_sz); |
2944 | } |
2945 | } |
2946 | |
2947 | if (rri_ele_sz == 0 && namespace_data->roster_rack_ids) { |
2948 | cf_free(namespace_data->roster_rack_ids); |
2949 | namespace_data->roster_rack_ids = NULL; |
2950 | } |
2951 | } |
2952 | } |
2953 | |
2954 | // Mark exchange data received from the source. |
2955 | node_state.received = true; |
2956 | exchange_node_state_update(msg_event->msg_source, &node_state); |
2957 | } |
2958 | else { |
2959 | // Duplicate pinfo received. Ignore. |
2960 | INFO("received duplicate exchange data from node %" PRIx64, |
2961 | msg_event->msg_source); |
2962 | } |
2963 | |
2964 | // Send an acknowledgement. |
2965 | exchange_data_ack_msg_send(msg_event->msg_source); |
2966 | |
2967 | // Check if we can switch to ready to commit state. |
2968 | exchange_exchanging_check_switch_ready_to_commit(); |
2969 | |
2970 | Exit: |
2971 | EXCHANGE_UNLOCK(); |
2972 | } |
2973 | |
2974 | /** |
2975 | * Handle incoming data ack message. |
2976 | * |
2977 | * Assumes the message has been checked for sanity. |
2978 | */ |
2979 | static void |
2980 | exchange_exchanging_data_ack_msg_handle(as_exchange_event* msg_event) |
2981 | { |
2982 | EXCHANGE_LOCK(); |
2983 | |
2984 | DEBUG("received exchange data ack from node %" PRIx64, |
2985 | msg_event->msg_source); |
2986 | |
2987 | as_exchange_node_state node_state; |
2988 | exchange_node_state_get_safe(msg_event->msg_source, &node_state); |
2989 | |
2990 | if (!node_state.send_acked) { |
2991 | // Mark send as acked in the node state. |
2992 | node_state.send_acked = true; |
2993 | exchange_node_state_update(msg_event->msg_source, &node_state); |
2994 | } |
2995 | else { |
2996 | // Duplicate ack. Ignore. |
2997 | DEBUG("received duplicate data ack from node %" PRIx64, |
2998 | msg_event->msg_source); |
2999 | } |
3000 | |
3001 | // We might have send and received all partition info. Check for completion. |
3002 | exchange_exchanging_check_switch_ready_to_commit(); |
3003 | |
3004 | EXCHANGE_UNLOCK(); |
3005 | } |
3006 | |
3007 | /** |
3008 | * Process a message event when in exchanging state. |
3009 | */ |
3010 | static void |
3011 | exchange_exchanging_msg_event_handle(as_exchange_event* msg_event) |
3012 | { |
3013 | EXCHANGE_LOCK(); |
3014 | |
3015 | if (!exchange_msg_is_sane(msg_event->msg_source, msg_event->msg)) { |
3016 | goto Exit; |
3017 | } |
3018 | |
3019 | as_exchange_msg_type msg_type; |
3020 | exchange_msg_type_get(msg_event->msg, &msg_type); |
3021 | |
3022 | switch (msg_type) { |
3023 | case AS_EXCHANGE_MSG_TYPE_DATA: |
3024 | exchange_exchanging_data_msg_handle(msg_event); |
3025 | break; |
3026 | case AS_EXCHANGE_MSG_TYPE_DATA_ACK: |
3027 | exchange_exchanging_data_ack_msg_handle(msg_event); |
3028 | break; |
3029 | default: |
3030 | DEBUG( |
3031 | "exchanging state received unexpected mesage of type %d from node %" PRIx64, |
3032 | msg_type, msg_event->msg_source); |
3033 | } |
3034 | Exit: |
3035 | EXCHANGE_UNLOCK(); |
3036 | } |
3037 | |
3038 | /** |
3039 | * Process a message event when in exchanging state. |
3040 | */ |
3041 | static void |
3042 | exchange_exchanging_timer_event_handle(as_exchange_event* msg_event) |
3043 | { |
3044 | EXCHANGE_LOCK(); |
3045 | bool send_data = false; |
3046 | |
3047 | cf_clock now = cf_getms(); |
3048 | |
3049 | // The timeout is a "linear" step function, where the timeout is constant |
3050 | // for the step interval. |
3051 | cf_clock min_timeout = EXCHANGE_SEND_MIN_TIMEOUT(); |
3052 | cf_clock max_timeout = EXCHANGE_SEND_MAX_TIMEOUT(); |
3053 | uint32_t step_interval = EXCHANGE_SEND_STEP_INTERVAL(); |
3054 | cf_clock timeout = MAX(min_timeout, |
3055 | MIN(max_timeout, |
3056 | min_timeout |
3057 | * ((now - g_exchange.send_ts) / step_interval))); |
3058 | |
3059 | if (g_exchange.send_ts + timeout < now) { |
3060 | send_data = true; |
3061 | } |
3062 | |
3063 | EXCHANGE_UNLOCK(); |
3064 | |
3065 | if (send_data) { |
3066 | exchange_data_msg_send_pending_ack(); |
3067 | } |
3068 | } |
3069 | |
3070 | /** |
3071 | * Event processing in the exchanging state. |
3072 | */ |
3073 | static void |
3074 | exchange_exchanging_event_handle(as_exchange_event* event) |
3075 | { |
3076 | switch (event->type) { |
3077 | case AS_EXCHANGE_EVENT_CLUSTER_CHANGE: |
3078 | exchange_clustering_event_handle(event); |
3079 | break; |
3080 | case AS_EXCHANGE_EVENT_MSG: |
3081 | exchange_exchanging_msg_event_handle(event); |
3082 | break; |
3083 | case AS_EXCHANGE_EVENT_TIMER: |
3084 | exchange_exchanging_timer_event_handle(event); |
3085 | break; |
3086 | } |
3087 | } |
3088 | |
3089 | /* |
3090 | * ---------------------------------------------------------------------------- |
3091 | * Ready_To_Commit state event handling |
3092 | * ---------------------------------------------------------------------------- |
3093 | */ |
3094 | |
3095 | /** |
3096 | * Handle incoming ready to commit message. |
3097 | * |
3098 | * Assumes the message has been checked for sanity. |
3099 | */ |
3100 | static void |
3101 | exchange_ready_to_commit_rtc_msg_handle(as_exchange_event* msg_event) |
3102 | { |
3103 | if (!exchange_self_is_principal()) { |
3104 | WARNING( |
3105 | "non-principal self received ready to commit message from %" PRIx64" - ignoring" , |
3106 | msg_event->msg_source); |
3107 | return; |
3108 | } |
3109 | |
3110 | EXCHANGE_LOCK(); |
3111 | |
3112 | DEBUG("received ready to commit from node %" PRIx64, msg_event->msg_source); |
3113 | |
3114 | as_exchange_node_state node_state; |
3115 | exchange_node_state_get_safe(msg_event->msg_source, &node_state); |
3116 | |
3117 | if (!node_state.is_ready_to_commit) { |
3118 | // Mark as ready to commit in the node state. |
3119 | node_state.is_ready_to_commit = true; |
3120 | exchange_node_state_update(msg_event->msg_source, &node_state); |
3121 | } |
3122 | else { |
3123 | // Duplicate ready to commit received. Ignore. |
3124 | INFO("received duplicate ready to commit message from node %" PRIx64, |
3125 | msg_event->msg_source); |
3126 | } |
3127 | |
3128 | cf_vector* node_vector = cf_vector_stack_create(cf_node); |
3129 | exchange_nodes_find_not_ready_to_commit(node_vector); |
3130 | |
3131 | if (cf_vector_size(node_vector) <= 0) { |
3132 | // Send a commit message to all nodes in succession list. |
3133 | cf_node* node_list = NULL; |
3134 | int num_node_list = 0; |
3135 | cf_vector_to_stack_array(&g_exchange.succession_list, &node_list, |
3136 | &num_node_list); |
3137 | exchange_commit_msg_send_all(node_list, num_node_list); |
3138 | } |
3139 | |
3140 | cf_vector_destroy(node_vector); |
3141 | |
3142 | EXCHANGE_UNLOCK(); |
3143 | } |
3144 | |
3145 | /** |
3146 | * Commit accumulated exchange data. |
3147 | */ |
3148 | static void |
3149 | exchange_data_commit() |
3150 | { |
3151 | EXCHANGE_LOCK(); |
3152 | |
3153 | INFO("data exchange completed with cluster key %" PRIx64, |
3154 | g_exchange.cluster_key); |
3155 | |
3156 | // Exchange is done, use the current cluster details as the committed |
3157 | // cluster details. |
3158 | exchange_commited_cluster_update(g_exchange.cluster_key, |
3159 | &g_exchange.succession_list); |
3160 | |
3161 | // Force an update of the skew, to ensure new nodes if any have been checked |
3162 | // for skew. |
3163 | as_skew_monitor_update(); |
3164 | |
3165 | // Must cover partition balance since it may manipulate ns->cluster_size. |
3166 | pthread_mutex_lock(&g_exchanged_info_lock); |
3167 | as_partition_balance(); |
3168 | pthread_mutex_unlock(&g_exchanged_info_lock); |
3169 | |
3170 | EXCHANGE_UNLOCK(); |
3171 | } |
3172 | |
3173 | /** |
3174 | * Handle incoming data ack message. |
3175 | * |
3176 | * Assumes the message has been checked for sanity. |
3177 | */ |
3178 | static void |
3179 | exchange_ready_to_commit_commit_msg_handle(as_exchange_event* msg_event) |
3180 | { |
3181 | EXCHANGE_LOCK(); |
3182 | |
3183 | if (msg_event->msg_source != g_exchange.principal) { |
3184 | WARNING( |
3185 | "ignoring commit message from node %" PRIx64" - expected message from %" PRIx64, |
3186 | msg_event->msg_source, g_exchange.principal); |
3187 | goto Exit; |
3188 | } |
3189 | |
3190 | INFO("received commit command from principal node %" PRIx64, |
3191 | msg_event->msg_source); |
3192 | |
3193 | // Commit exchanged data. |
3194 | exchange_data_commit(); |
3195 | |
3196 | // Move to the rest state. |
3197 | g_exchange.state = AS_EXCHANGE_STATE_REST; |
3198 | |
3199 | // Queue up a cluster change event for downstream sub systems. |
3200 | as_exchange_cluster_changed_event cluster_change_event; |
3201 | |
3202 | EXCHANGE_COMMITTED_CLUSTER_RLOCK(); |
3203 | cluster_change_event.cluster_key = g_exchange.committed_cluster_key; |
3204 | cluster_change_event.succession = vector_to_array( |
3205 | &g_exchange.committed_succession_list); |
3206 | cluster_change_event.cluster_size = g_exchange.committed_cluster_size; |
3207 | exchange_external_event_queue(&cluster_change_event); |
3208 | EXCHANGE_COMMITTED_CLUSTER_UNLOCK(); |
3209 | |
3210 | Exit: |
3211 | EXCHANGE_UNLOCK(); |
3212 | } |
3213 | |
3214 | /** |
3215 | * Handle incoming data message in ready to commit stage. |
3216 | * |
3217 | * Assumes the message has been checked for sanity. |
3218 | */ |
3219 | static void |
3220 | exchange_ready_to_commit_data_msg_handle(as_exchange_event* msg_event) |
3221 | { |
3222 | EXCHANGE_LOCK(); |
3223 | |
3224 | DEBUG("received exchange data from node %" PRIx64, msg_event->msg_source); |
3225 | |
3226 | // The source must have missed self node's data ack. Send an |
3227 | // acknowledgement. |
3228 | exchange_data_ack_msg_send(msg_event->msg_source); |
3229 | |
3230 | EXCHANGE_UNLOCK(); |
3231 | } |
3232 | |
3233 | /** |
3234 | * Process a message event when in ready_to_commit state. |
3235 | */ |
3236 | static void |
3237 | exchange_ready_to_commit_msg_event_handle(as_exchange_event* msg_event) |
3238 | { |
3239 | EXCHANGE_LOCK(); |
3240 | |
3241 | if (!exchange_msg_is_sane(msg_event->msg_source, msg_event->msg)) { |
3242 | goto Exit; |
3243 | } |
3244 | |
3245 | as_exchange_msg_type msg_type; |
3246 | exchange_msg_type_get(msg_event->msg, &msg_type); |
3247 | |
3248 | switch (msg_type) { |
3249 | case AS_EXCHANGE_MSG_TYPE_READY_TO_COMMIT: |
3250 | exchange_ready_to_commit_rtc_msg_handle(msg_event); |
3251 | break; |
3252 | case AS_EXCHANGE_MSG_TYPE_COMMIT: |
3253 | exchange_ready_to_commit_commit_msg_handle(msg_event); |
3254 | break; |
3255 | case AS_EXCHANGE_MSG_TYPE_DATA: |
3256 | exchange_ready_to_commit_data_msg_handle(msg_event); |
3257 | break; |
3258 | default: |
3259 | DEBUG( |
3260 | "ready to commit state received unexpected message of type %d from node %" PRIx64, |
3261 | msg_type, msg_event->msg_source); |
3262 | } |
3263 | Exit: |
3264 | EXCHANGE_UNLOCK(); |
3265 | } |
3266 | |
3267 | /** |
3268 | * Process a message event when in ready_to_commit state. |
3269 | */ |
3270 | static void |
3271 | exchange_ready_to_commit_timer_event_handle(as_exchange_event* msg_event) |
3272 | { |
3273 | EXCHANGE_LOCK(); |
3274 | |
3275 | if (g_exchange.ready_to_commit_send_ts + EXCHANGE_READY_TO_COMMIT_TIMEOUT() |
3276 | < cf_getms()) { |
3277 | // Its been a while since ready to commit has been sent to the |
3278 | // principal, retransmit it so that the principal gets it this time and |
3279 | // supplies a commit message. |
3280 | exchange_ready_to_commit_msg_send(); |
3281 | } |
3282 | EXCHANGE_UNLOCK(); |
3283 | } |
3284 | |
3285 | /** |
3286 | * Event processing in the ready_to_commit state. |
3287 | */ |
3288 | static void |
3289 | exchange_ready_to_commit_event_handle(as_exchange_event* event) |
3290 | { |
3291 | switch (event->type) { |
3292 | case AS_EXCHANGE_EVENT_CLUSTER_CHANGE: |
3293 | exchange_clustering_event_handle(event); |
3294 | break; |
3295 | case AS_EXCHANGE_EVENT_MSG: |
3296 | exchange_ready_to_commit_msg_event_handle(event); |
3297 | break; |
3298 | case AS_EXCHANGE_EVENT_TIMER: |
3299 | exchange_ready_to_commit_timer_event_handle(event); |
3300 | break; |
3301 | } |
3302 | } |
3303 | |
3304 | /* |
3305 | * ---------------------------------------------------------------------------- |
3306 | * Exchange core subsystem |
3307 | * ---------------------------------------------------------------------------- |
3308 | */ |
3309 | |
3310 | /** |
3311 | * Dispatch an exchange event inline to the relevant state handler. |
3312 | */ |
3313 | static void |
3314 | exchange_event_handle(as_exchange_event* event) |
3315 | { |
3316 | EXCHANGE_LOCK(); |
3317 | |
3318 | switch (g_exchange.state) { |
3319 | case AS_EXCHANGE_STATE_REST: |
3320 | exchange_rest_event_handle(event); |
3321 | break; |
3322 | case AS_EXCHANGE_STATE_EXCHANGING: |
3323 | exchange_exchanging_event_handle(event); |
3324 | break; |
3325 | case AS_EXCHANGE_STATE_READY_TO_COMMIT: |
3326 | exchange_ready_to_commit_event_handle(event); |
3327 | break; |
3328 | case AS_EXCHANGE_STATE_ORPHANED: |
3329 | exchange_orphan_event_handle(event); |
3330 | break; |
3331 | } |
3332 | |
3333 | EXCHANGE_UNLOCK(); |
3334 | } |
3335 | |
3336 | /** |
3337 | * Exchange timer event generator thread, to help with retries and retransmits |
3338 | * across all states. |
3339 | */ |
3340 | static void* |
3341 | exchange_timer_thr(void* arg) |
3342 | { |
3343 | as_exchange_event timer_event; |
3344 | memset(&timer_event, 0, sizeof(timer_event)); |
3345 | timer_event.type = AS_EXCHANGE_EVENT_TIMER; |
3346 | |
3347 | while (EXCHANGE_IS_RUNNING()) { |
3348 | // Wait for a while and retry. |
3349 | usleep(EXCHANGE_TIMER_TICK_INTERVAL() * 1000); |
3350 | exchange_event_handle(&timer_event); |
3351 | } |
3352 | return NULL; |
3353 | } |
3354 | |
3355 | /** |
3356 | * Handle incoming messages from fabric. |
3357 | */ |
3358 | static int |
3359 | exchange_fabric_msg_listener(cf_node source, msg* msg, void* udata) |
3360 | { |
3361 | if (!EXCHANGE_IS_RUNNING()) { |
3362 | // Ignore this message. |
3363 | DEBUG("exchange stopped - ignoring message from %" PRIx64, source); |
3364 | goto Exit; |
3365 | } |
3366 | |
3367 | as_exchange_event msg_event; |
3368 | memset(&msg_event, 0, sizeof(msg_event)); |
3369 | msg_event.type = AS_EXCHANGE_EVENT_MSG; |
3370 | msg_event.msg = msg; |
3371 | msg_event.msg_source = source; |
3372 | |
3373 | exchange_event_handle(&msg_event); |
3374 | Exit: |
3375 | as_fabric_msg_put(msg); |
3376 | return 0; |
3377 | } |
3378 | |
3379 | /** |
3380 | * Listener for cluster change events from clustering layer. |
3381 | */ |
3382 | void |
3383 | exchange_clustering_event_listener(as_clustering_event* event) |
3384 | { |
3385 | if (!EXCHANGE_IS_RUNNING()) { |
3386 | // Ignore this message. |
3387 | DEBUG("exchange stopped - ignoring cluster change event" ); |
3388 | return; |
3389 | } |
3390 | |
3391 | as_exchange_event clustering_event; |
3392 | memset(&clustering_event, 0, sizeof(clustering_event)); |
3393 | clustering_event.type = AS_EXCHANGE_EVENT_CLUSTER_CHANGE; |
3394 | clustering_event.clustering_event = event; |
3395 | |
3396 | // Dispatch the event. |
3397 | exchange_event_handle(&clustering_event); |
3398 | } |
3399 | |
3400 | /** |
3401 | * Initialize the template to be used for exchange messages. |
3402 | */ |
3403 | static void |
3404 | exchange_msg_init() |
3405 | { |
3406 | // Register fabric exchange msg type with no processing function. |
3407 | as_fabric_register_msg_fn(M_TYPE_EXCHANGE, exchange_msg_template, |
3408 | sizeof(exchange_msg_template), AS_EXCHANGE_MSG_SCRATCH_SIZE, |
3409 | exchange_fabric_msg_listener, NULL); |
3410 | } |
3411 | |
3412 | /** |
3413 | * Initialize exchange subsystem. |
3414 | */ |
3415 | static void |
3416 | exchange_init() |
3417 | { |
3418 | if (EXCHANGE_IS_INITIALIZED()) { |
3419 | return; |
3420 | } |
3421 | |
3422 | EXCHANGE_LOCK(); |
3423 | |
3424 | memset(&g_exchange, 0, sizeof(g_exchange)); |
3425 | |
3426 | // Start in the orphaned state. |
3427 | g_exchange.state = AS_EXCHANGE_STATE_ORPHANED; |
3428 | g_exchange.orphan_state_start_time = cf_getms(); |
3429 | g_exchange.orphan_state_are_transactions_blocked = true; |
3430 | |
3431 | // Initialize the adjacencies. |
3432 | g_exchange.nodeid_to_node_state = cf_shash_create(cf_nodeid_shash_fn, |
3433 | sizeof(cf_node), sizeof(as_exchange_node_state), |
3434 | AS_EXCHANGE_CLUSTER_MAX_SIZE_SOFT, 0); |
3435 | |
3436 | cf_vector_init(&g_exchange.succession_list, sizeof(cf_node), |
3437 | AS_EXCHANGE_CLUSTER_MAX_SIZE_SOFT, VECTOR_FLAG_INITZERO); |
3438 | |
3439 | EXCHANGE_COMMITTED_CLUSTER_WLOCK(); |
3440 | cf_vector_init(&g_exchange.committed_succession_list, sizeof(cf_node), |
3441 | AS_EXCHANGE_CLUSTER_MAX_SIZE_SOFT, VECTOR_FLAG_INITZERO); |
3442 | EXCHANGE_COMMITTED_CLUSTER_UNLOCK(); |
3443 | |
3444 | // Initialize exchange fabric messaging. |
3445 | exchange_msg_init(); |
3446 | |
3447 | // Initialize self exchange data dynamic buffers. |
3448 | for (uint32_t ns_ix = 0; ns_ix < g_config.n_namespaces; ns_ix++) { |
3449 | cf_dyn_buf_init_heap(&g_exchange.self_data_dyn_buf[ns_ix], |
3450 | AS_EXCHANGE_SELF_DYN_BUF_SIZE()); |
3451 | } |
3452 | |
3453 | // Initialize external event publishing. |
3454 | exchange_external_event_publisher_init(); |
3455 | |
3456 | // Get partition versions from storage. |
3457 | as_partition_balance_init(); |
3458 | |
3459 | DEBUG("exchange module initialized" ); |
3460 | |
3461 | EXCHANGE_UNLOCK(); |
3462 | } |
3463 | |
3464 | /** |
3465 | * Stop exchange subsystem. |
3466 | */ |
3467 | static void |
3468 | exchange_stop() |
3469 | { |
3470 | if (!EXCHANGE_IS_RUNNING()) { |
3471 | WARNING("exchange is already stopped" ); |
3472 | return; |
3473 | } |
3474 | |
3475 | // Unguarded state, but this should be ok. |
3476 | g_exchange.sys_state = AS_EXCHANGE_SYS_STATE_SHUTTING_DOWN; |
3477 | |
3478 | cf_thread_join(g_exchange.timer_tid); |
3479 | |
3480 | EXCHANGE_LOCK(); |
3481 | |
3482 | g_exchange.sys_state = AS_EXCHANGE_SYS_STATE_STOPPED; |
3483 | |
3484 | DEBUG("exchange module stopped" ); |
3485 | |
3486 | EXCHANGE_UNLOCK(); |
3487 | |
3488 | external_event_publisher_stop(); |
3489 | } |
3490 | |
3491 | /** |
3492 | * Start the exchange subsystem. |
3493 | */ |
3494 | static void |
3495 | exchange_start() |
3496 | { |
3497 | EXCHANGE_LOCK(); |
3498 | |
3499 | if (EXCHANGE_IS_RUNNING()) { |
3500 | // Shutdown the exchange subsystem. |
3501 | exchange_stop(); |
3502 | } |
3503 | |
3504 | g_exchange.sys_state = AS_EXCHANGE_SYS_STATE_RUNNING; |
3505 | |
3506 | g_exchange.timer_tid = cf_thread_create_joinable(exchange_timer_thr, |
3507 | (void*)&g_exchange); |
3508 | |
3509 | DEBUG("exchange module started" ); |
3510 | |
3511 | EXCHANGE_UNLOCK(); |
3512 | |
3513 | exchange_external_event_publisher_start(); |
3514 | } |
3515 | |
3516 | /* |
3517 | * ---------------------------------------------------------------------------- |
3518 | * Public API |
3519 | * ---------------------------------------------------------------------------- |
3520 | */ |
3521 | /** |
3522 | * Initialize exchange subsystem. |
3523 | */ |
3524 | void |
3525 | as_exchange_init() |
3526 | { |
3527 | exchange_init(); |
3528 | } |
3529 | |
3530 | /** |
3531 | * Start exchange subsystem. |
3532 | */ |
3533 | void |
3534 | as_exchange_start() |
3535 | { |
3536 | exchange_start(); |
3537 | } |
3538 | |
3539 | /** |
3540 | * Stop exchange subsystem. |
3541 | */ |
3542 | void |
3543 | as_exchange_stop() |
3544 | { |
3545 | } |
3546 | |
3547 | /** |
3548 | * Register to receive cluster-changed events. |
3549 | * TODO - may replace with simple static list someday. |
3550 | */ |
3551 | void |
3552 | as_exchange_register_listener(as_exchange_cluster_changed_cb cb, void* udata) |
3553 | { |
3554 | exchange_external_event_listener_register(cb, udata); |
3555 | } |
3556 | |
3557 | /** |
3558 | * Dump exchange state to log. |
3559 | */ |
3560 | void |
3561 | as_exchange_dump(bool verbose) |
3562 | { |
3563 | exchange_dump(CF_INFO, verbose); |
3564 | } |
3565 | |
3566 | /** |
3567 | * Member-access method. |
3568 | */ |
3569 | uint64_t |
3570 | as_exchange_cluster_key() |
3571 | { |
3572 | return (uint64_t)g_exchange.committed_cluster_key; |
3573 | } |
3574 | |
3575 | /** |
3576 | * Member-access method. |
3577 | */ |
3578 | uint32_t |
3579 | as_exchange_cluster_size() |
3580 | { |
3581 | return g_exchange.committed_cluster_size; |
3582 | } |
3583 | |
3584 | /** |
3585 | * Copy over the committed succession list. |
3586 | * Ensure the input vector has enough capacity. |
3587 | */ |
3588 | void |
3589 | as_exchange_succession(cf_vector* succession) |
3590 | { |
3591 | EXCHANGE_COMMITTED_CLUSTER_RLOCK(); |
3592 | vector_copy(succession, &g_exchange.committed_succession_list); |
3593 | EXCHANGE_COMMITTED_CLUSTER_UNLOCK(); |
3594 | } |
3595 | |
3596 | /** |
3597 | * Return the committed succession list. For internal use within the scope |
3598 | * exchange_data_commit function call. |
3599 | */ |
3600 | cf_node* |
3601 | as_exchange_succession_unsafe() |
3602 | { |
3603 | return vector_to_array(&g_exchange.committed_succession_list); |
3604 | } |
3605 | |
3606 | /** |
3607 | * Return the committed succession list as a string in a dyn-buf. |
3608 | */ |
3609 | void |
3610 | as_exchange_info_get_succession(cf_dyn_buf* db) |
3611 | { |
3612 | EXCHANGE_COMMITTED_CLUSTER_RLOCK(); |
3613 | |
3614 | cf_node* nodes = vector_to_array(&g_exchange.committed_succession_list); |
3615 | |
3616 | for (uint32_t i = 0; i < g_exchange.committed_cluster_size; i++) { |
3617 | cf_dyn_buf_append_uint64_x(db, nodes[i]); |
3618 | cf_dyn_buf_append_char(db, ','); |
3619 | } |
3620 | |
3621 | if (g_exchange.committed_cluster_size != 0) { |
3622 | cf_dyn_buf_chomp(db); |
3623 | } |
3624 | |
3625 | // Always succeeds. |
3626 | cf_dyn_buf_append_string(db, "\nok" ); |
3627 | |
3628 | EXCHANGE_COMMITTED_CLUSTER_UNLOCK(); |
3629 | } |
3630 | |
3631 | /** |
3632 | * Member-access method. |
3633 | */ |
3634 | cf_node |
3635 | as_exchange_principal() |
3636 | { |
3637 | return g_exchange.committed_principal; |
3638 | } |
3639 | |
3640 | /** |
3641 | * Used by exchange listeners during upgrades for compatibility purposes. |
3642 | */ |
3643 | uint32_t* |
3644 | as_exchange_compatibility_ids(void) |
3645 | { |
3646 | return (uint32_t*)g_exchange.compatibility_ids; |
3647 | } |
3648 | |
3649 | /** |
3650 | * Used by exchange listeners during upgrades for compatibility purposes. |
3651 | */ |
3652 | uint32_t |
3653 | as_exchange_min_compatibility_id(void) |
3654 | { |
3655 | return g_exchange.min_compatibility_id; |
3656 | } |
3657 | |
3658 | /** |
3659 | * Exchange cluster state output for info calls. |
3660 | */ |
3661 | void as_exchange_cluster_info(cf_dyn_buf* db) { |
3662 | EXCHANGE_COMMITTED_CLUSTER_RLOCK(); |
3663 | info_append_uint32(db, "cluster_size" , g_exchange.committed_cluster_size); |
3664 | info_append_uint64_x(db, "cluster_key" , g_exchange.committed_cluster_key); |
3665 | info_append_uint64(db, "cluster_generation" , g_exchange.committed_cluster_generation); |
3666 | info_append_uint64_x(db, "cluster_principal" , g_exchange.committed_principal); |
3667 | EXCHANGE_COMMITTED_CLUSTER_UNLOCK(); |
3668 | } |
3669 | |
3670 | /** |
3671 | * Lock before setting or getting exchanged info from non-exchange thread. |
3672 | */ |
3673 | void |
3674 | as_exchange_info_lock() |
3675 | { |
3676 | pthread_mutex_lock(&g_exchanged_info_lock); |
3677 | } |
3678 | |
3679 | /** |
3680 | * Unlock after setting or getting exchanged info from non-exchange thread. |
3681 | */ |
3682 | void |
3683 | as_exchange_info_unlock() |
3684 | { |
3685 | pthread_mutex_unlock(&g_exchanged_info_lock); |
3686 | } |
3687 | |