]> git.proxmox.com Git - mirror_frr.git/blob - zebra/zserv.c
Merge pull request #13649 from donaldsharp/unlock_the_node_or_else
[mirror_frr.git] / zebra / zserv.c
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3 * Zebra API server.
4 * Portions:
5 * Copyright (C) 1997-1999 Kunihiro Ishiguro
6 * Copyright (C) 2015-2018 Cumulus Networks, Inc.
7 * et al.
8 */
9
10 #include <zebra.h>
11
12 /* clang-format off */
13 #include <errno.h> /* for errno */
14 #include <netinet/in.h> /* for sockaddr_in */
15 #include <stdint.h> /* for uint8_t */
16 #include <stdio.h> /* for snprintf */
17 #include <sys/socket.h> /* for sockaddr_storage, AF_UNIX, accept... */
18 #include <sys/stat.h> /* for umask, mode_t */
19 #include <sys/un.h> /* for sockaddr_un */
20 #include <time.h> /* for NULL, tm, gmtime, time_t */
21 #include <unistd.h> /* for close, unlink, ssize_t */
22
23 #include "lib/buffer.h" /* for BUFFER_EMPTY, BUFFER_ERROR, BUFFE... */
24 #include "lib/command.h" /* for vty, install_element, CMD_SUCCESS... */
25 #include "lib/hook.h" /* for DEFINE_HOOK, DEFINE_KOOH, hook_call */
26 #include "lib/linklist.h" /* for ALL_LIST_ELEMENTS_RO, ALL_LIST_EL... */
27 #include "lib/libfrr.h" /* for frr_zclient_addr */
28 #include "lib/log.h" /* for zlog_warn, zlog_debug, safe_strerror */
29 #include "lib/memory.h" /* for MTYPE_TMP, XCALLOC, XFREE */
30 #include "lib/monotime.h" /* for monotime, ONE_DAY_SECOND, ONE_WEE... */
31 #include "lib/network.h" /* for set_nonblocking */
32 #include "lib/privs.h" /* for zebra_privs_t, ZPRIVS_LOWER, ZPRI... */
33 #include "lib/route_types.h" /* for ZEBRA_ROUTE_MAX */
34 #include "lib/sockopt.h" /* for setsockopt_so_recvbuf, setsockopt... */
35 #include "lib/sockunion.h" /* for sockopt_reuseaddr, sockopt_reuseport */
36 #include "lib/stream.h" /* for STREAM_SIZE, stream (ptr only), ... */
37 #include "frrevent.h" /* for thread (ptr only), EVENT_ARG, ... */
38 #include "lib/vrf.h" /* for vrf_info_lookup, VRF_DEFAULT */
39 #include "lib/vty.h" /* for vty_out, vty (ptr only) */
40 #include "lib/zclient.h" /* for zmsghdr, ZEBRA_HEADER_SIZE, ZEBRA... */
41 #include "lib/frr_pthread.h" /* for frr_pthread_new, frr_pthread_stop... */
42 #include "lib/frratomic.h" /* for atomic_load_explicit, atomic_stor... */
43 #include "lib/lib_errors.h" /* for generic ferr ids */
44 #include "lib/printfrr.h" /* for string functions */
45
46 #include "zebra/debug.h" /* for various debugging macros */
47 #include "zebra/rib.h" /* for rib_score_proto */
48 #include "zebra/zapi_msg.h" /* for zserv_handle_commands */
49 #include "zebra/zebra_vrf.h" /* for zebra_vrf_lookup_by_id, zvrf */
50 #include "zebra/zserv.h" /* for zserv */
51 #include "zebra/zebra_router.h"
52 #include "zebra/zebra_errors.h" /* for error messages */
53 /* clang-format on */
54
55 /* privileges */
56 extern struct zebra_privs_t zserv_privs;
57
58 /* The listener socket for clients connecting to us */
59 static int zsock;
60
61 /* The lock that protects access to zapi client objects */
62 static pthread_mutex_t client_mutex;
63
64 static struct zserv *find_client_internal(uint8_t proto,
65 unsigned short instance,
66 uint32_t session_id);
67
68 /* Mem type for zclients. */
69 DEFINE_MTYPE_STATIC(ZEBRA, ZSERV_CLIENT, "ZClients");
70
71 /*
72 * Client thread events.
73 *
74 * These are used almost exclusively by client threads to drive their own event
75 * loops. The only exception is in zserv_client_create(), which pushes an
76 * initial ZSERV_CLIENT_READ event to start the API handler loop.
77 */
78 enum zserv_client_event {
79 /* Schedule a socket read */
80 ZSERV_CLIENT_READ,
81 /* Schedule a buffer write */
82 ZSERV_CLIENT_WRITE,
83 };
84
85 /*
86 * Main thread events.
87 *
88 * These are used by client threads to notify the main thread about various
89 * events and to make processing requests.
90 */
91 enum zserv_event {
92 /* Schedule listen job on Zebra API socket */
93 ZSERV_ACCEPT,
94 /* The calling client has packets on its input buffer */
95 ZSERV_PROCESS_MESSAGES,
96 /* The calling client wishes to be killed */
97 ZSERV_HANDLE_CLIENT_FAIL,
98 };
99
100 /*
101 * Zebra server event driver for all client threads.
102 *
103 * This is essentially a wrapper around event_add_event() that centralizes
104 * those scheduling calls into one place.
105 *
106 * All calls to this function schedule an event on the pthread running the
107 * provided client.
108 *
109 * client
110 * the client in question, and thread target
111 *
112 * event
113 * the event to notify them about
114 */
115 static void zserv_client_event(struct zserv *client,
116 enum zserv_client_event event);
117
118 /*
119 * Zebra server event driver for the main thread.
120 *
121 * This is essentially a wrapper around event_add_event() that centralizes
122 * those scheduling calls into one place.
123 *
124 * All calls to this function schedule an event on Zebra's main pthread.
125 *
126 * client
127 * the client in question
128 *
129 * event
130 * the event to notify the main thread about
131 */
132 static void zserv_event(struct zserv *client, enum zserv_event event);
133
134
135 /* Client thread lifecycle -------------------------------------------------- */
136
137 /*
138 * Free a zserv client object.
139 */
140 void zserv_client_delete(struct zserv *client)
141 {
142 XFREE(MTYPE_ZSERV_CLIENT, client);
143 }
144
145 /*
146 * Log zapi message to zlog.
147 *
148 * errmsg (optional)
149 * Debugging message
150 *
151 * msg
152 * The message
153 *
154 * hdr (optional)
155 * The message header
156 */
157 void zserv_log_message(const char *errmsg, struct stream *msg,
158 struct zmsghdr *hdr)
159 {
160 zlog_debug("Rx'd ZAPI message");
161 if (errmsg)
162 zlog_debug("%s", errmsg);
163 if (hdr) {
164 zlog_debug(" Length: %d", hdr->length);
165 zlog_debug("Command: %s", zserv_command_string(hdr->command));
166 zlog_debug(" VRF: %u", hdr->vrf_id);
167 }
168 stream_hexdump(msg);
169 }
170
171 /*
172 * Gracefuly shut down a client connection.
173 *
174 * Cancel any pending tasks for the client's thread. Then schedule a task on
175 * the main thread to shut down the calling thread.
176 *
177 * It is not safe to close the client socket in this function. The socket is
178 * owned by the main thread.
179 *
180 * Must be called from the client pthread, never the main thread.
181 */
182 static void zserv_client_fail(struct zserv *client)
183 {
184 flog_warn(EC_ZEBRA_CLIENT_IO_ERROR,
185 "Client '%s' encountered an error and is shutting down.",
186 zebra_route_string(client->proto));
187
188 atomic_store_explicit(&client->pthread->running, false,
189 memory_order_relaxed);
190
191 EVENT_OFF(client->t_read);
192 EVENT_OFF(client->t_write);
193 zserv_event(client, ZSERV_HANDLE_CLIENT_FAIL);
194 }
195
196 /*
197 * Write all pending messages to client socket.
198 *
199 * This function first attempts to flush any buffered data. If unsuccessful,
200 * the function reschedules itself and returns. If successful, it pops all
201 * available messages from the output queue and continues to write data
202 * directly to the socket until the socket would block. If the socket never
203 * blocks and all data is written, the function returns without rescheduling
204 * itself. If the socket ends up throwing EWOULDBLOCK, the remaining data is
205 * buffered and the function reschedules itself.
206 *
207 * The utility of the buffer is that it allows us to vastly reduce lock
208 * contention by allowing us to pop *all* messages off the output queue at once
209 * instead of locking and unlocking each time we want to pop a single message
210 * off the queue. The same thing could arguably be accomplished faster by
211 * allowing the main thread to write directly into the buffer instead of
212 * enqueuing packets onto an intermediary queue, but the intermediary queue
213 * allows us to expose information about input and output queues to the user in
214 * terms of number of packets rather than size of data.
215 */
216 static void zserv_write(struct event *thread)
217 {
218 struct zserv *client = EVENT_ARG(thread);
219 struct stream *msg;
220 uint32_t wcmd = 0;
221 struct stream_fifo *cache;
222 uint64_t time_now = monotime(NULL);
223
224 /* If we have any data pending, try to flush it first */
225 switch (buffer_flush_all(client->wb, client->sock)) {
226 case BUFFER_ERROR:
227 goto zwrite_fail;
228 case BUFFER_PENDING:
229 frr_with_mutex (&client->stats_mtx) {
230 client->last_write_time = time_now;
231 }
232 zserv_client_event(client, ZSERV_CLIENT_WRITE);
233 return;
234 case BUFFER_EMPTY:
235 break;
236 }
237
238 cache = stream_fifo_new();
239
240 frr_with_mutex (&client->obuf_mtx) {
241 while (stream_fifo_head(client->obuf_fifo))
242 stream_fifo_push(cache,
243 stream_fifo_pop(client->obuf_fifo));
244 }
245
246 if (cache->tail) {
247 msg = cache->tail;
248 stream_set_getp(msg, 0);
249 wcmd = stream_getw_from(msg, ZAPI_HEADER_CMD_LOCATION);
250 }
251
252 while (stream_fifo_head(cache)) {
253 msg = stream_fifo_pop(cache);
254 buffer_put(client->wb, STREAM_DATA(msg), stream_get_endp(msg));
255 stream_free(msg);
256 }
257
258 stream_fifo_free(cache);
259
260 /* If we have any data pending, try to flush it first */
261 switch (buffer_flush_all(client->wb, client->sock)) {
262 case BUFFER_ERROR:
263 goto zwrite_fail;
264 case BUFFER_PENDING:
265 frr_with_mutex (&client->stats_mtx) {
266 client->last_write_time = time_now;
267 }
268 zserv_client_event(client, ZSERV_CLIENT_WRITE);
269 return;
270 case BUFFER_EMPTY:
271 break;
272 }
273
274 frr_with_mutex (&client->stats_mtx) {
275 client->last_write_cmd = wcmd;
276 client->last_write_time = time_now;
277 }
278 return;
279
280 zwrite_fail:
281 flog_warn(EC_ZEBRA_CLIENT_WRITE_FAILED,
282 "%s: could not write to %s [fd = %d], closing.", __func__,
283 zebra_route_string(client->proto), client->sock);
284 zserv_client_fail(client);
285 }
286
287 /*
288 * Read and process data from a client socket.
289 *
290 * The responsibilities here are to read raw data from the client socket,
291 * validate the header, encapsulate it into a single stream object, push it
292 * onto the input queue and then notify the main thread that there is new data
293 * available.
294 *
295 * This function first looks for any data in the client structure's working
296 * input buffer. If data is present, it is assumed that reading stopped in a
297 * previous invocation of this task and needs to be resumed to finish a message.
298 * Otherwise, the socket data stream is assumed to be at the beginning of a new
299 * ZAPI message (specifically at the header). The header is read and validated.
300 * If the header passed validation then the length field found in the header is
301 * used to compute the total length of the message. That much data is read (but
302 * not inspected), appended to the header, placed into a stream and pushed onto
303 * the client's input queue. A task is then scheduled on the main thread to
304 * process the client's input queue. Finally, if all of this was successful,
305 * this task reschedules itself.
306 *
307 * Any failure in any of these actions is handled by terminating the client.
308 */
309 static void zserv_read(struct event *thread)
310 {
311 struct zserv *client = EVENT_ARG(thread);
312 int sock;
313 size_t already;
314 struct stream_fifo *cache;
315 uint32_t p2p_orig;
316
317 uint32_t p2p;
318 struct zmsghdr hdr;
319
320 p2p_orig = atomic_load_explicit(&zrouter.packets_to_process,
321 memory_order_relaxed);
322 cache = stream_fifo_new();
323 p2p = p2p_orig;
324 sock = EVENT_FD(thread);
325
326 while (p2p) {
327 ssize_t nb;
328 bool hdrvalid;
329 char errmsg[256];
330
331 already = stream_get_endp(client->ibuf_work);
332
333 /* Read length and command (if we don't have it already). */
334 if (already < ZEBRA_HEADER_SIZE) {
335 nb = stream_read_try(client->ibuf_work, sock,
336 ZEBRA_HEADER_SIZE - already);
337 if ((nb == 0 || nb == -1)) {
338 if (IS_ZEBRA_DEBUG_EVENT)
339 zlog_debug("connection closed socket [%d]",
340 sock);
341 goto zread_fail;
342 }
343 if (nb != (ssize_t)(ZEBRA_HEADER_SIZE - already)) {
344 /* Try again later. */
345 break;
346 }
347 already = ZEBRA_HEADER_SIZE;
348 }
349
350 /* Reset to read from the beginning of the incoming packet. */
351 stream_set_getp(client->ibuf_work, 0);
352
353 /* Fetch header values */
354 hdrvalid = zapi_parse_header(client->ibuf_work, &hdr);
355
356 if (!hdrvalid) {
357 snprintf(errmsg, sizeof(errmsg),
358 "%s: Message has corrupt header", __func__);
359 zserv_log_message(errmsg, client->ibuf_work, NULL);
360 goto zread_fail;
361 }
362
363 /* Validate header */
364 if (hdr.marker != ZEBRA_HEADER_MARKER
365 || hdr.version != ZSERV_VERSION) {
366 snprintf(
367 errmsg, sizeof(errmsg),
368 "Message has corrupt header\n%s: socket %d version mismatch, marker %d, version %d",
369 __func__, sock, hdr.marker, hdr.version);
370 zserv_log_message(errmsg, client->ibuf_work, &hdr);
371 goto zread_fail;
372 }
373 if (hdr.length < ZEBRA_HEADER_SIZE) {
374 snprintf(
375 errmsg, sizeof(errmsg),
376 "Message has corrupt header\n%s: socket %d message length %u is less than header size %d",
377 __func__, sock, hdr.length, ZEBRA_HEADER_SIZE);
378 zserv_log_message(errmsg, client->ibuf_work, &hdr);
379 goto zread_fail;
380 }
381 if (hdr.length > STREAM_SIZE(client->ibuf_work)) {
382 snprintf(
383 errmsg, sizeof(errmsg),
384 "Message has corrupt header\n%s: socket %d message length %u exceeds buffer size %lu",
385 __func__, sock, hdr.length,
386 (unsigned long)STREAM_SIZE(client->ibuf_work));
387 zserv_log_message(errmsg, client->ibuf_work, &hdr);
388 goto zread_fail;
389 }
390
391 /* Read rest of data. */
392 if (already < hdr.length) {
393 nb = stream_read_try(client->ibuf_work, sock,
394 hdr.length - already);
395 if ((nb == 0 || nb == -1)) {
396 if (IS_ZEBRA_DEBUG_EVENT)
397 zlog_debug(
398 "connection closed [%d] when reading zebra data",
399 sock);
400 goto zread_fail;
401 }
402 if (nb != (ssize_t)(hdr.length - already)) {
403 /* Try again later. */
404 break;
405 }
406 }
407
408 /* Debug packet information. */
409 if (IS_ZEBRA_DEBUG_PACKET)
410 zlog_debug("zebra message[%s:%u:%u] comes from socket [%d]",
411 zserv_command_string(hdr.command),
412 hdr.vrf_id, hdr.length,
413 sock);
414
415 stream_set_getp(client->ibuf_work, 0);
416 struct stream *msg = stream_dup(client->ibuf_work);
417
418 stream_fifo_push(cache, msg);
419 stream_reset(client->ibuf_work);
420 p2p--;
421 }
422
423 if (p2p < p2p_orig) {
424 uint64_t time_now = monotime(NULL);
425
426 /* update session statistics */
427 frr_with_mutex (&client->stats_mtx) {
428 client->last_read_time = time_now;
429 client->last_read_cmd = hdr.command;
430 }
431
432 /* publish read packets on client's input queue */
433 frr_with_mutex (&client->ibuf_mtx) {
434 while (cache->head)
435 stream_fifo_push(client->ibuf_fifo,
436 stream_fifo_pop(cache));
437 }
438
439 /* Schedule job to process those packets */
440 zserv_event(client, ZSERV_PROCESS_MESSAGES);
441
442 }
443
444 if (IS_ZEBRA_DEBUG_PACKET)
445 zlog_debug("Read %d packets from client: %s", p2p_orig - p2p,
446 zebra_route_string(client->proto));
447
448 /* Reschedule ourselves */
449 zserv_client_event(client, ZSERV_CLIENT_READ);
450
451 stream_fifo_free(cache);
452
453 return;
454
455 zread_fail:
456 stream_fifo_free(cache);
457 zserv_client_fail(client);
458 }
459
460 static void zserv_client_event(struct zserv *client,
461 enum zserv_client_event event)
462 {
463 switch (event) {
464 case ZSERV_CLIENT_READ:
465 event_add_read(client->pthread->master, zserv_read, client,
466 client->sock, &client->t_read);
467 break;
468 case ZSERV_CLIENT_WRITE:
469 event_add_write(client->pthread->master, zserv_write, client,
470 client->sock, &client->t_write);
471 break;
472 }
473 }
474
475 /* Main thread lifecycle ---------------------------------------------------- */
476
477 /*
478 * Read and process messages from a client.
479 *
480 * This task runs on the main pthread. It is scheduled by client pthreads when
481 * they have new messages available on their input queues. The client is passed
482 * as the task argument.
483 *
484 * Each message is popped off the client's input queue and the action associated
485 * with the message is executed. This proceeds until there are no more messages,
486 * an error occurs, or the processing limit is reached.
487 *
488 * The client's I/O thread can push at most zrouter.packets_to_process messages
489 * onto the input buffer before notifying us there are packets to read. As long
490 * as we always process zrouter.packets_to_process messages here, then we can
491 * rely on the read thread to handle queuing this task enough times to process
492 * everything on the input queue.
493 */
494 static void zserv_process_messages(struct event *thread)
495 {
496 struct zserv *client = EVENT_ARG(thread);
497 struct stream *msg;
498 struct stream_fifo *cache = stream_fifo_new();
499 uint32_t p2p = zrouter.packets_to_process;
500 bool need_resched = false;
501
502 frr_with_mutex (&client->ibuf_mtx) {
503 uint32_t i;
504 for (i = 0; i < p2p && stream_fifo_head(client->ibuf_fifo);
505 ++i) {
506 msg = stream_fifo_pop(client->ibuf_fifo);
507 stream_fifo_push(cache, msg);
508 }
509
510 /* Need to reschedule processing work if there are still
511 * packets in the fifo.
512 */
513 if (stream_fifo_head(client->ibuf_fifo))
514 need_resched = true;
515 }
516
517 /* Process the batch of messages */
518 if (stream_fifo_head(cache))
519 zserv_handle_commands(client, cache);
520
521 stream_fifo_free(cache);
522
523 /* Reschedule ourselves if necessary */
524 if (need_resched)
525 zserv_event(client, ZSERV_PROCESS_MESSAGES);
526 }
527
528 int zserv_send_message(struct zserv *client, struct stream *msg)
529 {
530 frr_with_mutex (&client->obuf_mtx) {
531 stream_fifo_push(client->obuf_fifo, msg);
532 }
533
534 zserv_client_event(client, ZSERV_CLIENT_WRITE);
535
536 return 0;
537 }
538
539 /*
540 * Send a batch of messages to a connected Zebra API client.
541 */
542 int zserv_send_batch(struct zserv *client, struct stream_fifo *fifo)
543 {
544 struct stream *msg;
545
546 frr_with_mutex (&client->obuf_mtx) {
547 msg = stream_fifo_pop(fifo);
548 while (msg) {
549 stream_fifo_push(client->obuf_fifo, msg);
550 msg = stream_fifo_pop(fifo);
551 }
552 }
553
554 zserv_client_event(client, ZSERV_CLIENT_WRITE);
555
556 return 0;
557 }
558
559 /* Hooks for client connect / disconnect */
560 DEFINE_HOOK(zserv_client_connect, (struct zserv *client), (client));
561 DEFINE_KOOH(zserv_client_close, (struct zserv *client), (client));
562
563 /*
564 * Deinitialize zebra client.
565 *
566 * - Deregister and deinitialize related internal resources
567 * - Gracefuly close socket
568 * - Free associated resources
569 * - Free client structure
570 *
571 * This does *not* take any action on the struct event * fields. These are
572 * managed by the owning pthread and any tasks associated with them must have
573 * been stopped prior to invoking this function.
574 */
575 static void zserv_client_free(struct zserv *client)
576 {
577 if (client == NULL)
578 return;
579
580 hook_call(zserv_client_close, client);
581
582 /* Close file descriptor. */
583 if (client->sock) {
584 unsigned long nroutes;
585 unsigned long nnhgs;
586
587 close(client->sock);
588
589 if (DYNAMIC_CLIENT_GR_DISABLED(client)) {
590 zebra_mpls_client_cleanup_vrf_label(client->proto);
591
592 nroutes = rib_score_proto(client->proto,
593 client->instance);
594 zlog_notice(
595 "client %d disconnected %lu %s routes removed from the rib",
596 client->sock, nroutes,
597 zebra_route_string(client->proto));
598
599 /* Not worrying about instance for now */
600 nnhgs = zebra_nhg_score_proto(client->proto);
601 zlog_notice(
602 "client %d disconnected %lu %s nhgs removed from the rib",
603 client->sock, nnhgs,
604 zebra_route_string(client->proto));
605 }
606 client->sock = -1;
607 }
608
609 /* Free stream buffers. */
610 if (client->ibuf_work)
611 stream_free(client->ibuf_work);
612 if (client->obuf_work)
613 stream_free(client->obuf_work);
614 if (client->ibuf_fifo)
615 stream_fifo_free(client->ibuf_fifo);
616 if (client->obuf_fifo)
617 stream_fifo_free(client->obuf_fifo);
618 if (client->wb)
619 buffer_free(client->wb);
620
621 /* Free buffer mutexes */
622 pthread_mutex_destroy(&client->stats_mtx);
623 pthread_mutex_destroy(&client->obuf_mtx);
624 pthread_mutex_destroy(&client->ibuf_mtx);
625
626 /* Free bitmaps. */
627 for (afi_t afi = AFI_IP; afi < AFI_MAX; afi++) {
628 for (int i = 0; i < ZEBRA_ROUTE_MAX; i++) {
629 vrf_bitmap_free(client->redist[afi][i]);
630 redist_del_all_instances(&client->mi_redist[afi][i]);
631 }
632
633 vrf_bitmap_free(client->redist_default[afi]);
634 vrf_bitmap_free(client->ridinfo[afi]);
635 vrf_bitmap_free(client->nhrp_neighinfo[afi]);
636 }
637
638 /*
639 * If any instance are graceful restart enabled,
640 * client is not deleted
641 */
642 if (DYNAMIC_CLIENT_GR_DISABLED(client)) {
643 if (IS_ZEBRA_DEBUG_EVENT)
644 zlog_debug("%s: Deleting client %s", __func__,
645 zebra_route_string(client->proto));
646 zserv_client_delete(client);
647 } else {
648 /* Handle cases where client has GR instance. */
649 if (IS_ZEBRA_DEBUG_EVENT)
650 zlog_debug("%s: client %s restart enabled", __func__,
651 zebra_route_string(client->proto));
652 if (zebra_gr_client_disconnect(client) < 0)
653 zlog_err(
654 "%s: GR enabled but could not handle disconnect event",
655 __func__);
656 }
657 }
658
659 void zserv_close_client(struct zserv *client)
660 {
661 bool free_p = true;
662
663 if (client->pthread) {
664 /* synchronously stop and join pthread */
665 frr_pthread_stop(client->pthread, NULL);
666
667 if (IS_ZEBRA_DEBUG_EVENT)
668 zlog_debug("Closing client '%s'",
669 zebra_route_string(client->proto));
670
671 event_cancel_event(zrouter.master, client);
672 EVENT_OFF(client->t_cleanup);
673 EVENT_OFF(client->t_process);
674
675 /* destroy pthread */
676 frr_pthread_destroy(client->pthread);
677 client->pthread = NULL;
678 }
679
680 /*
681 * Final check in case the client struct is in use in another
682 * pthread: if not in-use, continue and free the client
683 */
684 frr_with_mutex (&client_mutex) {
685 if (client->busy_count <= 0) {
686 /* remove from client list */
687 listnode_delete(zrouter.client_list, client);
688 } else {
689 /*
690 * The client session object may be in use, although
691 * the associated pthread is gone. Defer final
692 * cleanup.
693 */
694 client->is_closed = true;
695 free_p = false;
696 }
697 }
698
699 /* delete client */
700 if (free_p)
701 zserv_client_free(client);
702 }
703
704 /*
705 * This task is scheduled by a ZAPI client pthread on the main pthread when it
706 * wants to stop itself. When this executes, the client connection should
707 * already have been closed and the thread will most likely have died, but its
708 * resources still need to be cleaned up.
709 */
710 static void zserv_handle_client_fail(struct event *thread)
711 {
712 struct zserv *client = EVENT_ARG(thread);
713
714 zserv_close_client(client);
715 }
716
717 /*
718 * Create a new client.
719 *
720 * This is called when a new connection is accept()'d on the ZAPI socket. It
721 * initializes new client structure, notifies any subscribers of the connection
722 * event and spawns the client's thread.
723 *
724 * sock
725 * client's socket file descriptor
726 */
727 static struct zserv *zserv_client_create(int sock)
728 {
729 struct zserv *client;
730 size_t stream_size =
731 MAX(ZEBRA_MAX_PACKET_SIZ, sizeof(struct zapi_route));
732 int i;
733 afi_t afi;
734
735 client = XCALLOC(MTYPE_ZSERV_CLIENT, sizeof(struct zserv));
736
737 /* Make client input/output buffer. */
738 client->sock = sock;
739 client->ibuf_fifo = stream_fifo_new();
740 client->obuf_fifo = stream_fifo_new();
741 client->ibuf_work = stream_new(stream_size);
742 client->obuf_work = stream_new(stream_size);
743 client->connect_time = monotime(NULL);
744 pthread_mutex_init(&client->ibuf_mtx, NULL);
745 pthread_mutex_init(&client->obuf_mtx, NULL);
746 pthread_mutex_init(&client->stats_mtx, NULL);
747 client->wb = buffer_new(0);
748 TAILQ_INIT(&(client->gr_info_queue));
749
750 /* Initialize flags */
751 for (afi = AFI_IP; afi < AFI_MAX; afi++) {
752 for (i = 0; i < ZEBRA_ROUTE_MAX; i++)
753 client->redist[afi][i] = vrf_bitmap_init();
754 client->redist_default[afi] = vrf_bitmap_init();
755 client->ridinfo[afi] = vrf_bitmap_init();
756 client->nhrp_neighinfo[afi] = vrf_bitmap_init();
757 }
758
759 /* Add this client to linked list. */
760 frr_with_mutex (&client_mutex) {
761 listnode_add(zrouter.client_list, client);
762 }
763
764 struct frr_pthread_attr zclient_pthr_attrs = {
765 .start = frr_pthread_attr_default.start,
766 .stop = frr_pthread_attr_default.stop
767 };
768 client->pthread =
769 frr_pthread_new(&zclient_pthr_attrs, "Zebra API client thread",
770 "zebra_apic");
771
772 /* start read loop */
773 zserv_client_event(client, ZSERV_CLIENT_READ);
774
775 /* call callbacks */
776 hook_call(zserv_client_connect, client);
777
778 /* start pthread */
779 frr_pthread_run(client->pthread, NULL);
780
781 return client;
782 }
783
784 /*
785 * Retrieve a client object by the complete tuple of
786 * {protocol, instance, session}. This version supports use
787 * from a different pthread: the object will be returned marked
788 * in-use. The caller *must* release the client object with the
789 * release_client() api, to ensure that the in-use marker is cleared properly.
790 */
791 struct zserv *zserv_acquire_client(uint8_t proto, unsigned short instance,
792 uint32_t session_id)
793 {
794 struct zserv *client = NULL;
795
796 frr_with_mutex (&client_mutex) {
797 client = find_client_internal(proto, instance, session_id);
798 if (client) {
799 /* Don't return a dead/closed client object */
800 if (client->is_closed)
801 client = NULL;
802 else
803 client->busy_count++;
804 }
805 }
806
807 return client;
808 }
809
810 /*
811 * Release a client object that was acquired with the acquire_client() api.
812 * After this has been called, the caller must not use the client pointer -
813 * it may be freed if the client has closed.
814 */
815 void zserv_release_client(struct zserv *client)
816 {
817 /*
818 * Once we've decremented the client object's refcount, it's possible
819 * for it to be deleted as soon as we release the lock, so we won't
820 * touch the object again.
821 */
822 frr_with_mutex (&client_mutex) {
823 client->busy_count--;
824
825 if (client->busy_count <= 0) {
826 /*
827 * No more users of the client object. If the client
828 * session is closed, schedule cleanup on the zebra
829 * main pthread.
830 */
831 if (client->is_closed)
832 event_add_event(zrouter.master,
833 zserv_handle_client_fail,
834 client, 0, &client->t_cleanup);
835 }
836 }
837
838 /*
839 * Cleanup must take place on the zebra main pthread, so we've
840 * scheduled an event.
841 */
842 }
843
844 /*
845 * Accept socket connection.
846 */
847 static void zserv_accept(struct event *thread)
848 {
849 int accept_sock;
850 int client_sock;
851 struct sockaddr_in client;
852 socklen_t len;
853
854 accept_sock = EVENT_FD(thread);
855
856 /* Reregister myself. */
857 zserv_event(NULL, ZSERV_ACCEPT);
858
859 len = sizeof(struct sockaddr_in);
860 client_sock = accept(accept_sock, (struct sockaddr *)&client, &len);
861
862 if (client_sock < 0) {
863 flog_err_sys(EC_LIB_SOCKET, "Can't accept zebra socket: %s",
864 safe_strerror(errno));
865 return;
866 }
867
868 /* Make client socket non-blocking. */
869 set_nonblocking(client_sock);
870
871 /* Create new zebra client. */
872 zserv_client_create(client_sock);
873 }
874
875 void zserv_close(void)
876 {
877 /*
878 * On shutdown, let's close the socket down
879 * so that long running processes of killing the
880 * routing table doesn't leave us in a bad
881 * state where a client tries to reconnect
882 */
883 close(zsock);
884 zsock = -1;
885
886 /* Free client list's mutex */
887 pthread_mutex_destroy(&client_mutex);
888 }
889
890 void zserv_start(char *path)
891 {
892 int ret;
893 mode_t old_mask;
894 struct sockaddr_storage sa;
895 socklen_t sa_len;
896
897 if (!frr_zclient_addr(&sa, &sa_len, path))
898 /* should be caught in zebra main() */
899 return;
900
901 /* Set umask */
902 old_mask = umask(0077);
903
904 /* Make UNIX domain socket. */
905 zsock = socket(sa.ss_family, SOCK_STREAM, 0);
906 if (zsock < 0) {
907 flog_err_sys(EC_LIB_SOCKET, "Can't create zserv socket: %s",
908 safe_strerror(errno));
909 return;
910 }
911
912 if (sa.ss_family != AF_UNIX) {
913 sockopt_reuseaddr(zsock);
914 sockopt_reuseport(zsock);
915 } else {
916 struct sockaddr_un *suna = (struct sockaddr_un *)&sa;
917 if (suna->sun_path[0])
918 unlink(suna->sun_path);
919 }
920
921 setsockopt_so_recvbuf(zsock, 1048576);
922 setsockopt_so_sendbuf(zsock, 1048576);
923
924 frr_with_privs((sa.ss_family != AF_UNIX) ? &zserv_privs : NULL) {
925 ret = bind(zsock, (struct sockaddr *)&sa, sa_len);
926 }
927 if (ret < 0) {
928 flog_err_sys(EC_LIB_SOCKET, "Can't bind zserv socket on %s: %s",
929 path, safe_strerror(errno));
930 close(zsock);
931 zsock = -1;
932 return;
933 }
934
935 ret = listen(zsock, 5);
936 if (ret < 0) {
937 flog_err_sys(EC_LIB_SOCKET,
938 "Can't listen to zserv socket %s: %s", path,
939 safe_strerror(errno));
940 close(zsock);
941 zsock = -1;
942 return;
943 }
944
945 umask(old_mask);
946
947 zserv_event(NULL, ZSERV_ACCEPT);
948 }
949
950 void zserv_event(struct zserv *client, enum zserv_event event)
951 {
952 switch (event) {
953 case ZSERV_ACCEPT:
954 event_add_read(zrouter.master, zserv_accept, NULL, zsock, NULL);
955 break;
956 case ZSERV_PROCESS_MESSAGES:
957 event_add_event(zrouter.master, zserv_process_messages, client,
958 0, &client->t_process);
959 break;
960 case ZSERV_HANDLE_CLIENT_FAIL:
961 event_add_event(zrouter.master, zserv_handle_client_fail,
962 client, 0, &client->t_cleanup);
963 }
964 }
965
966
967 /* General purpose ---------------------------------------------------------- */
968
969 #define ZEBRA_TIME_BUF 32
970 static char *zserv_time_buf(time_t *time1, char *buf, int buflen)
971 {
972 time_t now;
973
974 assert(buf != NULL);
975 assert(buflen >= ZEBRA_TIME_BUF);
976 assert(time1 != NULL);
977
978 if (!*time1) {
979 snprintf(buf, buflen, "never ");
980 return (buf);
981 }
982
983 now = monotime(NULL);
984 now -= *time1;
985
986 frrtime_to_interval(now, buf, buflen);
987
988 return buf;
989 }
990
991 /* Display client info details */
992 static void zebra_show_client_detail(struct vty *vty, struct zserv *client)
993 {
994 char cbuf[ZEBRA_TIME_BUF], rbuf[ZEBRA_TIME_BUF];
995 char wbuf[ZEBRA_TIME_BUF], nhbuf[ZEBRA_TIME_BUF], mbuf[ZEBRA_TIME_BUF];
996 time_t connect_time, last_read_time, last_write_time;
997 uint32_t last_read_cmd, last_write_cmd;
998
999 vty_out(vty, "Client: %s", zebra_route_string(client->proto));
1000 if (client->instance)
1001 vty_out(vty, " Instance: %u", client->instance);
1002 if (client->session_id)
1003 vty_out(vty, " [%u]", client->session_id);
1004 vty_out(vty, "\n");
1005
1006 vty_out(vty, "------------------------ \n");
1007 vty_out(vty, "FD: %d \n", client->sock);
1008
1009 frr_with_mutex (&client->stats_mtx) {
1010 connect_time = client->connect_time;
1011 last_read_time = client->last_read_time;
1012 last_write_time = client->last_write_time;
1013
1014 last_read_cmd = client->last_read_cmd;
1015 last_write_cmd = client->last_write_cmd;
1016 }
1017
1018 vty_out(vty, "Connect Time: %s \n",
1019 zserv_time_buf(&connect_time, cbuf, ZEBRA_TIME_BUF));
1020 if (client->nh_reg_time) {
1021 vty_out(vty, "Nexthop Registry Time: %s \n",
1022 zserv_time_buf(&client->nh_reg_time, nhbuf,
1023 ZEBRA_TIME_BUF));
1024 if (client->nh_last_upd_time)
1025 vty_out(vty, "Nexthop Last Update Time: %s \n",
1026 zserv_time_buf(&client->nh_last_upd_time, mbuf,
1027 ZEBRA_TIME_BUF));
1028 else
1029 vty_out(vty, "No Nexthop Update sent\n");
1030 } else
1031 vty_out(vty, "Not registered for Nexthop Updates\n");
1032
1033 vty_out(vty,
1034 "Client will %sbe notified about the status of its routes.\n",
1035 client->notify_owner ? "" : "Not ");
1036
1037 vty_out(vty, "Last Msg Rx Time: %s \n",
1038 zserv_time_buf(&last_read_time, rbuf, ZEBRA_TIME_BUF));
1039 vty_out(vty, "Last Msg Tx Time: %s \n",
1040 zserv_time_buf(&last_write_time, wbuf, ZEBRA_TIME_BUF));
1041 if (last_read_cmd)
1042 vty_out(vty, "Last Rcvd Cmd: %s \n",
1043 zserv_command_string(last_read_cmd));
1044 if (last_write_cmd)
1045 vty_out(vty, "Last Sent Cmd: %s \n",
1046 zserv_command_string(last_write_cmd));
1047 vty_out(vty, "\n");
1048
1049 vty_out(vty, "Type Add Update Del \n");
1050 vty_out(vty, "================================================== \n");
1051 vty_out(vty, "IPv4 %-12u%-12u%-12u\n", client->v4_route_add_cnt,
1052 client->v4_route_upd8_cnt, client->v4_route_del_cnt);
1053 vty_out(vty, "IPv6 %-12u%-12u%-12u\n", client->v6_route_add_cnt,
1054 client->v6_route_upd8_cnt, client->v6_route_del_cnt);
1055 vty_out(vty, "Redist:v4 %-12u%-12u%-12u\n", client->redist_v4_add_cnt,
1056 0, client->redist_v4_del_cnt);
1057 vty_out(vty, "Redist:v6 %-12u%-12u%-12u\n", client->redist_v6_add_cnt,
1058 0, client->redist_v6_del_cnt);
1059 vty_out(vty, "VRF %-12u%-12u%-12u\n", client->vrfadd_cnt, 0,
1060 client->vrfdel_cnt);
1061 vty_out(vty, "Connected %-12u%-12u%-12u\n", client->ifadd_cnt, 0,
1062 client->ifdel_cnt);
1063 vty_out(vty, "Interface %-12u%-12u%-12u\n", client->ifup_cnt, 0,
1064 client->ifdown_cnt);
1065 vty_out(vty, "Intf Addr %-12u%-12u%-12u\n",
1066 client->connected_rt_add_cnt, 0, client->connected_rt_del_cnt);
1067 vty_out(vty, "BFD peer %-12u%-12u%-12u\n", client->bfd_peer_add_cnt,
1068 client->bfd_peer_upd8_cnt, client->bfd_peer_del_cnt);
1069 vty_out(vty, "NHT v4 %-12u%-12u%-12u\n",
1070 client->v4_nh_watch_add_cnt, 0, client->v4_nh_watch_rem_cnt);
1071 vty_out(vty, "NHT v6 %-12u%-12u%-12u\n",
1072 client->v6_nh_watch_add_cnt, 0, client->v6_nh_watch_rem_cnt);
1073 vty_out(vty, "VxLAN SG %-12u%-12u%-12u\n", client->vxlan_sg_add_cnt,
1074 0, client->vxlan_sg_del_cnt);
1075 vty_out(vty, "VNI %-12u%-12u%-12u\n", client->vniadd_cnt, 0,
1076 client->vnidel_cnt);
1077 vty_out(vty, "L3-VNI %-12u%-12u%-12u\n", client->l3vniadd_cnt, 0,
1078 client->l3vnidel_cnt);
1079 vty_out(vty, "MAC-IP %-12u%-12u%-12u\n", client->macipadd_cnt, 0,
1080 client->macipdel_cnt);
1081 vty_out(vty, "ES %-12u%-12u%-12u\n", client->local_es_add_cnt,
1082 0, client->local_es_del_cnt);
1083 vty_out(vty, "ES-EVI %-12u%-12u%-12u\n",
1084 client->local_es_evi_add_cnt, 0, client->local_es_evi_del_cnt);
1085 vty_out(vty, "Errors: %u\n", client->error_cnt);
1086
1087 #if defined DEV_BUILD
1088 vty_out(vty, "Input Fifo: %zu:%zu Output Fifo: %zu:%zu\n",
1089 client->ibuf_fifo->count, client->ibuf_fifo->max_count,
1090 client->obuf_fifo->count, client->obuf_fifo->max_count);
1091 #endif
1092 vty_out(vty, "\n");
1093 }
1094
1095 /* Display stale client information */
1096 static void zebra_show_stale_client_detail(struct vty *vty,
1097 struct zserv *client)
1098 {
1099 char buf[PREFIX2STR_BUFFER];
1100 time_t uptime;
1101 struct client_gr_info *info = NULL;
1102 struct zserv *s = NULL;
1103 bool first_p = true;
1104
1105 TAILQ_FOREACH (info, &client->gr_info_queue, gr_info) {
1106 if (first_p) {
1107 vty_out(vty, "Stale Client Information\n");
1108 vty_out(vty, "------------------------\n");
1109
1110 if (client->instance)
1111 vty_out(vty, " Instance: %u", client->instance);
1112 if (client->session_id)
1113 vty_out(vty, " [%u]", client->session_id);
1114
1115 first_p = false;
1116 }
1117
1118 vty_out(vty, "VRF : %s\n", vrf_id_to_name(info->vrf_id));
1119 vty_out(vty, "Capabilities : ");
1120 switch (info->capabilities) {
1121 case ZEBRA_CLIENT_GR_CAPABILITIES:
1122 vty_out(vty, "Graceful Restart(%u seconds)\n",
1123 info->stale_removal_time);
1124 break;
1125 case ZEBRA_CLIENT_ROUTE_UPDATE_COMPLETE:
1126 case ZEBRA_CLIENT_ROUTE_UPDATE_PENDING:
1127 case ZEBRA_CLIENT_GR_DISABLE:
1128 case ZEBRA_CLIENT_RIB_STALE_TIME:
1129 vty_out(vty, "None\n");
1130 break;
1131 }
1132
1133 if (ZEBRA_CLIENT_GR_ENABLED(info->capabilities)) {
1134 if (info->stale_client_ptr) {
1135 s = (struct zserv *)(info->stale_client_ptr);
1136 uptime = monotime(NULL);
1137 uptime -= s->restart_time;
1138
1139 frrtime_to_interval(uptime, buf, sizeof(buf));
1140
1141 vty_out(vty, "Last restart time : %s ago\n",
1142 buf);
1143
1144 vty_out(vty, "Stalepath removal time: %d sec\n",
1145 info->stale_removal_time);
1146 if (info->t_stale_removal) {
1147 vty_out(vty,
1148 "Stale delete timer: %ld sec\n",
1149 event_timer_remain_second(
1150 info->t_stale_removal));
1151 }
1152 }
1153 }
1154 }
1155 vty_out(vty, "\n");
1156 return;
1157 }
1158
1159 static void zebra_show_client_brief(struct vty *vty, struct zserv *client)
1160 {
1161 char client_string[80];
1162 char cbuf[ZEBRA_TIME_BUF], rbuf[ZEBRA_TIME_BUF];
1163 char wbuf[ZEBRA_TIME_BUF];
1164 time_t connect_time, last_read_time, last_write_time;
1165
1166 frr_with_mutex (&client->stats_mtx) {
1167 connect_time = client->connect_time;
1168 last_read_time = client->last_read_time;
1169 last_write_time = client->last_write_time;
1170 }
1171
1172 if (client->instance || client->session_id)
1173 snprintfrr(client_string, sizeof(client_string), "%s[%u:%u]",
1174 zebra_route_string(client->proto), client->instance,
1175 client->session_id);
1176 else
1177 snprintfrr(client_string, sizeof(client_string), "%s",
1178 zebra_route_string(client->proto));
1179
1180 vty_out(vty, "%-10s%12s %12s%12s %10d/%-10d %10d/%-10d\n",
1181 client_string,
1182 zserv_time_buf(&connect_time, cbuf, ZEBRA_TIME_BUF),
1183 zserv_time_buf(&last_read_time, rbuf, ZEBRA_TIME_BUF),
1184 zserv_time_buf(&last_write_time, wbuf, ZEBRA_TIME_BUF),
1185 client->v4_route_add_cnt + client->v4_route_upd8_cnt,
1186 client->v4_route_del_cnt,
1187 client->v6_route_add_cnt + client->v6_route_upd8_cnt,
1188 client->v6_route_del_cnt);
1189 }
1190
1191 /*
1192 * Common logic that searches the client list for a zapi client; this
1193 * MUST be called holding the client list mutex.
1194 */
1195 static struct zserv *find_client_internal(uint8_t proto,
1196 unsigned short instance,
1197 uint32_t session_id)
1198 {
1199 struct listnode *node, *nnode;
1200 struct zserv *client = NULL;
1201
1202 for (ALL_LIST_ELEMENTS(zrouter.client_list, node, nnode, client)) {
1203 if (client->proto == proto && client->instance == instance &&
1204 client->session_id == session_id)
1205 break;
1206 }
1207
1208 return client;
1209 }
1210
1211 /*
1212 * Public api that searches for a client session; this version is
1213 * used from the zebra main pthread.
1214 */
1215 struct zserv *zserv_find_client(uint8_t proto, unsigned short instance)
1216 {
1217 struct zserv *client;
1218
1219 frr_with_mutex (&client_mutex) {
1220 client = find_client_internal(proto, instance, 0);
1221 }
1222
1223 return client;
1224 }
1225
1226 /*
1227 * Retrieve a client by its protocol, instance number, and session id.
1228 */
1229 struct zserv *zserv_find_client_session(uint8_t proto, unsigned short instance,
1230 uint32_t session_id)
1231 {
1232 struct zserv *client;
1233
1234 frr_with_mutex (&client_mutex) {
1235 client = find_client_internal(proto, instance, session_id);
1236 }
1237
1238 return client;
1239
1240 }
1241
1242 /* This command is for debugging purpose. */
1243 DEFUN (show_zebra_client,
1244 show_zebra_client_cmd,
1245 "show zebra client",
1246 SHOW_STR
1247 ZEBRA_STR
1248 "Client information\n")
1249 {
1250 struct listnode *node;
1251 struct zserv *client;
1252
1253 for (ALL_LIST_ELEMENTS_RO(zrouter.client_list, node, client)) {
1254 zebra_show_client_detail(vty, client);
1255 /* Show GR info if present */
1256 zebra_show_stale_client_detail(vty, client);
1257 }
1258
1259 return CMD_SUCCESS;
1260 }
1261
1262 /* This command is for debugging purpose. */
1263 DEFUN (show_zebra_client_summary,
1264 show_zebra_client_summary_cmd,
1265 "show zebra client summary",
1266 SHOW_STR
1267 ZEBRA_STR
1268 "Client information brief\n"
1269 "Brief Summary\n")
1270 {
1271 struct listnode *node;
1272 struct zserv *client;
1273
1274 vty_out(vty,
1275 "Name Connect Time Last Read Last Write IPv4 Routes IPv6 Routes\n");
1276 vty_out(vty,
1277 "------------------------------------------------------------------------------------------\n");
1278
1279 for (ALL_LIST_ELEMENTS_RO(zrouter.client_list, node, client))
1280 zebra_show_client_brief(vty, client);
1281
1282 vty_out(vty, "Routes column shows (added+updated)/deleted\n");
1283 return CMD_SUCCESS;
1284 }
1285
1286 static int zserv_client_close_cb(struct zserv *closed_client)
1287 {
1288 struct listnode *node, *nnode;
1289 struct zserv *client = NULL;
1290
1291 for (ALL_LIST_ELEMENTS(zrouter.client_list, node, nnode, client)) {
1292 if (client->proto == closed_client->proto)
1293 continue;
1294
1295 zsend_client_close_notify(client, closed_client);
1296 }
1297
1298 return 0;
1299 }
1300
1301 void zserv_init(void)
1302 {
1303 /* Client list init. */
1304 zrouter.client_list = list_new();
1305 zrouter.stale_client_list = list_new();
1306
1307 /* Misc init. */
1308 zsock = -1;
1309 pthread_mutex_init(&client_mutex, NULL);
1310
1311 install_element(ENABLE_NODE, &show_zebra_client_cmd);
1312 install_element(ENABLE_NODE, &show_zebra_client_summary_cmd);
1313
1314 hook_register(zserv_client_close, zserv_client_close_cb);
1315 }