|  | // SPDX-License-Identifier: GPL-2.0-only | 
|  | #define _GNU_SOURCE | 
|  |  | 
|  | #include <arpa/inet.h> | 
|  | #include <bpf/bpf.h> | 
|  | #include <bpf/libbpf.h> | 
|  | #include <errno.h> | 
|  | #include <fcntl.h> | 
|  | #include <getopt.h> | 
|  | #include <linux/ethtool.h> | 
|  | #include <linux/hashtable.h> | 
|  | #include <linux/if_link.h> | 
|  | #include <linux/jhash.h> | 
|  | #include <linux/limits.h> | 
|  | #include <linux/list.h> | 
|  | #include <linux/sockios.h> | 
|  | #include <locale.h> | 
|  | #include <math.h> | 
|  | #include <net/if.h> | 
|  | #include <poll.h> | 
|  | #include <signal.h> | 
|  | #include <stdbool.h> | 
|  | #include <stdio.h> | 
|  | #include <stdlib.h> | 
|  | #include <string.h> | 
|  | #include <sys/ioctl.h> | 
|  | #include <sys/mman.h> | 
|  | #include <sys/signalfd.h> | 
|  | #include <sys/sysinfo.h> | 
|  | #include <sys/timerfd.h> | 
|  | #include <sys/utsname.h> | 
|  | #include <time.h> | 
|  | #include <unistd.h> | 
|  |  | 
|  | #include "bpf_util.h" | 
|  | #include "xdp_sample_user.h" | 
|  |  | 
|  | #define __sample_print(fmt, cond, ...)                                         \ | 
|  | ({                                                                     \ | 
|  | if (cond)                                                      \ | 
|  | printf(fmt, ##__VA_ARGS__);                            \ | 
|  | }) | 
|  |  | 
|  | #define print_always(fmt, ...) __sample_print(fmt, 1, ##__VA_ARGS__) | 
|  | #define print_default(fmt, ...)                                                \ | 
|  | __sample_print(fmt, sample_log_level & LL_DEFAULT, ##__VA_ARGS__) | 
|  | #define __print_err(err, fmt, ...)                                             \ | 
|  | ({                                                                     \ | 
|  | __sample_print(fmt, err > 0 || sample_log_level & LL_DEFAULT,  \ | 
|  | ##__VA_ARGS__);                                 \ | 
|  | sample_err_exp = sample_err_exp ? true : err > 0;              \ | 
|  | }) | 
|  | #define print_err(err, fmt, ...) __print_err(err, fmt, ##__VA_ARGS__) | 
|  |  | 
|  | #define __COLUMN(x) "%'10" x " %-13s" | 
|  | #define FMT_COLUMNf __COLUMN(".0f") | 
|  | #define FMT_COLUMNd __COLUMN("d") | 
|  | #define FMT_COLUMNl __COLUMN("llu") | 
|  | #define RX(rx) rx, "rx/s" | 
|  | #define PPS(pps) pps, "pkt/s" | 
|  | #define DROP(drop) drop, "drop/s" | 
|  | #define ERR(err) err, "error/s" | 
|  | #define HITS(hits) hits, "hit/s" | 
|  | #define XMIT(xmit) xmit, "xmit/s" | 
|  | #define PASS(pass) pass, "pass/s" | 
|  | #define REDIR(redir) redir, "redir/s" | 
|  | #define NANOSEC_PER_SEC 1000000000 /* 10^9 */ | 
|  |  | 
|  | #define XDP_UNKNOWN (XDP_REDIRECT + 1) | 
|  | #define XDP_ACTION_MAX (XDP_UNKNOWN + 1) | 
|  | #define XDP_REDIRECT_ERR_MAX 7 | 
|  |  | 
|  | enum map_type { | 
|  | MAP_RX, | 
|  | MAP_REDIRECT_ERR, | 
|  | MAP_CPUMAP_ENQUEUE, | 
|  | MAP_CPUMAP_KTHREAD, | 
|  | MAP_EXCEPTION, | 
|  | MAP_DEVMAP_XMIT, | 
|  | MAP_DEVMAP_XMIT_MULTI, | 
|  | NUM_MAP, | 
|  | }; | 
|  |  | 
|  | enum log_level { | 
|  | LL_DEFAULT = 1U << 0, | 
|  | LL_SIMPLE = 1U << 1, | 
|  | LL_DEBUG = 1U << 2, | 
|  | }; | 
|  |  | 
|  | struct record { | 
|  | __u64 timestamp; | 
|  | struct datarec total; | 
|  | struct datarec *cpu; | 
|  | }; | 
|  |  | 
|  | struct map_entry { | 
|  | struct hlist_node node; | 
|  | __u64 pair; | 
|  | struct record val; | 
|  | }; | 
|  |  | 
|  | struct stats_record { | 
|  | struct record rx_cnt; | 
|  | struct record redir_err[XDP_REDIRECT_ERR_MAX]; | 
|  | struct record kthread; | 
|  | struct record exception[XDP_ACTION_MAX]; | 
|  | struct record devmap_xmit; | 
|  | DECLARE_HASHTABLE(xmit_map, 5); | 
|  | struct record enq[]; | 
|  | }; | 
|  |  | 
|  | struct sample_output { | 
|  | struct { | 
|  | __u64 rx; | 
|  | __u64 redir; | 
|  | __u64 drop; | 
|  | __u64 drop_xmit; | 
|  | __u64 err; | 
|  | __u64 xmit; | 
|  | } totals; | 
|  | struct { | 
|  | union { | 
|  | __u64 pps; | 
|  | __u64 num; | 
|  | }; | 
|  | __u64 drop; | 
|  | __u64 err; | 
|  | } rx_cnt; | 
|  | struct { | 
|  | __u64 suc; | 
|  | __u64 err; | 
|  | } redir_cnt; | 
|  | struct { | 
|  | __u64 hits; | 
|  | } except_cnt; | 
|  | struct { | 
|  | __u64 pps; | 
|  | __u64 drop; | 
|  | __u64 err; | 
|  | double bavg; | 
|  | } xmit_cnt; | 
|  | }; | 
|  |  | 
|  | struct xdp_desc { | 
|  | int ifindex; | 
|  | __u32 prog_id; | 
|  | int flags; | 
|  | } sample_xdp_progs[32]; | 
|  |  | 
|  | struct datarec *sample_mmap[NUM_MAP]; | 
|  | struct bpf_map *sample_map[NUM_MAP]; | 
|  | size_t sample_map_count[NUM_MAP]; | 
|  | enum log_level sample_log_level; | 
|  | struct sample_output sample_out; | 
|  | unsigned long sample_interval; | 
|  | bool sample_err_exp; | 
|  | int sample_xdp_cnt; | 
|  | int sample_n_cpus; | 
|  | int sample_sig_fd; | 
|  | int sample_mask; | 
|  |  | 
|  | static const char *xdp_redirect_err_names[XDP_REDIRECT_ERR_MAX] = { | 
|  | /* Key=1 keeps unknown errors */ | 
|  | "Success", | 
|  | "Unknown", | 
|  | "EINVAL", | 
|  | "ENETDOWN", | 
|  | "EMSGSIZE", | 
|  | "EOPNOTSUPP", | 
|  | "ENOSPC", | 
|  | }; | 
|  |  | 
|  | /* Keyed from Unknown */ | 
|  | static const char *xdp_redirect_err_help[XDP_REDIRECT_ERR_MAX - 1] = { | 
|  | "Unknown error", | 
|  | "Invalid redirection", | 
|  | "Device being redirected to is down", | 
|  | "Packet length too large for device", | 
|  | "Operation not supported", | 
|  | "No space in ptr_ring of cpumap kthread", | 
|  | }; | 
|  |  | 
|  | static const char *xdp_action_names[XDP_ACTION_MAX] = { | 
|  | [XDP_ABORTED]  = "XDP_ABORTED", | 
|  | [XDP_DROP]     = "XDP_DROP", | 
|  | [XDP_PASS]     = "XDP_PASS", | 
|  | [XDP_TX]       = "XDP_TX", | 
|  | [XDP_REDIRECT] = "XDP_REDIRECT", | 
|  | [XDP_UNKNOWN]  = "XDP_UNKNOWN", | 
|  | }; | 
|  |  | 
|  | static __u64 gettime(void) | 
|  | { | 
|  | struct timespec t; | 
|  | int res; | 
|  |  | 
|  | res = clock_gettime(CLOCK_MONOTONIC, &t); | 
|  | if (res < 0) { | 
|  | fprintf(stderr, "Error with gettimeofday! (%i)\n", res); | 
|  | return UINT64_MAX; | 
|  | } | 
|  | return (__u64)t.tv_sec * NANOSEC_PER_SEC + t.tv_nsec; | 
|  | } | 
|  |  | 
|  | static const char *action2str(int action) | 
|  | { | 
|  | if (action < XDP_ACTION_MAX) | 
|  | return xdp_action_names[action]; | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static void sample_print_help(int mask) | 
|  | { | 
|  | printf("Output format description\n\n" | 
|  | "By default, redirect success statistics are disabled, use -s to enable.\n" | 
|  | "The terse output mode is default, verbose mode can be activated using -v\n" | 
|  | "Use SIGQUIT (Ctrl + \\) to switch the mode dynamically at runtime\n\n" | 
|  | "Terse mode displays at most the following fields:\n" | 
|  | "  rx/s        Number of packets received per second\n" | 
|  | "  redir/s     Number of packets successfully redirected per second\n" | 
|  | "  err,drop/s  Aggregated count of errors per second (including dropped packets)\n" | 
|  | "  xmit/s      Number of packets transmitted on the output device per second\n\n" | 
|  | "Output description for verbose mode:\n" | 
|  | "  FIELD                 DESCRIPTION\n"); | 
|  |  | 
|  | if (mask & SAMPLE_RX_CNT) { | 
|  | printf("  receive\t\tDisplays the number of packets received & errors encountered\n" | 
|  | " \t\t\tWhenever an error or packet drop occurs, details of per CPU error\n" | 
|  | " \t\t\tand drop statistics will be expanded inline in terse mode.\n" | 
|  | " \t\t\t\tpkt/s     - Packets received per second\n" | 
|  | " \t\t\t\tdrop/s    - Packets dropped per second\n" | 
|  | " \t\t\t\terror/s   - Errors encountered per second\n\n"); | 
|  | } | 
|  | if (mask & (SAMPLE_REDIRECT_CNT | SAMPLE_REDIRECT_ERR_CNT)) { | 
|  | printf("  redirect\t\tDisplays the number of packets successfully redirected\n" | 
|  | "  \t\t\tErrors encountered are expanded under redirect_err field\n" | 
|  | "  \t\t\tNote that passing -s to enable it has a per packet overhead\n" | 
|  | "  \t\t\t\tredir/s   - Packets redirected successfully per second\n\n" | 
|  | "  redirect_err\t\tDisplays the number of packets that failed redirection\n" | 
|  | "  \t\t\tThe errno is expanded under this field with per CPU count\n" | 
|  | "  \t\t\tThe recognized errors are:\n"); | 
|  |  | 
|  | for (int i = 2; i < XDP_REDIRECT_ERR_MAX; i++) | 
|  | printf("\t\t\t  %s: %s\n", xdp_redirect_err_names[i], | 
|  | xdp_redirect_err_help[i - 1]); | 
|  |  | 
|  | printf("  \n\t\t\t\terror/s   - Packets that failed redirection per second\n\n"); | 
|  | } | 
|  |  | 
|  | if (mask & SAMPLE_CPUMAP_ENQUEUE_CNT) { | 
|  | printf("  enqueue to cpu N\tDisplays the number of packets enqueued to bulk queue of CPU N\n" | 
|  | "  \t\t\tExpands to cpu:FROM->N to display enqueue stats for each CPU enqueuing to CPU N\n" | 
|  | "  \t\t\tReceived packets can be associated with the CPU redirect program is enqueuing \n" | 
|  | "  \t\t\tpackets to.\n" | 
|  | "  \t\t\t\tpkt/s    - Packets enqueued per second from other CPU to CPU N\n" | 
|  | "  \t\t\t\tdrop/s   - Packets dropped when trying to enqueue to CPU N\n" | 
|  | "  \t\t\t\tbulk-avg - Average number of packets processed for each event\n\n"); | 
|  | } | 
|  |  | 
|  | if (mask & SAMPLE_CPUMAP_KTHREAD_CNT) { | 
|  | printf("  kthread\t\tDisplays the number of packets processed in CPUMAP kthread for each CPU\n" | 
|  | "  \t\t\tPackets consumed from ptr_ring in kthread, and its xdp_stats (after calling \n" | 
|  | "  \t\t\tCPUMAP bpf prog) are expanded below this. xdp_stats are expanded as a total and\n" | 
|  | "  \t\t\tthen per-CPU to associate it to each CPU's pinned CPUMAP kthread.\n" | 
|  | "  \t\t\t\tpkt/s    - Packets consumed per second from ptr_ring\n" | 
|  | "  \t\t\t\tdrop/s   - Packets dropped per second in kthread\n" | 
|  | "  \t\t\t\tsched    - Number of times kthread called schedule()\n\n" | 
|  | "  \t\t\txdp_stats (also expands to per-CPU counts)\n" | 
|  | "  \t\t\t\tpass/s  - XDP_PASS count for CPUMAP program execution\n" | 
|  | "  \t\t\t\tdrop/s  - XDP_DROP count for CPUMAP program execution\n" | 
|  | "  \t\t\t\tredir/s - XDP_REDIRECT count for CPUMAP program execution\n\n"); | 
|  | } | 
|  |  | 
|  | if (mask & SAMPLE_EXCEPTION_CNT) { | 
|  | printf("  xdp_exception\t\tDisplays xdp_exception tracepoint events\n" | 
|  | "  \t\t\tThis can occur due to internal driver errors, unrecognized\n" | 
|  | "  \t\t\tXDP actions and due to explicit user trigger by use of XDP_ABORTED\n" | 
|  | "  \t\t\tEach action is expanded below this field with its count\n" | 
|  | "  \t\t\t\thit/s     - Number of times the tracepoint was hit per second\n\n"); | 
|  | } | 
|  |  | 
|  | if (mask & SAMPLE_DEVMAP_XMIT_CNT) { | 
|  | printf("  devmap_xmit\t\tDisplays devmap_xmit tracepoint events\n" | 
|  | "  \t\t\tThis tracepoint is invoked for successful transmissions on output\n" | 
|  | "  \t\t\tdevice but these statistics are not available for generic XDP mode,\n" | 
|  | "  \t\t\thence they will be omitted from the output when using SKB mode\n" | 
|  | "  \t\t\t\txmit/s    - Number of packets that were transmitted per second\n" | 
|  | "  \t\t\t\tdrop/s    - Number of packets that failed transmissions per second\n" | 
|  | "  \t\t\t\tdrv_err/s - Number of internal driver errors per second\n" | 
|  | "  \t\t\t\tbulk-avg  - Average number of packets processed for each event\n\n"); | 
|  | } | 
|  | } | 
|  |  | 
|  | void sample_usage(char *argv[], const struct option *long_options, | 
|  | const char *doc, int mask, bool error) | 
|  | { | 
|  | int i; | 
|  |  | 
|  | if (!error) | 
|  | sample_print_help(mask); | 
|  |  | 
|  | printf("\n%s\nOption for %s:\n", doc, argv[0]); | 
|  | for (i = 0; long_options[i].name != 0; i++) { | 
|  | printf(" --%-15s", long_options[i].name); | 
|  | if (long_options[i].flag != NULL) | 
|  | printf(" flag (internal value: %d)", | 
|  | *long_options[i].flag); | 
|  | else | 
|  | printf("\t short-option: -%c", long_options[i].val); | 
|  | printf("\n"); | 
|  | } | 
|  | printf("\n"); | 
|  | } | 
|  |  | 
|  | static struct datarec *alloc_record_per_cpu(void) | 
|  | { | 
|  | unsigned int nr_cpus = libbpf_num_possible_cpus(); | 
|  | struct datarec *array; | 
|  |  | 
|  | array = calloc(nr_cpus, sizeof(*array)); | 
|  | if (!array) { | 
|  | fprintf(stderr, "Failed to allocate memory (nr_cpus: %u)\n", | 
|  | nr_cpus); | 
|  | return NULL; | 
|  | } | 
|  | return array; | 
|  | } | 
|  |  | 
|  | static int map_entry_init(struct map_entry *e, __u64 pair) | 
|  | { | 
|  | e->pair = pair; | 
|  | INIT_HLIST_NODE(&e->node); | 
|  | e->val.timestamp = gettime(); | 
|  | e->val.cpu = alloc_record_per_cpu(); | 
|  | if (!e->val.cpu) | 
|  | return -ENOMEM; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void map_collect_percpu(struct datarec *values, struct record *rec) | 
|  | { | 
|  | /* For percpu maps, userspace gets a value per possible CPU */ | 
|  | unsigned int nr_cpus = libbpf_num_possible_cpus(); | 
|  | __u64 sum_xdp_redirect = 0; | 
|  | __u64 sum_processed = 0; | 
|  | __u64 sum_xdp_pass = 0; | 
|  | __u64 sum_xdp_drop = 0; | 
|  | __u64 sum_dropped = 0; | 
|  | __u64 sum_issue = 0; | 
|  | int i; | 
|  |  | 
|  | /* Get time as close as possible to reading map contents */ | 
|  | rec->timestamp = gettime(); | 
|  |  | 
|  | /* Record and sum values from each CPU */ | 
|  | for (i = 0; i < nr_cpus; i++) { | 
|  | rec->cpu[i].processed = READ_ONCE(values[i].processed); | 
|  | rec->cpu[i].dropped = READ_ONCE(values[i].dropped); | 
|  | rec->cpu[i].issue = READ_ONCE(values[i].issue); | 
|  | rec->cpu[i].xdp_pass = READ_ONCE(values[i].xdp_pass); | 
|  | rec->cpu[i].xdp_drop = READ_ONCE(values[i].xdp_drop); | 
|  | rec->cpu[i].xdp_redirect = READ_ONCE(values[i].xdp_redirect); | 
|  |  | 
|  | sum_processed += rec->cpu[i].processed; | 
|  | sum_dropped += rec->cpu[i].dropped; | 
|  | sum_issue += rec->cpu[i].issue; | 
|  | sum_xdp_pass += rec->cpu[i].xdp_pass; | 
|  | sum_xdp_drop += rec->cpu[i].xdp_drop; | 
|  | sum_xdp_redirect += rec->cpu[i].xdp_redirect; | 
|  | } | 
|  |  | 
|  | rec->total.processed = sum_processed; | 
|  | rec->total.dropped = sum_dropped; | 
|  | rec->total.issue = sum_issue; | 
|  | rec->total.xdp_pass = sum_xdp_pass; | 
|  | rec->total.xdp_drop = sum_xdp_drop; | 
|  | rec->total.xdp_redirect = sum_xdp_redirect; | 
|  | } | 
|  |  | 
|  | static int map_collect_percpu_devmap(int map_fd, struct stats_record *rec) | 
|  | { | 
|  | unsigned int nr_cpus = bpf_num_possible_cpus(); | 
|  | __u32 batch, count = 32; | 
|  | struct datarec *values; | 
|  | bool init = false; | 
|  | __u64 *keys; | 
|  | int i, ret; | 
|  |  | 
|  | keys = calloc(count, sizeof(__u64)); | 
|  | if (!keys) | 
|  | return -ENOMEM; | 
|  | values = calloc(count * nr_cpus, sizeof(struct datarec)); | 
|  | if (!values) { | 
|  | free(keys); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | for (;;) { | 
|  | bool exit = false; | 
|  |  | 
|  | ret = bpf_map_lookup_batch(map_fd, init ? &batch : NULL, &batch, | 
|  | keys, values, &count, NULL); | 
|  | if (ret < 0 && errno != ENOENT) | 
|  | break; | 
|  | if (errno == ENOENT) | 
|  | exit = true; | 
|  |  | 
|  | init = true; | 
|  | for (i = 0; i < count; i++) { | 
|  | struct map_entry *e, *x = NULL; | 
|  | __u64 pair = keys[i]; | 
|  | struct datarec *arr; | 
|  |  | 
|  | arr = &values[i * nr_cpus]; | 
|  | hash_for_each_possible(rec->xmit_map, e, node, pair) { | 
|  | if (e->pair == pair) { | 
|  | x = e; | 
|  | break; | 
|  | } | 
|  | } | 
|  | if (!x) { | 
|  | x = calloc(1, sizeof(*x)); | 
|  | if (!x) | 
|  | goto cleanup; | 
|  | if (map_entry_init(x, pair) < 0) { | 
|  | free(x); | 
|  | goto cleanup; | 
|  | } | 
|  | hash_add(rec->xmit_map, &x->node, pair); | 
|  | } | 
|  | map_collect_percpu(arr, &x->val); | 
|  | } | 
|  |  | 
|  | if (exit) | 
|  | break; | 
|  | count = 32; | 
|  | } | 
|  |  | 
|  | free(values); | 
|  | free(keys); | 
|  | return 0; | 
|  | cleanup: | 
|  | free(values); | 
|  | free(keys); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | static struct stats_record *alloc_stats_record(void) | 
|  | { | 
|  | struct stats_record *rec; | 
|  | int i; | 
|  |  | 
|  | rec = calloc(1, sizeof(*rec) + sample_n_cpus * sizeof(struct record)); | 
|  | if (!rec) { | 
|  | fprintf(stderr, "Failed to allocate memory\n"); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | if (sample_mask & SAMPLE_RX_CNT) { | 
|  | rec->rx_cnt.cpu = alloc_record_per_cpu(); | 
|  | if (!rec->rx_cnt.cpu) { | 
|  | fprintf(stderr, | 
|  | "Failed to allocate rx_cnt per-CPU array\n"); | 
|  | goto end_rec; | 
|  | } | 
|  | } | 
|  | if (sample_mask & (SAMPLE_REDIRECT_CNT | SAMPLE_REDIRECT_ERR_CNT)) { | 
|  | for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++) { | 
|  | rec->redir_err[i].cpu = alloc_record_per_cpu(); | 
|  | if (!rec->redir_err[i].cpu) { | 
|  | fprintf(stderr, | 
|  | "Failed to allocate redir_err per-CPU array for " | 
|  | "\"%s\" case\n", | 
|  | xdp_redirect_err_names[i]); | 
|  | while (i--) | 
|  | free(rec->redir_err[i].cpu); | 
|  | goto end_rx_cnt; | 
|  | } | 
|  | } | 
|  | } | 
|  | if (sample_mask & SAMPLE_CPUMAP_KTHREAD_CNT) { | 
|  | rec->kthread.cpu = alloc_record_per_cpu(); | 
|  | if (!rec->kthread.cpu) { | 
|  | fprintf(stderr, | 
|  | "Failed to allocate kthread per-CPU array\n"); | 
|  | goto end_redir; | 
|  | } | 
|  | } | 
|  | if (sample_mask & SAMPLE_EXCEPTION_CNT) { | 
|  | for (i = 0; i < XDP_ACTION_MAX; i++) { | 
|  | rec->exception[i].cpu = alloc_record_per_cpu(); | 
|  | if (!rec->exception[i].cpu) { | 
|  | fprintf(stderr, | 
|  | "Failed to allocate exception per-CPU array for " | 
|  | "\"%s\" case\n", | 
|  | action2str(i)); | 
|  | while (i--) | 
|  | free(rec->exception[i].cpu); | 
|  | goto end_kthread; | 
|  | } | 
|  | } | 
|  | } | 
|  | if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT) { | 
|  | rec->devmap_xmit.cpu = alloc_record_per_cpu(); | 
|  | if (!rec->devmap_xmit.cpu) { | 
|  | fprintf(stderr, | 
|  | "Failed to allocate devmap_xmit per-CPU array\n"); | 
|  | goto end_exception; | 
|  | } | 
|  | } | 
|  | if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI) | 
|  | hash_init(rec->xmit_map); | 
|  | if (sample_mask & SAMPLE_CPUMAP_ENQUEUE_CNT) { | 
|  | for (i = 0; i < sample_n_cpus; i++) { | 
|  | rec->enq[i].cpu = alloc_record_per_cpu(); | 
|  | if (!rec->enq[i].cpu) { | 
|  | fprintf(stderr, | 
|  | "Failed to allocate enqueue per-CPU array for " | 
|  | "CPU %d\n", | 
|  | i); | 
|  | while (i--) | 
|  | free(rec->enq[i].cpu); | 
|  | goto end_devmap_xmit; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | return rec; | 
|  |  | 
|  | end_devmap_xmit: | 
|  | free(rec->devmap_xmit.cpu); | 
|  | end_exception: | 
|  | for (i = 0; i < XDP_ACTION_MAX; i++) | 
|  | free(rec->exception[i].cpu); | 
|  | end_kthread: | 
|  | free(rec->kthread.cpu); | 
|  | end_redir: | 
|  | for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++) | 
|  | free(rec->redir_err[i].cpu); | 
|  | end_rx_cnt: | 
|  | free(rec->rx_cnt.cpu); | 
|  | end_rec: | 
|  | free(rec); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static void free_stats_record(struct stats_record *r) | 
|  | { | 
|  | struct hlist_node *tmp; | 
|  | struct map_entry *e; | 
|  | int i; | 
|  |  | 
|  | for (i = 0; i < sample_n_cpus; i++) | 
|  | free(r->enq[i].cpu); | 
|  | hash_for_each_safe(r->xmit_map, i, tmp, e, node) { | 
|  | hash_del(&e->node); | 
|  | free(e->val.cpu); | 
|  | free(e); | 
|  | } | 
|  | free(r->devmap_xmit.cpu); | 
|  | for (i = 0; i < XDP_ACTION_MAX; i++) | 
|  | free(r->exception[i].cpu); | 
|  | free(r->kthread.cpu); | 
|  | for (i = 0; i < XDP_REDIRECT_ERR_MAX; i++) | 
|  | free(r->redir_err[i].cpu); | 
|  | free(r->rx_cnt.cpu); | 
|  | free(r); | 
|  | } | 
|  |  | 
|  | static double calc_period(struct record *r, struct record *p) | 
|  | { | 
|  | double period_ = 0; | 
|  | __u64 period = 0; | 
|  |  | 
|  | period = r->timestamp - p->timestamp; | 
|  | if (period > 0) | 
|  | period_ = ((double)period / NANOSEC_PER_SEC); | 
|  |  | 
|  | return period_; | 
|  | } | 
|  |  | 
|  | static double sample_round(double val) | 
|  | { | 
|  | if (val - floor(val) < 0.5) | 
|  | return floor(val); | 
|  | return ceil(val); | 
|  | } | 
|  |  | 
|  | static __u64 calc_pps(struct datarec *r, struct datarec *p, double period_) | 
|  | { | 
|  | __u64 packets = 0; | 
|  | __u64 pps = 0; | 
|  |  | 
|  | if (period_ > 0) { | 
|  | packets = r->processed - p->processed; | 
|  | pps = sample_round(packets / period_); | 
|  | } | 
|  | return pps; | 
|  | } | 
|  |  | 
|  | static __u64 calc_drop_pps(struct datarec *r, struct datarec *p, double period_) | 
|  | { | 
|  | __u64 packets = 0; | 
|  | __u64 pps = 0; | 
|  |  | 
|  | if (period_ > 0) { | 
|  | packets = r->dropped - p->dropped; | 
|  | pps = sample_round(packets / period_); | 
|  | } | 
|  | return pps; | 
|  | } | 
|  |  | 
|  | static __u64 calc_errs_pps(struct datarec *r, struct datarec *p, double period_) | 
|  | { | 
|  | __u64 packets = 0; | 
|  | __u64 pps = 0; | 
|  |  | 
|  | if (period_ > 0) { | 
|  | packets = r->issue - p->issue; | 
|  | pps = sample_round(packets / period_); | 
|  | } | 
|  | return pps; | 
|  | } | 
|  |  | 
|  | static __u64 calc_info_pps(struct datarec *r, struct datarec *p, double period_) | 
|  | { | 
|  | __u64 packets = 0; | 
|  | __u64 pps = 0; | 
|  |  | 
|  | if (period_ > 0) { | 
|  | packets = r->info - p->info; | 
|  | pps = sample_round(packets / period_); | 
|  | } | 
|  | return pps; | 
|  | } | 
|  |  | 
|  | static void calc_xdp_pps(struct datarec *r, struct datarec *p, double *xdp_pass, | 
|  | double *xdp_drop, double *xdp_redirect, double period_) | 
|  | { | 
|  | *xdp_pass = 0, *xdp_drop = 0, *xdp_redirect = 0; | 
|  | if (period_ > 0) { | 
|  | *xdp_redirect = (r->xdp_redirect - p->xdp_redirect) / period_; | 
|  | *xdp_pass = (r->xdp_pass - p->xdp_pass) / period_; | 
|  | *xdp_drop = (r->xdp_drop - p->xdp_drop) / period_; | 
|  | } | 
|  | } | 
|  |  | 
|  | static void stats_get_rx_cnt(struct stats_record *stats_rec, | 
|  | struct stats_record *stats_prev, | 
|  | unsigned int nr_cpus, struct sample_output *out) | 
|  | { | 
|  | struct record *rec, *prev; | 
|  | double t, pps, drop, err; | 
|  | int i; | 
|  |  | 
|  | rec = &stats_rec->rx_cnt; | 
|  | prev = &stats_prev->rx_cnt; | 
|  | t = calc_period(rec, prev); | 
|  |  | 
|  | for (i = 0; i < nr_cpus; i++) { | 
|  | struct datarec *r = &rec->cpu[i]; | 
|  | struct datarec *p = &prev->cpu[i]; | 
|  | char str[64]; | 
|  |  | 
|  | pps = calc_pps(r, p, t); | 
|  | drop = calc_drop_pps(r, p, t); | 
|  | err = calc_errs_pps(r, p, t); | 
|  | if (!pps && !drop && !err) | 
|  | continue; | 
|  |  | 
|  | snprintf(str, sizeof(str), "cpu:%d", i); | 
|  | print_default("    %-18s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf | 
|  | "\n", | 
|  | str, PPS(pps), DROP(drop), ERR(err)); | 
|  | } | 
|  |  | 
|  | if (out) { | 
|  | pps = calc_pps(&rec->total, &prev->total, t); | 
|  | drop = calc_drop_pps(&rec->total, &prev->total, t); | 
|  | err = calc_errs_pps(&rec->total, &prev->total, t); | 
|  |  | 
|  | out->rx_cnt.pps = pps; | 
|  | out->rx_cnt.drop = drop; | 
|  | out->rx_cnt.err = err; | 
|  | out->totals.rx += pps; | 
|  | out->totals.drop += drop; | 
|  | out->totals.err += err; | 
|  | } | 
|  | } | 
|  |  | 
|  | static void stats_get_cpumap_enqueue(struct stats_record *stats_rec, | 
|  | struct stats_record *stats_prev, | 
|  | unsigned int nr_cpus) | 
|  | { | 
|  | struct record *rec, *prev; | 
|  | double t, pps, drop, err; | 
|  | int i, to_cpu; | 
|  |  | 
|  | /* cpumap enqueue stats */ | 
|  | for (to_cpu = 0; to_cpu < sample_n_cpus; to_cpu++) { | 
|  | rec = &stats_rec->enq[to_cpu]; | 
|  | prev = &stats_prev->enq[to_cpu]; | 
|  | t = calc_period(rec, prev); | 
|  |  | 
|  | pps = calc_pps(&rec->total, &prev->total, t); | 
|  | drop = calc_drop_pps(&rec->total, &prev->total, t); | 
|  | err = calc_errs_pps(&rec->total, &prev->total, t); | 
|  |  | 
|  | if (pps > 0 || drop > 0) { | 
|  | char str[64]; | 
|  |  | 
|  | snprintf(str, sizeof(str), "enqueue to cpu %d", to_cpu); | 
|  |  | 
|  | if (err > 0) | 
|  | err = pps / err; /* calc average bulk size */ | 
|  |  | 
|  | print_err(drop, | 
|  | "  %-20s " FMT_COLUMNf FMT_COLUMNf __COLUMN( | 
|  | ".2f") "\n", | 
|  | str, PPS(pps), DROP(drop), err, "bulk-avg"); | 
|  | } | 
|  |  | 
|  | for (i = 0; i < nr_cpus; i++) { | 
|  | struct datarec *r = &rec->cpu[i]; | 
|  | struct datarec *p = &prev->cpu[i]; | 
|  | char str[64]; | 
|  |  | 
|  | pps = calc_pps(r, p, t); | 
|  | drop = calc_drop_pps(r, p, t); | 
|  | err = calc_errs_pps(r, p, t); | 
|  | if (!pps && !drop && !err) | 
|  | continue; | 
|  |  | 
|  | snprintf(str, sizeof(str), "cpu:%d->%d", i, to_cpu); | 
|  | if (err > 0) | 
|  | err = pps / err; /* calc average bulk size */ | 
|  | print_default( | 
|  | "    %-18s " FMT_COLUMNf FMT_COLUMNf __COLUMN( | 
|  | ".2f") "\n", | 
|  | str, PPS(pps), DROP(drop), err, "bulk-avg"); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | static void stats_get_cpumap_remote(struct stats_record *stats_rec, | 
|  | struct stats_record *stats_prev, | 
|  | unsigned int nr_cpus) | 
|  | { | 
|  | double xdp_pass, xdp_drop, xdp_redirect; | 
|  | struct record *rec, *prev; | 
|  | double t; | 
|  | int i; | 
|  |  | 
|  | rec = &stats_rec->kthread; | 
|  | prev = &stats_prev->kthread; | 
|  | t = calc_period(rec, prev); | 
|  |  | 
|  | calc_xdp_pps(&rec->total, &prev->total, &xdp_pass, &xdp_drop, | 
|  | &xdp_redirect, t); | 
|  | if (xdp_pass || xdp_drop || xdp_redirect) { | 
|  | print_err(xdp_drop, | 
|  | "    %-18s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf "\n", | 
|  | "xdp_stats", PASS(xdp_pass), DROP(xdp_drop), | 
|  | REDIR(xdp_redirect)); | 
|  | } | 
|  |  | 
|  | for (i = 0; i < nr_cpus; i++) { | 
|  | struct datarec *r = &rec->cpu[i]; | 
|  | struct datarec *p = &prev->cpu[i]; | 
|  | char str[64]; | 
|  |  | 
|  | calc_xdp_pps(r, p, &xdp_pass, &xdp_drop, &xdp_redirect, t); | 
|  | if (!xdp_pass && !xdp_drop && !xdp_redirect) | 
|  | continue; | 
|  |  | 
|  | snprintf(str, sizeof(str), "cpu:%d", i); | 
|  | print_default("      %-16s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf | 
|  | "\n", | 
|  | str, PASS(xdp_pass), DROP(xdp_drop), | 
|  | REDIR(xdp_redirect)); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void stats_get_cpumap_kthread(struct stats_record *stats_rec, | 
|  | struct stats_record *stats_prev, | 
|  | unsigned int nr_cpus) | 
|  | { | 
|  | struct record *rec, *prev; | 
|  | double t, pps, drop, err; | 
|  | int i; | 
|  |  | 
|  | rec = &stats_rec->kthread; | 
|  | prev = &stats_prev->kthread; | 
|  | t = calc_period(rec, prev); | 
|  |  | 
|  | pps = calc_pps(&rec->total, &prev->total, t); | 
|  | drop = calc_drop_pps(&rec->total, &prev->total, t); | 
|  | err = calc_errs_pps(&rec->total, &prev->total, t); | 
|  |  | 
|  | print_err(drop, "  %-20s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf "\n", | 
|  | pps ? "kthread total" : "kthread", PPS(pps), DROP(drop), err, | 
|  | "sched"); | 
|  |  | 
|  | for (i = 0; i < nr_cpus; i++) { | 
|  | struct datarec *r = &rec->cpu[i]; | 
|  | struct datarec *p = &prev->cpu[i]; | 
|  | char str[64]; | 
|  |  | 
|  | pps = calc_pps(r, p, t); | 
|  | drop = calc_drop_pps(r, p, t); | 
|  | err = calc_errs_pps(r, p, t); | 
|  | if (!pps && !drop && !err) | 
|  | continue; | 
|  |  | 
|  | snprintf(str, sizeof(str), "cpu:%d", i); | 
|  | print_default("    %-18s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf | 
|  | "\n", | 
|  | str, PPS(pps), DROP(drop), err, "sched"); | 
|  | } | 
|  | } | 
|  |  | 
|  | static void stats_get_redirect_cnt(struct stats_record *stats_rec, | 
|  | struct stats_record *stats_prev, | 
|  | unsigned int nr_cpus, | 
|  | struct sample_output *out) | 
|  | { | 
|  | struct record *rec, *prev; | 
|  | double t, pps; | 
|  | int i; | 
|  |  | 
|  | rec = &stats_rec->redir_err[0]; | 
|  | prev = &stats_prev->redir_err[0]; | 
|  | t = calc_period(rec, prev); | 
|  | for (i = 0; i < nr_cpus; i++) { | 
|  | struct datarec *r = &rec->cpu[i]; | 
|  | struct datarec *p = &prev->cpu[i]; | 
|  | char str[64]; | 
|  |  | 
|  | pps = calc_pps(r, p, t); | 
|  | if (!pps) | 
|  | continue; | 
|  |  | 
|  | snprintf(str, sizeof(str), "cpu:%d", i); | 
|  | print_default("    %-18s " FMT_COLUMNf "\n", str, REDIR(pps)); | 
|  | } | 
|  |  | 
|  | if (out) { | 
|  | pps = calc_pps(&rec->total, &prev->total, t); | 
|  | out->redir_cnt.suc = pps; | 
|  | out->totals.redir += pps; | 
|  | } | 
|  | } | 
|  |  | 
|  | static void stats_get_redirect_err_cnt(struct stats_record *stats_rec, | 
|  | struct stats_record *stats_prev, | 
|  | unsigned int nr_cpus, | 
|  | struct sample_output *out) | 
|  | { | 
|  | struct record *rec, *prev; | 
|  | double t, drop, sum = 0; | 
|  | int rec_i, i; | 
|  |  | 
|  | for (rec_i = 1; rec_i < XDP_REDIRECT_ERR_MAX; rec_i++) { | 
|  | char str[64]; | 
|  |  | 
|  | rec = &stats_rec->redir_err[rec_i]; | 
|  | prev = &stats_prev->redir_err[rec_i]; | 
|  | t = calc_period(rec, prev); | 
|  |  | 
|  | drop = calc_drop_pps(&rec->total, &prev->total, t); | 
|  | if (drop > 0 && !out) { | 
|  | snprintf(str, sizeof(str), | 
|  | sample_log_level & LL_DEFAULT ? "%s total" : | 
|  | "%s", | 
|  | xdp_redirect_err_names[rec_i]); | 
|  | print_err(drop, "    %-18s " FMT_COLUMNf "\n", str, | 
|  | ERR(drop)); | 
|  | } | 
|  |  | 
|  | for (i = 0; i < nr_cpus; i++) { | 
|  | struct datarec *r = &rec->cpu[i]; | 
|  | struct datarec *p = &prev->cpu[i]; | 
|  | double drop; | 
|  |  | 
|  | drop = calc_drop_pps(r, p, t); | 
|  | if (!drop) | 
|  | continue; | 
|  |  | 
|  | snprintf(str, sizeof(str), "cpu:%d", i); | 
|  | print_default("       %-16s" FMT_COLUMNf "\n", str, | 
|  | ERR(drop)); | 
|  | } | 
|  |  | 
|  | sum += drop; | 
|  | } | 
|  |  | 
|  | if (out) { | 
|  | out->redir_cnt.err = sum; | 
|  | out->totals.err += sum; | 
|  | } | 
|  | } | 
|  |  | 
|  | static void stats_get_exception_cnt(struct stats_record *stats_rec, | 
|  | struct stats_record *stats_prev, | 
|  | unsigned int nr_cpus, | 
|  | struct sample_output *out) | 
|  | { | 
|  | double t, drop, sum = 0; | 
|  | struct record *rec, *prev; | 
|  | int rec_i, i; | 
|  |  | 
|  | for (rec_i = 0; rec_i < XDP_ACTION_MAX; rec_i++) { | 
|  | rec = &stats_rec->exception[rec_i]; | 
|  | prev = &stats_prev->exception[rec_i]; | 
|  | t = calc_period(rec, prev); | 
|  |  | 
|  | drop = calc_drop_pps(&rec->total, &prev->total, t); | 
|  | /* Fold out errors after heading */ | 
|  | sum += drop; | 
|  |  | 
|  | if (drop > 0 && !out) { | 
|  | print_always("    %-18s " FMT_COLUMNf "\n", | 
|  | action2str(rec_i), ERR(drop)); | 
|  |  | 
|  | for (i = 0; i < nr_cpus; i++) { | 
|  | struct datarec *r = &rec->cpu[i]; | 
|  | struct datarec *p = &prev->cpu[i]; | 
|  | char str[64]; | 
|  | double drop; | 
|  |  | 
|  | drop = calc_drop_pps(r, p, t); | 
|  | if (!drop) | 
|  | continue; | 
|  |  | 
|  | snprintf(str, sizeof(str), "cpu:%d", i); | 
|  | print_default("       %-16s" FMT_COLUMNf "\n", | 
|  | str, ERR(drop)); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | if (out) { | 
|  | out->except_cnt.hits = sum; | 
|  | out->totals.err += sum; | 
|  | } | 
|  | } | 
|  |  | 
|  | static void stats_get_devmap_xmit(struct stats_record *stats_rec, | 
|  | struct stats_record *stats_prev, | 
|  | unsigned int nr_cpus, | 
|  | struct sample_output *out) | 
|  | { | 
|  | double pps, drop, info, err; | 
|  | struct record *rec, *prev; | 
|  | double t; | 
|  | int i; | 
|  |  | 
|  | rec = &stats_rec->devmap_xmit; | 
|  | prev = &stats_prev->devmap_xmit; | 
|  | t = calc_period(rec, prev); | 
|  | for (i = 0; i < nr_cpus; i++) { | 
|  | struct datarec *r = &rec->cpu[i]; | 
|  | struct datarec *p = &prev->cpu[i]; | 
|  | char str[64]; | 
|  |  | 
|  | pps = calc_pps(r, p, t); | 
|  | drop = calc_drop_pps(r, p, t); | 
|  | err = calc_errs_pps(r, p, t); | 
|  |  | 
|  | if (!pps && !drop && !err) | 
|  | continue; | 
|  |  | 
|  | snprintf(str, sizeof(str), "cpu:%d", i); | 
|  | info = calc_info_pps(r, p, t); | 
|  | if (info > 0) | 
|  | info = (pps + drop) / info; /* calc avg bulk */ | 
|  | print_default("     %-18s" FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf | 
|  | __COLUMN(".2f") "\n", | 
|  | str, XMIT(pps), DROP(drop), err, "drv_err/s", | 
|  | info, "bulk-avg"); | 
|  | } | 
|  | if (out) { | 
|  | pps = calc_pps(&rec->total, &prev->total, t); | 
|  | drop = calc_drop_pps(&rec->total, &prev->total, t); | 
|  | info = calc_info_pps(&rec->total, &prev->total, t); | 
|  | if (info > 0) | 
|  | info = (pps + drop) / info; /* calc avg bulk */ | 
|  | err = calc_errs_pps(&rec->total, &prev->total, t); | 
|  |  | 
|  | out->xmit_cnt.pps = pps; | 
|  | out->xmit_cnt.drop = drop; | 
|  | out->xmit_cnt.bavg = info; | 
|  | out->xmit_cnt.err = err; | 
|  | out->totals.xmit += pps; | 
|  | out->totals.drop_xmit += drop; | 
|  | out->totals.err += err; | 
|  | } | 
|  | } | 
|  |  | 
|  | static void stats_get_devmap_xmit_multi(struct stats_record *stats_rec, | 
|  | struct stats_record *stats_prev, | 
|  | unsigned int nr_cpus, | 
|  | struct sample_output *out, | 
|  | bool xmit_total) | 
|  | { | 
|  | double pps, drop, info, err; | 
|  | struct map_entry *entry; | 
|  | struct record *r, *p; | 
|  | double t; | 
|  | int bkt; | 
|  |  | 
|  | hash_for_each(stats_rec->xmit_map, bkt, entry, node) { | 
|  | struct map_entry *e, *x = NULL; | 
|  | char ifname_from[IFNAMSIZ]; | 
|  | char ifname_to[IFNAMSIZ]; | 
|  | const char *fstr, *tstr; | 
|  | unsigned long prev_time; | 
|  | struct record beg = {}; | 
|  | __u32 from_idx, to_idx; | 
|  | char str[128]; | 
|  | __u64 pair; | 
|  | int i; | 
|  |  | 
|  | prev_time = sample_interval * NANOSEC_PER_SEC; | 
|  |  | 
|  | pair = entry->pair; | 
|  | from_idx = pair >> 32; | 
|  | to_idx = pair & 0xFFFFFFFF; | 
|  |  | 
|  | r = &entry->val; | 
|  | beg.timestamp = r->timestamp - prev_time; | 
|  |  | 
|  | /* Find matching entry from stats_prev map */ | 
|  | hash_for_each_possible(stats_prev->xmit_map, e, node, pair) { | 
|  | if (e->pair == pair) { | 
|  | x = e; | 
|  | break; | 
|  | } | 
|  | } | 
|  | if (x) | 
|  | p = &x->val; | 
|  | else | 
|  | p = &beg; | 
|  | t = calc_period(r, p); | 
|  | pps = calc_pps(&r->total, &p->total, t); | 
|  | drop = calc_drop_pps(&r->total, &p->total, t); | 
|  | info = calc_info_pps(&r->total, &p->total, t); | 
|  | if (info > 0) | 
|  | info = (pps + drop) / info; /* calc avg bulk */ | 
|  | err = calc_errs_pps(&r->total, &p->total, t); | 
|  |  | 
|  | if (out) { | 
|  | /* We are responsible for filling out totals */ | 
|  | out->totals.xmit += pps; | 
|  | out->totals.drop_xmit += drop; | 
|  | out->totals.err += err; | 
|  | continue; | 
|  | } | 
|  |  | 
|  | fstr = tstr = NULL; | 
|  | if (if_indextoname(from_idx, ifname_from)) | 
|  | fstr = ifname_from; | 
|  | if (if_indextoname(to_idx, ifname_to)) | 
|  | tstr = ifname_to; | 
|  |  | 
|  | snprintf(str, sizeof(str), "xmit %s->%s", fstr ?: "?", | 
|  | tstr ?: "?"); | 
|  | /* Skip idle streams of redirection */ | 
|  | if (pps || drop || err) { | 
|  | print_err(drop, | 
|  | "  %-20s " FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf | 
|  | __COLUMN(".2f") "\n", str, XMIT(pps), DROP(drop), | 
|  | err, "drv_err/s", info, "bulk-avg"); | 
|  | } | 
|  |  | 
|  | for (i = 0; i < nr_cpus; i++) { | 
|  | struct datarec *rc = &r->cpu[i]; | 
|  | struct datarec *pc, p_beg = {}; | 
|  | char str[64]; | 
|  |  | 
|  | pc = p == &beg ? &p_beg : &p->cpu[i]; | 
|  |  | 
|  | pps = calc_pps(rc, pc, t); | 
|  | drop = calc_drop_pps(rc, pc, t); | 
|  | err = calc_errs_pps(rc, pc, t); | 
|  |  | 
|  | if (!pps && !drop && !err) | 
|  | continue; | 
|  |  | 
|  | snprintf(str, sizeof(str), "cpu:%d", i); | 
|  | info = calc_info_pps(rc, pc, t); | 
|  | if (info > 0) | 
|  | info = (pps + drop) / info; /* calc avg bulk */ | 
|  |  | 
|  | print_default("     %-18s" FMT_COLUMNf FMT_COLUMNf FMT_COLUMNf | 
|  | __COLUMN(".2f") "\n", str, XMIT(pps), | 
|  | DROP(drop), err, "drv_err/s", info, "bulk-avg"); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | static void stats_print(const char *prefix, int mask, struct stats_record *r, | 
|  | struct stats_record *p, struct sample_output *out) | 
|  | { | 
|  | int nr_cpus = libbpf_num_possible_cpus(); | 
|  | const char *str; | 
|  |  | 
|  | print_always("%-23s", prefix ?: "Summary"); | 
|  | if (mask & SAMPLE_RX_CNT) | 
|  | print_always(FMT_COLUMNl, RX(out->totals.rx)); | 
|  | if (mask & SAMPLE_REDIRECT_CNT) | 
|  | print_always(FMT_COLUMNl, REDIR(out->totals.redir)); | 
|  | printf(FMT_COLUMNl, | 
|  | out->totals.err + out->totals.drop + out->totals.drop_xmit, | 
|  | "err,drop/s"); | 
|  | if (mask & SAMPLE_DEVMAP_XMIT_CNT || | 
|  | mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI) | 
|  | printf(FMT_COLUMNl, XMIT(out->totals.xmit)); | 
|  | printf("\n"); | 
|  |  | 
|  | if (mask & SAMPLE_RX_CNT) { | 
|  | str = (sample_log_level & LL_DEFAULT) && out->rx_cnt.pps ? | 
|  | "receive total" : | 
|  | "receive"; | 
|  | print_err((out->rx_cnt.err || out->rx_cnt.drop), | 
|  | "  %-20s " FMT_COLUMNl FMT_COLUMNl FMT_COLUMNl "\n", | 
|  | str, PPS(out->rx_cnt.pps), DROP(out->rx_cnt.drop), | 
|  | ERR(out->rx_cnt.err)); | 
|  |  | 
|  | stats_get_rx_cnt(r, p, nr_cpus, NULL); | 
|  | } | 
|  |  | 
|  | if (mask & SAMPLE_CPUMAP_ENQUEUE_CNT) | 
|  | stats_get_cpumap_enqueue(r, p, nr_cpus); | 
|  |  | 
|  | if (mask & SAMPLE_CPUMAP_KTHREAD_CNT) { | 
|  | stats_get_cpumap_kthread(r, p, nr_cpus); | 
|  | stats_get_cpumap_remote(r, p, nr_cpus); | 
|  | } | 
|  |  | 
|  | if (mask & SAMPLE_REDIRECT_CNT) { | 
|  | str = out->redir_cnt.suc ? "redirect total" : "redirect"; | 
|  | print_default("  %-20s " FMT_COLUMNl "\n", str, | 
|  | REDIR(out->redir_cnt.suc)); | 
|  |  | 
|  | stats_get_redirect_cnt(r, p, nr_cpus, NULL); | 
|  | } | 
|  |  | 
|  | if (mask & SAMPLE_REDIRECT_ERR_CNT) { | 
|  | str = (sample_log_level & LL_DEFAULT) && out->redir_cnt.err ? | 
|  | "redirect_err total" : | 
|  | "redirect_err"; | 
|  | print_err(out->redir_cnt.err, "  %-20s " FMT_COLUMNl "\n", str, | 
|  | ERR(out->redir_cnt.err)); | 
|  |  | 
|  | stats_get_redirect_err_cnt(r, p, nr_cpus, NULL); | 
|  | } | 
|  |  | 
|  | if (mask & SAMPLE_EXCEPTION_CNT) { | 
|  | str = out->except_cnt.hits ? "xdp_exception total" : | 
|  | "xdp_exception"; | 
|  |  | 
|  | print_err(out->except_cnt.hits, "  %-20s " FMT_COLUMNl "\n", str, | 
|  | HITS(out->except_cnt.hits)); | 
|  |  | 
|  | stats_get_exception_cnt(r, p, nr_cpus, NULL); | 
|  | } | 
|  |  | 
|  | if (mask & SAMPLE_DEVMAP_XMIT_CNT) { | 
|  | str = (sample_log_level & LL_DEFAULT) && out->xmit_cnt.pps ? | 
|  | "devmap_xmit total" : | 
|  | "devmap_xmit"; | 
|  |  | 
|  | print_err(out->xmit_cnt.err || out->xmit_cnt.drop, | 
|  | "  %-20s " FMT_COLUMNl FMT_COLUMNl FMT_COLUMNl | 
|  | __COLUMN(".2f") "\n", | 
|  | str, XMIT(out->xmit_cnt.pps), | 
|  | DROP(out->xmit_cnt.drop), out->xmit_cnt.err, | 
|  | "drv_err/s", out->xmit_cnt.bavg, "bulk-avg"); | 
|  |  | 
|  | stats_get_devmap_xmit(r, p, nr_cpus, NULL); | 
|  | } | 
|  |  | 
|  | if (mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI) | 
|  | stats_get_devmap_xmit_multi(r, p, nr_cpus, NULL, | 
|  | mask & SAMPLE_DEVMAP_XMIT_CNT); | 
|  |  | 
|  | if (sample_log_level & LL_DEFAULT || | 
|  | ((sample_log_level & LL_SIMPLE) && sample_err_exp)) { | 
|  | sample_err_exp = false; | 
|  | printf("\n"); | 
|  | } | 
|  | } | 
|  |  | 
|  | int sample_setup_maps(struct bpf_map **maps) | 
|  | { | 
|  | sample_n_cpus = libbpf_num_possible_cpus(); | 
|  |  | 
|  | for (int i = 0; i < MAP_DEVMAP_XMIT_MULTI; i++) { | 
|  | sample_map[i] = maps[i]; | 
|  |  | 
|  | switch (i) { | 
|  | case MAP_RX: | 
|  | case MAP_CPUMAP_KTHREAD: | 
|  | case MAP_DEVMAP_XMIT: | 
|  | sample_map_count[i] = sample_n_cpus; | 
|  | break; | 
|  | case MAP_REDIRECT_ERR: | 
|  | sample_map_count[i] = | 
|  | XDP_REDIRECT_ERR_MAX * sample_n_cpus; | 
|  | break; | 
|  | case MAP_EXCEPTION: | 
|  | sample_map_count[i] = XDP_ACTION_MAX * sample_n_cpus; | 
|  | case MAP_CPUMAP_ENQUEUE: | 
|  | sample_map_count[i] = sample_n_cpus * sample_n_cpus; | 
|  | break; | 
|  | default: | 
|  | return -EINVAL; | 
|  | } | 
|  | if (bpf_map__set_max_entries(sample_map[i], sample_map_count[i]) < 0) | 
|  | return -errno; | 
|  | } | 
|  | sample_map[MAP_DEVMAP_XMIT_MULTI] = maps[MAP_DEVMAP_XMIT_MULTI]; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int sample_setup_maps_mappings(void) | 
|  | { | 
|  | for (int i = 0; i < MAP_DEVMAP_XMIT_MULTI; i++) { | 
|  | size_t size = sample_map_count[i] * sizeof(struct datarec); | 
|  |  | 
|  | sample_mmap[i] = mmap(NULL, size, PROT_READ | PROT_WRITE, | 
|  | MAP_SHARED, bpf_map__fd(sample_map[i]), 0); | 
|  | if (sample_mmap[i] == MAP_FAILED) | 
|  | return -errno; | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int __sample_init(int mask) | 
|  | { | 
|  | sigset_t st; | 
|  |  | 
|  | sigemptyset(&st); | 
|  | sigaddset(&st, SIGQUIT); | 
|  | sigaddset(&st, SIGINT); | 
|  | sigaddset(&st, SIGTERM); | 
|  |  | 
|  | if (sigprocmask(SIG_BLOCK, &st, NULL) < 0) | 
|  | return -errno; | 
|  |  | 
|  | sample_sig_fd = signalfd(-1, &st, SFD_CLOEXEC | SFD_NONBLOCK); | 
|  | if (sample_sig_fd < 0) | 
|  | return -errno; | 
|  |  | 
|  | sample_mask = mask; | 
|  |  | 
|  | return sample_setup_maps_mappings(); | 
|  | } | 
|  |  | 
|  | static int __sample_remove_xdp(int ifindex, __u32 prog_id, int xdp_flags) | 
|  | { | 
|  | __u32 cur_prog_id = 0; | 
|  | int ret; | 
|  |  | 
|  | if (prog_id) { | 
|  | ret = bpf_xdp_query_id(ifindex, xdp_flags, &cur_prog_id); | 
|  | if (ret < 0) | 
|  | return -errno; | 
|  |  | 
|  | if (prog_id != cur_prog_id) { | 
|  | print_always( | 
|  | "Program on ifindex %d does not match installed " | 
|  | "program, skipping unload\n", | 
|  | ifindex); | 
|  | return -ENOENT; | 
|  | } | 
|  | } | 
|  |  | 
|  | return bpf_xdp_detach(ifindex, xdp_flags, NULL); | 
|  | } | 
|  |  | 
|  | int sample_install_xdp(struct bpf_program *xdp_prog, int ifindex, bool generic, | 
|  | bool force) | 
|  | { | 
|  | int ret, xdp_flags = 0; | 
|  | __u32 prog_id = 0; | 
|  |  | 
|  | if (sample_xdp_cnt == 32) { | 
|  | fprintf(stderr, | 
|  | "Total limit for installed XDP programs in a sample reached\n"); | 
|  | return -ENOTSUP; | 
|  | } | 
|  |  | 
|  | xdp_flags |= !force ? XDP_FLAGS_UPDATE_IF_NOEXIST : 0; | 
|  | xdp_flags |= generic ? XDP_FLAGS_SKB_MODE : XDP_FLAGS_DRV_MODE; | 
|  | ret = bpf_xdp_attach(ifindex, bpf_program__fd(xdp_prog), xdp_flags, NULL); | 
|  | if (ret < 0) { | 
|  | ret = -errno; | 
|  | fprintf(stderr, | 
|  | "Failed to install program \"%s\" on ifindex %d, mode = %s, " | 
|  | "force = %s: %s\n", | 
|  | bpf_program__name(xdp_prog), ifindex, | 
|  | generic ? "skb" : "native", force ? "true" : "false", | 
|  | strerror(-ret)); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | ret = bpf_xdp_query_id(ifindex, xdp_flags, &prog_id); | 
|  | if (ret < 0) { | 
|  | ret = -errno; | 
|  | fprintf(stderr, | 
|  | "Failed to get XDP program id for ifindex %d, removing program: %s\n", | 
|  | ifindex, strerror(errno)); | 
|  | __sample_remove_xdp(ifindex, 0, xdp_flags); | 
|  | return ret; | 
|  | } | 
|  | sample_xdp_progs[sample_xdp_cnt++] = | 
|  | (struct xdp_desc){ ifindex, prog_id, xdp_flags }; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void sample_summary_print(void) | 
|  | { | 
|  | double num = sample_out.rx_cnt.num; | 
|  |  | 
|  | if (sample_out.totals.rx) { | 
|  | double pkts = sample_out.totals.rx; | 
|  |  | 
|  | print_always("  Packets received    : %'-10llu\n", | 
|  | sample_out.totals.rx); | 
|  | print_always("  Average packets/s   : %'-10.0f\n", | 
|  | sample_round(pkts / num)); | 
|  | } | 
|  | if (sample_out.totals.redir) { | 
|  | double pkts = sample_out.totals.redir; | 
|  |  | 
|  | print_always("  Packets redirected  : %'-10llu\n", | 
|  | sample_out.totals.redir); | 
|  | print_always("  Average redir/s     : %'-10.0f\n", | 
|  | sample_round(pkts / num)); | 
|  | } | 
|  | if (sample_out.totals.drop) | 
|  | print_always("  Rx dropped          : %'-10llu\n", | 
|  | sample_out.totals.drop); | 
|  | if (sample_out.totals.drop_xmit) | 
|  | print_always("  Tx dropped          : %'-10llu\n", | 
|  | sample_out.totals.drop_xmit); | 
|  | if (sample_out.totals.err) | 
|  | print_always("  Errors recorded     : %'-10llu\n", | 
|  | sample_out.totals.err); | 
|  | if (sample_out.totals.xmit) { | 
|  | double pkts = sample_out.totals.xmit; | 
|  |  | 
|  | print_always("  Packets transmitted : %'-10llu\n", | 
|  | sample_out.totals.xmit); | 
|  | print_always("  Average transmit/s  : %'-10.0f\n", | 
|  | sample_round(pkts / num)); | 
|  | } | 
|  | } | 
|  |  | 
|  | void sample_exit(int status) | 
|  | { | 
|  | size_t size; | 
|  |  | 
|  | for (int i = 0; i < NUM_MAP; i++) { | 
|  | size = sample_map_count[i] * sizeof(**sample_mmap); | 
|  | munmap(sample_mmap[i], size); | 
|  | } | 
|  | while (sample_xdp_cnt--) { | 
|  | int i = sample_xdp_cnt, ifindex, xdp_flags; | 
|  | __u32 prog_id; | 
|  |  | 
|  | prog_id = sample_xdp_progs[i].prog_id; | 
|  | ifindex = sample_xdp_progs[i].ifindex; | 
|  | xdp_flags = sample_xdp_progs[i].flags; | 
|  |  | 
|  | __sample_remove_xdp(ifindex, prog_id, xdp_flags); | 
|  | } | 
|  | sample_summary_print(); | 
|  | close(sample_sig_fd); | 
|  | exit(status); | 
|  | } | 
|  |  | 
|  | static int sample_stats_collect(struct stats_record *rec) | 
|  | { | 
|  | int i; | 
|  |  | 
|  | if (sample_mask & SAMPLE_RX_CNT) | 
|  | map_collect_percpu(sample_mmap[MAP_RX], &rec->rx_cnt); | 
|  |  | 
|  | if (sample_mask & SAMPLE_REDIRECT_CNT) | 
|  | map_collect_percpu(sample_mmap[MAP_REDIRECT_ERR], &rec->redir_err[0]); | 
|  |  | 
|  | if (sample_mask & SAMPLE_REDIRECT_ERR_CNT) { | 
|  | for (i = 1; i < XDP_REDIRECT_ERR_MAX; i++) | 
|  | map_collect_percpu(&sample_mmap[MAP_REDIRECT_ERR][i * sample_n_cpus], | 
|  | &rec->redir_err[i]); | 
|  | } | 
|  |  | 
|  | if (sample_mask & SAMPLE_CPUMAP_ENQUEUE_CNT) | 
|  | for (i = 0; i < sample_n_cpus; i++) | 
|  | map_collect_percpu(&sample_mmap[MAP_CPUMAP_ENQUEUE][i * sample_n_cpus], | 
|  | &rec->enq[i]); | 
|  |  | 
|  | if (sample_mask & SAMPLE_CPUMAP_KTHREAD_CNT) | 
|  | map_collect_percpu(sample_mmap[MAP_CPUMAP_KTHREAD], | 
|  | &rec->kthread); | 
|  |  | 
|  | if (sample_mask & SAMPLE_EXCEPTION_CNT) | 
|  | for (i = 0; i < XDP_ACTION_MAX; i++) | 
|  | map_collect_percpu(&sample_mmap[MAP_EXCEPTION][i * sample_n_cpus], | 
|  | &rec->exception[i]); | 
|  |  | 
|  | if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT) | 
|  | map_collect_percpu(sample_mmap[MAP_DEVMAP_XMIT], &rec->devmap_xmit); | 
|  |  | 
|  | if (sample_mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI) { | 
|  | if (map_collect_percpu_devmap(bpf_map__fd(sample_map[MAP_DEVMAP_XMIT_MULTI]), rec) < 0) | 
|  | return -EINVAL; | 
|  | } | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void sample_summary_update(struct sample_output *out) | 
|  | { | 
|  | sample_out.totals.rx += out->totals.rx; | 
|  | sample_out.totals.redir += out->totals.redir; | 
|  | sample_out.totals.drop += out->totals.drop; | 
|  | sample_out.totals.drop_xmit += out->totals.drop_xmit; | 
|  | sample_out.totals.err += out->totals.err; | 
|  | sample_out.totals.xmit += out->totals.xmit; | 
|  | sample_out.rx_cnt.num++; | 
|  | } | 
|  |  | 
|  | static void sample_stats_print(int mask, struct stats_record *cur, | 
|  | struct stats_record *prev, char *prog_name) | 
|  | { | 
|  | struct sample_output out = {}; | 
|  |  | 
|  | if (mask & SAMPLE_RX_CNT) | 
|  | stats_get_rx_cnt(cur, prev, 0, &out); | 
|  | if (mask & SAMPLE_REDIRECT_CNT) | 
|  | stats_get_redirect_cnt(cur, prev, 0, &out); | 
|  | if (mask & SAMPLE_REDIRECT_ERR_CNT) | 
|  | stats_get_redirect_err_cnt(cur, prev, 0, &out); | 
|  | if (mask & SAMPLE_EXCEPTION_CNT) | 
|  | stats_get_exception_cnt(cur, prev, 0, &out); | 
|  | if (mask & SAMPLE_DEVMAP_XMIT_CNT) | 
|  | stats_get_devmap_xmit(cur, prev, 0, &out); | 
|  | else if (mask & SAMPLE_DEVMAP_XMIT_CNT_MULTI) | 
|  | stats_get_devmap_xmit_multi(cur, prev, 0, &out, | 
|  | mask & SAMPLE_DEVMAP_XMIT_CNT); | 
|  | sample_summary_update(&out); | 
|  |  | 
|  | stats_print(prog_name, mask, cur, prev, &out); | 
|  | } | 
|  |  | 
|  | void sample_switch_mode(void) | 
|  | { | 
|  | sample_log_level ^= LL_DEBUG - 1; | 
|  | } | 
|  |  | 
|  | static int sample_signal_cb(void) | 
|  | { | 
|  | struct signalfd_siginfo si; | 
|  | int r; | 
|  |  | 
|  | r = read(sample_sig_fd, &si, sizeof(si)); | 
|  | if (r < 0) | 
|  | return -errno; | 
|  |  | 
|  | switch (si.ssi_signo) { | 
|  | case SIGQUIT: | 
|  | sample_switch_mode(); | 
|  | printf("\n"); | 
|  | break; | 
|  | default: | 
|  | printf("\n"); | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* Pointer swap trick */ | 
|  | static void swap(struct stats_record **a, struct stats_record **b) | 
|  | { | 
|  | struct stats_record *tmp; | 
|  |  | 
|  | tmp = *a; | 
|  | *a = *b; | 
|  | *b = tmp; | 
|  | } | 
|  |  | 
|  | static int sample_timer_cb(int timerfd, struct stats_record **rec, | 
|  | struct stats_record **prev) | 
|  | { | 
|  | char line[64] = "Summary"; | 
|  | int ret; | 
|  | __u64 t; | 
|  |  | 
|  | ret = read(timerfd, &t, sizeof(t)); | 
|  | if (ret < 0) | 
|  | return -errno; | 
|  |  | 
|  | swap(prev, rec); | 
|  | ret = sample_stats_collect(*rec); | 
|  | if (ret < 0) | 
|  | return ret; | 
|  |  | 
|  | if (sample_xdp_cnt == 2 && !(sample_mask & SAMPLE_SKIP_HEADING)) { | 
|  | char fi[IFNAMSIZ]; | 
|  | char to[IFNAMSIZ]; | 
|  | const char *f, *t; | 
|  |  | 
|  | f = t = NULL; | 
|  | if (if_indextoname(sample_xdp_progs[0].ifindex, fi)) | 
|  | f = fi; | 
|  | if (if_indextoname(sample_xdp_progs[1].ifindex, to)) | 
|  | t = to; | 
|  |  | 
|  | snprintf(line, sizeof(line), "%s->%s", f ?: "?", t ?: "?"); | 
|  | } | 
|  |  | 
|  | sample_stats_print(sample_mask, *rec, *prev, line); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | int sample_run(int interval, void (*post_cb)(void *), void *ctx) | 
|  | { | 
|  | struct timespec ts = { interval, 0 }; | 
|  | struct itimerspec its = { ts, ts }; | 
|  | struct stats_record *rec, *prev; | 
|  | struct pollfd pfd[2] = {}; | 
|  | int timerfd, ret; | 
|  |  | 
|  | if (!interval) { | 
|  | fprintf(stderr, "Incorrect interval 0\n"); | 
|  | return -EINVAL; | 
|  | } | 
|  | sample_interval = interval; | 
|  | /* Pretty print numbers */ | 
|  | setlocale(LC_NUMERIC, "en_US.UTF-8"); | 
|  |  | 
|  | timerfd = timerfd_create(CLOCK_MONOTONIC, TFD_CLOEXEC | TFD_NONBLOCK); | 
|  | if (timerfd < 0) | 
|  | return -errno; | 
|  | timerfd_settime(timerfd, 0, &its, NULL); | 
|  |  | 
|  | pfd[0].fd = sample_sig_fd; | 
|  | pfd[0].events = POLLIN; | 
|  |  | 
|  | pfd[1].fd = timerfd; | 
|  | pfd[1].events = POLLIN; | 
|  |  | 
|  | ret = -ENOMEM; | 
|  | rec = alloc_stats_record(); | 
|  | if (!rec) | 
|  | goto end; | 
|  | prev = alloc_stats_record(); | 
|  | if (!prev) | 
|  | goto end_rec; | 
|  |  | 
|  | ret = sample_stats_collect(rec); | 
|  | if (ret < 0) | 
|  | goto end_rec_prev; | 
|  |  | 
|  | for (;;) { | 
|  | ret = poll(pfd, 2, -1); | 
|  | if (ret < 0) { | 
|  | if (errno == EINTR) | 
|  | continue; | 
|  | else | 
|  | break; | 
|  | } | 
|  |  | 
|  | if (pfd[0].revents & POLLIN) | 
|  | ret = sample_signal_cb(); | 
|  | else if (pfd[1].revents & POLLIN) | 
|  | ret = sample_timer_cb(timerfd, &rec, &prev); | 
|  |  | 
|  | if (ret) | 
|  | break; | 
|  |  | 
|  | if (post_cb) | 
|  | post_cb(ctx); | 
|  | } | 
|  |  | 
|  | end_rec_prev: | 
|  | free_stats_record(prev); | 
|  | end_rec: | 
|  | free_stats_record(rec); | 
|  | end: | 
|  | close(timerfd); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | const char *get_driver_name(int ifindex) | 
|  | { | 
|  | struct ethtool_drvinfo drv = {}; | 
|  | char ifname[IF_NAMESIZE]; | 
|  | static char drvname[32]; | 
|  | struct ifreq ifr = {}; | 
|  | int fd, r = 0; | 
|  |  | 
|  | fd = socket(AF_INET, SOCK_DGRAM, 0); | 
|  | if (fd < 0) | 
|  | return "[error]"; | 
|  |  | 
|  | if (!if_indextoname(ifindex, ifname)) | 
|  | goto end; | 
|  |  | 
|  | drv.cmd = ETHTOOL_GDRVINFO; | 
|  | safe_strncpy(ifr.ifr_name, ifname, sizeof(ifr.ifr_name)); | 
|  | ifr.ifr_data = (void *)&drv; | 
|  |  | 
|  | r = ioctl(fd, SIOCETHTOOL, &ifr); | 
|  | if (r) | 
|  | goto end; | 
|  |  | 
|  | safe_strncpy(drvname, drv.driver, sizeof(drvname)); | 
|  |  | 
|  | close(fd); | 
|  | return drvname; | 
|  |  | 
|  | end: | 
|  | r = errno; | 
|  | close(fd); | 
|  | return r == EOPNOTSUPP ? "loopback" : "[error]"; | 
|  | } | 
|  |  | 
|  | int get_mac_addr(int ifindex, void *mac_addr) | 
|  | { | 
|  | char ifname[IF_NAMESIZE]; | 
|  | struct ifreq ifr = {}; | 
|  | int fd, r; | 
|  |  | 
|  | fd = socket(AF_INET, SOCK_DGRAM, 0); | 
|  | if (fd < 0) | 
|  | return -errno; | 
|  |  | 
|  | if (!if_indextoname(ifindex, ifname)) { | 
|  | r = -errno; | 
|  | goto end; | 
|  | } | 
|  |  | 
|  | safe_strncpy(ifr.ifr_name, ifname, sizeof(ifr.ifr_name)); | 
|  |  | 
|  | r = ioctl(fd, SIOCGIFHWADDR, &ifr); | 
|  | if (r) { | 
|  | r = -errno; | 
|  | goto end; | 
|  | } | 
|  |  | 
|  | memcpy(mac_addr, ifr.ifr_hwaddr.sa_data, 6 * sizeof(char)); | 
|  |  | 
|  | end: | 
|  | close(fd); | 
|  | return r; | 
|  | } | 
|  |  | 
|  | __attribute__((constructor)) static void sample_ctor(void) | 
|  | { | 
|  | if (libbpf_set_strict_mode(LIBBPF_STRICT_ALL) < 0) { | 
|  | fprintf(stderr, "Failed to set libbpf strict mode: %s\n", | 
|  | strerror(errno)); | 
|  | /* Just exit, nothing to cleanup right now */ | 
|  | exit(EXIT_FAIL_BPF); | 
|  | } | 
|  | } |