| // SPDX-License-Identifier: GPL-2.0-only |
| /* Copyright(c) 2017 Jesper Dangaard Brouer, Red Hat, Inc. |
| */ |
| static const char *__doc__ = |
| " XDP redirect with a CPU-map type \"BPF_MAP_TYPE_CPUMAP\""; |
| |
| #include <errno.h> |
| #include <signal.h> |
| #include <stdio.h> |
| #include <stdlib.h> |
| #include <stdbool.h> |
| #include <string.h> |
| #include <unistd.h> |
| #include <locale.h> |
| #include <sys/resource.h> |
| #include <sys/sysinfo.h> |
| #include <getopt.h> |
| #include <net/if.h> |
| #include <time.h> |
| #include <linux/limits.h> |
| |
| #include <arpa/inet.h> |
| #include <linux/if_link.h> |
| |
| /* How many xdp_progs are defined in _kern.c */ |
| #define MAX_PROG 6 |
| |
| #include <bpf/bpf.h> |
| #include <bpf/libbpf.h> |
| |
| #include "bpf_util.h" |
| |
| static int ifindex = -1; |
| static char ifname_buf[IF_NAMESIZE]; |
| static char *ifname; |
| static __u32 prog_id; |
| |
| static __u32 xdp_flags = XDP_FLAGS_UPDATE_IF_NOEXIST; |
| static int n_cpus; |
| |
| enum map_type { |
| CPU_MAP, |
| RX_CNT, |
| REDIRECT_ERR_CNT, |
| CPUMAP_ENQUEUE_CNT, |
| CPUMAP_KTHREAD_CNT, |
| CPUS_AVAILABLE, |
| CPUS_COUNT, |
| CPUS_ITERATOR, |
| EXCEPTION_CNT, |
| }; |
| |
| static const char *const map_type_strings[] = { |
| [CPU_MAP] = "cpu_map", |
| [RX_CNT] = "rx_cnt", |
| [REDIRECT_ERR_CNT] = "redirect_err_cnt", |
| [CPUMAP_ENQUEUE_CNT] = "cpumap_enqueue_cnt", |
| [CPUMAP_KTHREAD_CNT] = "cpumap_kthread_cnt", |
| [CPUS_AVAILABLE] = "cpus_available", |
| [CPUS_COUNT] = "cpus_count", |
| [CPUS_ITERATOR] = "cpus_iterator", |
| [EXCEPTION_CNT] = "exception_cnt", |
| }; |
| |
| #define NUM_TP 5 |
| #define NUM_MAP 9 |
| struct bpf_link *tp_links[NUM_TP] = {}; |
| static int map_fds[NUM_MAP]; |
| static int tp_cnt = 0; |
| |
| /* Exit return codes */ |
| #define EXIT_OK 0 |
| #define EXIT_FAIL 1 |
| #define EXIT_FAIL_OPTION 2 |
| #define EXIT_FAIL_XDP 3 |
| #define EXIT_FAIL_BPF 4 |
| #define EXIT_FAIL_MEM 5 |
| |
| static const struct option long_options[] = { |
| {"help", no_argument, NULL, 'h' }, |
| {"dev", required_argument, NULL, 'd' }, |
| {"skb-mode", no_argument, NULL, 'S' }, |
| {"sec", required_argument, NULL, 's' }, |
| {"progname", required_argument, NULL, 'p' }, |
| {"qsize", required_argument, NULL, 'q' }, |
| {"cpu", required_argument, NULL, 'c' }, |
| {"stress-mode", no_argument, NULL, 'x' }, |
| {"no-separators", no_argument, NULL, 'z' }, |
| {"force", no_argument, NULL, 'F' }, |
| {"mprog-disable", no_argument, NULL, 'n' }, |
| {"mprog-name", required_argument, NULL, 'e' }, |
| {"mprog-filename", required_argument, NULL, 'f' }, |
| {"redirect-device", required_argument, NULL, 'r' }, |
| {"redirect-map", required_argument, NULL, 'm' }, |
| {0, 0, NULL, 0 } |
| }; |
| |
| static void int_exit(int sig) |
| { |
| __u32 curr_prog_id = 0; |
| |
| if (ifindex > -1) { |
| if (bpf_get_link_xdp_id(ifindex, &curr_prog_id, xdp_flags)) { |
| printf("bpf_get_link_xdp_id failed\n"); |
| exit(EXIT_FAIL); |
| } |
| if (prog_id == curr_prog_id) { |
| fprintf(stderr, |
| "Interrupted: Removing XDP program on ifindex:%d device:%s\n", |
| ifindex, ifname); |
| bpf_set_link_xdp_fd(ifindex, -1, xdp_flags); |
| } else if (!curr_prog_id) { |
| printf("couldn't find a prog id on a given iface\n"); |
| } else { |
| printf("program on interface changed, not removing\n"); |
| } |
| } |
| /* Detach tracepoints */ |
| while (tp_cnt) |
| bpf_link__destroy(tp_links[--tp_cnt]); |
| |
| exit(EXIT_OK); |
| } |
| |
| static void print_avail_progs(struct bpf_object *obj) |
| { |
| struct bpf_program *pos; |
| |
| bpf_object__for_each_program(pos, obj) { |
| if (bpf_program__is_xdp(pos)) |
| printf(" %s\n", bpf_program__section_name(pos)); |
| } |
| } |
| |
| static void usage(char *argv[], struct bpf_object *obj) |
| { |
| int i; |
| |
| printf("\nDOCUMENTATION:\n%s\n", __doc__); |
| printf("\n"); |
| printf(" Usage: %s (options-see-below)\n", argv[0]); |
| printf(" Listing options:\n"); |
| for (i = 0; long_options[i].name != 0; i++) { |
| printf(" --%-12s", long_options[i].name); |
| if (long_options[i].flag != NULL) |
| printf(" flag (internal value:%d)", |
| *long_options[i].flag); |
| else |
| printf(" short-option: -%c", |
| long_options[i].val); |
| printf("\n"); |
| } |
| printf("\n Programs to be used for --progname:\n"); |
| print_avail_progs(obj); |
| printf("\n"); |
| } |
| |
| /* gettime returns the current time of day in nanoseconds. |
| * Cost: clock_gettime (ns) => 26ns (CLOCK_MONOTONIC) |
| * clock_gettime (ns) => 9ns (CLOCK_MONOTONIC_COARSE) |
| */ |
| #define NANOSEC_PER_SEC 1000000000 /* 10^9 */ |
| static __u64 gettime(void) |
| { |
| struct timespec t; |
| int res; |
| |
| res = clock_gettime(CLOCK_MONOTONIC, &t); |
| if (res < 0) { |
| fprintf(stderr, "Error with gettimeofday! (%i)\n", res); |
| exit(EXIT_FAIL); |
| } |
| return (__u64) t.tv_sec * NANOSEC_PER_SEC + t.tv_nsec; |
| } |
| |
| /* Common stats data record shared with _kern.c */ |
| struct datarec { |
| __u64 processed; |
| __u64 dropped; |
| __u64 issue; |
| __u64 xdp_pass; |
| __u64 xdp_drop; |
| __u64 xdp_redirect; |
| }; |
| struct record { |
| __u64 timestamp; |
| struct datarec total; |
| struct datarec *cpu; |
| }; |
| struct stats_record { |
| struct record rx_cnt; |
| struct record redir_err; |
| struct record kthread; |
| struct record exception; |
| struct record enq[]; |
| }; |
| |
| static bool map_collect_percpu(int fd, __u32 key, struct record *rec) |
| { |
| /* For percpu maps, userspace gets a value per possible CPU */ |
| unsigned int nr_cpus = bpf_num_possible_cpus(); |
| struct datarec values[nr_cpus]; |
| __u64 sum_xdp_redirect = 0; |
| __u64 sum_xdp_pass = 0; |
| __u64 sum_xdp_drop = 0; |
| __u64 sum_processed = 0; |
| __u64 sum_dropped = 0; |
| __u64 sum_issue = 0; |
| int i; |
| |
| if ((bpf_map_lookup_elem(fd, &key, values)) != 0) { |
| fprintf(stderr, |
| "ERR: bpf_map_lookup_elem failed key:0x%X\n", key); |
| return false; |
| } |
| /* Get time as close as possible to reading map contents */ |
| rec->timestamp = gettime(); |
| |
| /* Record and sum values from each CPU */ |
| for (i = 0; i < nr_cpus; i++) { |
| rec->cpu[i].processed = values[i].processed; |
| sum_processed += values[i].processed; |
| rec->cpu[i].dropped = values[i].dropped; |
| sum_dropped += values[i].dropped; |
| rec->cpu[i].issue = values[i].issue; |
| sum_issue += values[i].issue; |
| rec->cpu[i].xdp_pass = values[i].xdp_pass; |
| sum_xdp_pass += values[i].xdp_pass; |
| rec->cpu[i].xdp_drop = values[i].xdp_drop; |
| sum_xdp_drop += values[i].xdp_drop; |
| rec->cpu[i].xdp_redirect = values[i].xdp_redirect; |
| sum_xdp_redirect += values[i].xdp_redirect; |
| } |
| rec->total.processed = sum_processed; |
| rec->total.dropped = sum_dropped; |
| rec->total.issue = sum_issue; |
| rec->total.xdp_pass = sum_xdp_pass; |
| rec->total.xdp_drop = sum_xdp_drop; |
| rec->total.xdp_redirect = sum_xdp_redirect; |
| return true; |
| } |
| |
| static struct datarec *alloc_record_per_cpu(void) |
| { |
| unsigned int nr_cpus = bpf_num_possible_cpus(); |
| struct datarec *array; |
| |
| array = calloc(nr_cpus, sizeof(struct datarec)); |
| if (!array) { |
| fprintf(stderr, "Mem alloc error (nr_cpus:%u)\n", nr_cpus); |
| exit(EXIT_FAIL_MEM); |
| } |
| return array; |
| } |
| |
| static struct stats_record *alloc_stats_record(void) |
| { |
| struct stats_record *rec; |
| int i, size; |
| |
| size = sizeof(*rec) + n_cpus * sizeof(struct record); |
| rec = malloc(size); |
| if (!rec) { |
| fprintf(stderr, "Mem alloc error\n"); |
| exit(EXIT_FAIL_MEM); |
| } |
| memset(rec, 0, size); |
| rec->rx_cnt.cpu = alloc_record_per_cpu(); |
| rec->redir_err.cpu = alloc_record_per_cpu(); |
| rec->kthread.cpu = alloc_record_per_cpu(); |
| rec->exception.cpu = alloc_record_per_cpu(); |
| for (i = 0; i < n_cpus; i++) |
| rec->enq[i].cpu = alloc_record_per_cpu(); |
| |
| return rec; |
| } |
| |
| static void free_stats_record(struct stats_record *r) |
| { |
| int i; |
| |
| for (i = 0; i < n_cpus; i++) |
| free(r->enq[i].cpu); |
| free(r->exception.cpu); |
| free(r->kthread.cpu); |
| free(r->redir_err.cpu); |
| free(r->rx_cnt.cpu); |
| free(r); |
| } |
| |
| static double calc_period(struct record *r, struct record *p) |
| { |
| double period_ = 0; |
| __u64 period = 0; |
| |
| period = r->timestamp - p->timestamp; |
| if (period > 0) |
| period_ = ((double) period / NANOSEC_PER_SEC); |
| |
| return period_; |
| } |
| |
| static __u64 calc_pps(struct datarec *r, struct datarec *p, double period_) |
| { |
| __u64 packets = 0; |
| __u64 pps = 0; |
| |
| if (period_ > 0) { |
| packets = r->processed - p->processed; |
| pps = packets / period_; |
| } |
| return pps; |
| } |
| |
| static __u64 calc_drop_pps(struct datarec *r, struct datarec *p, double period_) |
| { |
| __u64 packets = 0; |
| __u64 pps = 0; |
| |
| if (period_ > 0) { |
| packets = r->dropped - p->dropped; |
| pps = packets / period_; |
| } |
| return pps; |
| } |
| |
| static __u64 calc_errs_pps(struct datarec *r, |
| struct datarec *p, double period_) |
| { |
| __u64 packets = 0; |
| __u64 pps = 0; |
| |
| if (period_ > 0) { |
| packets = r->issue - p->issue; |
| pps = packets / period_; |
| } |
| return pps; |
| } |
| |
| static void calc_xdp_pps(struct datarec *r, struct datarec *p, |
| double *xdp_pass, double *xdp_drop, |
| double *xdp_redirect, double period_) |
| { |
| *xdp_pass = 0, *xdp_drop = 0, *xdp_redirect = 0; |
| if (period_ > 0) { |
| *xdp_redirect = (r->xdp_redirect - p->xdp_redirect) / period_; |
| *xdp_pass = (r->xdp_pass - p->xdp_pass) / period_; |
| *xdp_drop = (r->xdp_drop - p->xdp_drop) / period_; |
| } |
| } |
| |
| static void stats_print(struct stats_record *stats_rec, |
| struct stats_record *stats_prev, |
| char *prog_name, char *mprog_name, int mprog_fd) |
| { |
| unsigned int nr_cpus = bpf_num_possible_cpus(); |
| double pps = 0, drop = 0, err = 0; |
| bool mprog_enabled = false; |
| struct record *rec, *prev; |
| int to_cpu; |
| double t; |
| int i; |
| |
| if (mprog_fd > 0) |
| mprog_enabled = true; |
| |
| /* Header */ |
| printf("Running XDP/eBPF prog_name:%s\n", prog_name); |
| printf("%-15s %-7s %-14s %-11s %-9s\n", |
| "XDP-cpumap", "CPU:to", "pps", "drop-pps", "extra-info"); |
| |
| /* XDP rx_cnt */ |
| { |
| char *fmt_rx = "%-15s %-7d %'-14.0f %'-11.0f %'-10.0f %s\n"; |
| char *fm2_rx = "%-15s %-7s %'-14.0f %'-11.0f\n"; |
| char *errstr = ""; |
| |
| rec = &stats_rec->rx_cnt; |
| prev = &stats_prev->rx_cnt; |
| t = calc_period(rec, prev); |
| for (i = 0; i < nr_cpus; i++) { |
| struct datarec *r = &rec->cpu[i]; |
| struct datarec *p = &prev->cpu[i]; |
| |
| pps = calc_pps(r, p, t); |
| drop = calc_drop_pps(r, p, t); |
| err = calc_errs_pps(r, p, t); |
| if (err > 0) |
| errstr = "cpu-dest/err"; |
| if (pps > 0) |
| printf(fmt_rx, "XDP-RX", |
| i, pps, drop, err, errstr); |
| } |
| pps = calc_pps(&rec->total, &prev->total, t); |
| drop = calc_drop_pps(&rec->total, &prev->total, t); |
| err = calc_errs_pps(&rec->total, &prev->total, t); |
| printf(fm2_rx, "XDP-RX", "total", pps, drop); |
| } |
| |
| /* cpumap enqueue stats */ |
| for (to_cpu = 0; to_cpu < n_cpus; to_cpu++) { |
| char *fmt = "%-15s %3d:%-3d %'-14.0f %'-11.0f %'-10.2f %s\n"; |
| char *fm2 = "%-15s %3s:%-3d %'-14.0f %'-11.0f %'-10.2f %s\n"; |
| char *errstr = ""; |
| |
| rec = &stats_rec->enq[to_cpu]; |
| prev = &stats_prev->enq[to_cpu]; |
| t = calc_period(rec, prev); |
| for (i = 0; i < nr_cpus; i++) { |
| struct datarec *r = &rec->cpu[i]; |
| struct datarec *p = &prev->cpu[i]; |
| |
| pps = calc_pps(r, p, t); |
| drop = calc_drop_pps(r, p, t); |
| err = calc_errs_pps(r, p, t); |
| if (err > 0) { |
| errstr = "bulk-average"; |
| err = pps / err; /* calc average bulk size */ |
| } |
| if (pps > 0) |
| printf(fmt, "cpumap-enqueue", |
| i, to_cpu, pps, drop, err, errstr); |
| } |
| pps = calc_pps(&rec->total, &prev->total, t); |
| if (pps > 0) { |
| drop = calc_drop_pps(&rec->total, &prev->total, t); |
| err = calc_errs_pps(&rec->total, &prev->total, t); |
| if (err > 0) { |
| errstr = "bulk-average"; |
| err = pps / err; /* calc average bulk size */ |
| } |
| printf(fm2, "cpumap-enqueue", |
| "sum", to_cpu, pps, drop, err, errstr); |
| } |
| } |
| |
| /* cpumap kthread stats */ |
| { |
| char *fmt_k = "%-15s %-7d %'-14.0f %'-11.0f %'-10.0f %s\n"; |
| char *fm2_k = "%-15s %-7s %'-14.0f %'-11.0f %'-10.0f %s\n"; |
| char *e_str = ""; |
| |
| rec = &stats_rec->kthread; |
| prev = &stats_prev->kthread; |
| t = calc_period(rec, prev); |
| for (i = 0; i < nr_cpus; i++) { |
| struct datarec *r = &rec->cpu[i]; |
| struct datarec *p = &prev->cpu[i]; |
| |
| pps = calc_pps(r, p, t); |
| drop = calc_drop_pps(r, p, t); |
| err = calc_errs_pps(r, p, t); |
| if (err > 0) |
| e_str = "sched"; |
| if (pps > 0) |
| printf(fmt_k, "cpumap_kthread", |
| i, pps, drop, err, e_str); |
| } |
| pps = calc_pps(&rec->total, &prev->total, t); |
| drop = calc_drop_pps(&rec->total, &prev->total, t); |
| err = calc_errs_pps(&rec->total, &prev->total, t); |
| if (err > 0) |
| e_str = "sched-sum"; |
| printf(fm2_k, "cpumap_kthread", "total", pps, drop, err, e_str); |
| } |
| |
| /* XDP redirect err tracepoints (very unlikely) */ |
| { |
| char *fmt_err = "%-15s %-7d %'-14.0f %'-11.0f\n"; |
| char *fm2_err = "%-15s %-7s %'-14.0f %'-11.0f\n"; |
| |
| rec = &stats_rec->redir_err; |
| prev = &stats_prev->redir_err; |
| t = calc_period(rec, prev); |
| for (i = 0; i < nr_cpus; i++) { |
| struct datarec *r = &rec->cpu[i]; |
| struct datarec *p = &prev->cpu[i]; |
| |
| pps = calc_pps(r, p, t); |
| drop = calc_drop_pps(r, p, t); |
| if (pps > 0) |
| printf(fmt_err, "redirect_err", i, pps, drop); |
| } |
| pps = calc_pps(&rec->total, &prev->total, t); |
| drop = calc_drop_pps(&rec->total, &prev->total, t); |
| printf(fm2_err, "redirect_err", "total", pps, drop); |
| } |
| |
| /* XDP general exception tracepoints */ |
| { |
| char *fmt_err = "%-15s %-7d %'-14.0f %'-11.0f\n"; |
| char *fm2_err = "%-15s %-7s %'-14.0f %'-11.0f\n"; |
| |
| rec = &stats_rec->exception; |
| prev = &stats_prev->exception; |
| t = calc_period(rec, prev); |
| for (i = 0; i < nr_cpus; i++) { |
| struct datarec *r = &rec->cpu[i]; |
| struct datarec *p = &prev->cpu[i]; |
| |
| pps = calc_pps(r, p, t); |
| drop = calc_drop_pps(r, p, t); |
| if (pps > 0) |
| printf(fmt_err, "xdp_exception", i, pps, drop); |
| } |
| pps = calc_pps(&rec->total, &prev->total, t); |
| drop = calc_drop_pps(&rec->total, &prev->total, t); |
| printf(fm2_err, "xdp_exception", "total", pps, drop); |
| } |
| |
| /* CPUMAP attached XDP program that runs on remote/destination CPU */ |
| if (mprog_enabled) { |
| char *fmt_k = "%-15s %-7d %'-14.0f %'-11.0f %'-10.0f\n"; |
| char *fm2_k = "%-15s %-7s %'-14.0f %'-11.0f %'-10.0f\n"; |
| double xdp_pass, xdp_drop, xdp_redirect; |
| |
| printf("\n2nd remote XDP/eBPF prog_name: %s\n", mprog_name); |
| printf("%-15s %-7s %-14s %-11s %-9s\n", |
| "XDP-cpumap", "CPU:to", "xdp-pass", "xdp-drop", "xdp-redir"); |
| |
| rec = &stats_rec->kthread; |
| prev = &stats_prev->kthread; |
| t = calc_period(rec, prev); |
| for (i = 0; i < nr_cpus; i++) { |
| struct datarec *r = &rec->cpu[i]; |
| struct datarec *p = &prev->cpu[i]; |
| |
| calc_xdp_pps(r, p, &xdp_pass, &xdp_drop, |
| &xdp_redirect, t); |
| if (xdp_pass > 0 || xdp_drop > 0 || xdp_redirect > 0) |
| printf(fmt_k, "xdp-in-kthread", i, xdp_pass, xdp_drop, |
| xdp_redirect); |
| } |
| calc_xdp_pps(&rec->total, &prev->total, &xdp_pass, &xdp_drop, |
| &xdp_redirect, t); |
| printf(fm2_k, "xdp-in-kthread", "total", xdp_pass, xdp_drop, xdp_redirect); |
| } |
| |
| printf("\n"); |
| fflush(stdout); |
| } |
| |
| static void stats_collect(struct stats_record *rec) |
| { |
| int fd, i; |
| |
| fd = map_fds[RX_CNT]; |
| map_collect_percpu(fd, 0, &rec->rx_cnt); |
| |
| fd = map_fds[REDIRECT_ERR_CNT]; |
| map_collect_percpu(fd, 1, &rec->redir_err); |
| |
| fd = map_fds[CPUMAP_ENQUEUE_CNT]; |
| for (i = 0; i < n_cpus; i++) |
| map_collect_percpu(fd, i, &rec->enq[i]); |
| |
| fd = map_fds[CPUMAP_KTHREAD_CNT]; |
| map_collect_percpu(fd, 0, &rec->kthread); |
| |
| fd = map_fds[EXCEPTION_CNT]; |
| map_collect_percpu(fd, 0, &rec->exception); |
| } |
| |
| |
| /* Pointer swap trick */ |
| static inline void swap(struct stats_record **a, struct stats_record **b) |
| { |
| struct stats_record *tmp; |
| |
| tmp = *a; |
| *a = *b; |
| *b = tmp; |
| } |
| |
| static int create_cpu_entry(__u32 cpu, struct bpf_cpumap_val *value, |
| __u32 avail_idx, bool new) |
| { |
| __u32 curr_cpus_count = 0; |
| __u32 key = 0; |
| int ret; |
| |
| /* Add a CPU entry to cpumap, as this allocate a cpu entry in |
| * the kernel for the cpu. |
| */ |
| ret = bpf_map_update_elem(map_fds[CPU_MAP], &cpu, value, 0); |
| if (ret) { |
| fprintf(stderr, "Create CPU entry failed (err:%d)\n", ret); |
| exit(EXIT_FAIL_BPF); |
| } |
| |
| /* Inform bpf_prog's that a new CPU is available to select |
| * from via some control maps. |
| */ |
| ret = bpf_map_update_elem(map_fds[CPUS_AVAILABLE], &avail_idx, &cpu, 0); |
| if (ret) { |
| fprintf(stderr, "Add to avail CPUs failed\n"); |
| exit(EXIT_FAIL_BPF); |
| } |
| |
| /* When not replacing/updating existing entry, bump the count */ |
| ret = bpf_map_lookup_elem(map_fds[CPUS_COUNT], &key, &curr_cpus_count); |
| if (ret) { |
| fprintf(stderr, "Failed reading curr cpus_count\n"); |
| exit(EXIT_FAIL_BPF); |
| } |
| if (new) { |
| curr_cpus_count++; |
| ret = bpf_map_update_elem(map_fds[CPUS_COUNT], &key, |
| &curr_cpus_count, 0); |
| if (ret) { |
| fprintf(stderr, "Failed write curr cpus_count\n"); |
| exit(EXIT_FAIL_BPF); |
| } |
| } |
| /* map_fd[7] = cpus_iterator */ |
| printf("%s CPU:%u as idx:%u qsize:%d prog_fd: %d (cpus_count:%u)\n", |
| new ? "Add-new":"Replace", cpu, avail_idx, |
| value->qsize, value->bpf_prog.fd, curr_cpus_count); |
| |
| return 0; |
| } |
| |
| /* CPUs are zero-indexed. Thus, add a special sentinel default value |
| * in map cpus_available to mark CPU index'es not configured |
| */ |
| static void mark_cpus_unavailable(void) |
| { |
| __u32 invalid_cpu = n_cpus; |
| int ret, i; |
| |
| for (i = 0; i < n_cpus; i++) { |
| ret = bpf_map_update_elem(map_fds[CPUS_AVAILABLE], &i, |
| &invalid_cpu, 0); |
| if (ret) { |
| fprintf(stderr, "Failed marking CPU unavailable\n"); |
| exit(EXIT_FAIL_BPF); |
| } |
| } |
| } |
| |
| /* Stress cpumap management code by concurrently changing underlying cpumap */ |
| static void stress_cpumap(struct bpf_cpumap_val *value) |
| { |
| /* Changing qsize will cause kernel to free and alloc a new |
| * bpf_cpu_map_entry, with an associated/complicated tear-down |
| * procedure. |
| */ |
| value->qsize = 1024; |
| create_cpu_entry(1, value, 0, false); |
| value->qsize = 8; |
| create_cpu_entry(1, value, 0, false); |
| value->qsize = 16000; |
| create_cpu_entry(1, value, 0, false); |
| } |
| |
| static void stats_poll(int interval, bool use_separators, char *prog_name, |
| char *mprog_name, struct bpf_cpumap_val *value, |
| bool stress_mode) |
| { |
| struct stats_record *record, *prev; |
| int mprog_fd; |
| |
| record = alloc_stats_record(); |
| prev = alloc_stats_record(); |
| stats_collect(record); |
| |
| /* Trick to pretty printf with thousands separators use %' */ |
| if (use_separators) |
| setlocale(LC_NUMERIC, "en_US"); |
| |
| while (1) { |
| swap(&prev, &record); |
| mprog_fd = value->bpf_prog.fd; |
| stats_collect(record); |
| stats_print(record, prev, prog_name, mprog_name, mprog_fd); |
| sleep(interval); |
| if (stress_mode) |
| stress_cpumap(value); |
| } |
| |
| free_stats_record(record); |
| free_stats_record(prev); |
| } |
| |
| static int init_tracepoints(struct bpf_object *obj) |
| { |
| struct bpf_program *prog; |
| |
| bpf_object__for_each_program(prog, obj) { |
| if (bpf_program__is_tracepoint(prog) != true) |
| continue; |
| |
| tp_links[tp_cnt] = bpf_program__attach(prog); |
| if (libbpf_get_error(tp_links[tp_cnt])) { |
| tp_links[tp_cnt] = NULL; |
| return -EINVAL; |
| } |
| tp_cnt++; |
| } |
| |
| return 0; |
| } |
| |
| static int init_map_fds(struct bpf_object *obj) |
| { |
| enum map_type type; |
| |
| for (type = 0; type < NUM_MAP; type++) { |
| map_fds[type] = |
| bpf_object__find_map_fd_by_name(obj, |
| map_type_strings[type]); |
| |
| if (map_fds[type] < 0) |
| return -ENOENT; |
| } |
| |
| return 0; |
| } |
| |
| static int load_cpumap_prog(char *file_name, char *prog_name, |
| char *redir_interface, char *redir_map) |
| { |
| struct bpf_prog_load_attr prog_load_attr = { |
| .prog_type = BPF_PROG_TYPE_XDP, |
| .expected_attach_type = BPF_XDP_CPUMAP, |
| .file = file_name, |
| }; |
| struct bpf_program *prog; |
| struct bpf_object *obj; |
| int fd; |
| |
| if (bpf_prog_load_xattr(&prog_load_attr, &obj, &fd)) |
| return -1; |
| |
| if (fd < 0) { |
| fprintf(stderr, "ERR: bpf_prog_load_xattr: %s\n", |
| strerror(errno)); |
| return fd; |
| } |
| |
| if (redir_interface && redir_map) { |
| int err, map_fd, ifindex_out, key = 0; |
| |
| map_fd = bpf_object__find_map_fd_by_name(obj, redir_map); |
| if (map_fd < 0) |
| return map_fd; |
| |
| ifindex_out = if_nametoindex(redir_interface); |
| if (!ifindex_out) |
| return -1; |
| |
| err = bpf_map_update_elem(map_fd, &key, &ifindex_out, 0); |
| if (err < 0) |
| return err; |
| } |
| |
| prog = bpf_object__find_program_by_title(obj, prog_name); |
| if (!prog) { |
| fprintf(stderr, "bpf_object__find_program_by_title failed\n"); |
| return EXIT_FAIL; |
| } |
| |
| return bpf_program__fd(prog); |
| } |
| |
| int main(int argc, char **argv) |
| { |
| struct rlimit r = {RLIM_INFINITY, RLIM_INFINITY}; |
| char *prog_name = "xdp_cpu_map5_lb_hash_ip_pairs"; |
| char *mprog_filename = "xdp_redirect_kern.o"; |
| char *redir_interface = NULL, *redir_map = NULL; |
| char *mprog_name = "xdp_redirect_dummy"; |
| bool mprog_disable = false; |
| struct bpf_prog_load_attr prog_load_attr = { |
| .prog_type = BPF_PROG_TYPE_UNSPEC, |
| }; |
| struct bpf_prog_info info = {}; |
| __u32 info_len = sizeof(info); |
| struct bpf_cpumap_val value; |
| bool use_separators = true; |
| bool stress_mode = false; |
| struct bpf_program *prog; |
| struct bpf_object *obj; |
| int err = EXIT_FAIL; |
| char filename[256]; |
| int added_cpus = 0; |
| int longindex = 0; |
| int interval = 2; |
| int add_cpu = -1; |
| int opt, prog_fd; |
| int *cpu, i; |
| __u32 qsize; |
| |
| n_cpus = get_nprocs_conf(); |
| |
| /* Notice: choosing he queue size is very important with the |
| * ixgbe driver, because it's driver page recycling trick is |
| * dependend on pages being returned quickly. The number of |
| * out-standing packets in the system must be less-than 2x |
| * RX-ring size. |
| */ |
| qsize = 128+64; |
| |
| snprintf(filename, sizeof(filename), "%s_kern.o", argv[0]); |
| prog_load_attr.file = filename; |
| |
| if (setrlimit(RLIMIT_MEMLOCK, &r)) { |
| perror("setrlimit(RLIMIT_MEMLOCK)"); |
| return 1; |
| } |
| |
| if (bpf_prog_load_xattr(&prog_load_attr, &obj, &prog_fd)) |
| return err; |
| |
| if (prog_fd < 0) { |
| fprintf(stderr, "ERR: bpf_prog_load_xattr: %s\n", |
| strerror(errno)); |
| return err; |
| } |
| |
| if (init_tracepoints(obj) < 0) { |
| fprintf(stderr, "ERR: bpf_program__attach failed\n"); |
| return err; |
| } |
| |
| if (init_map_fds(obj) < 0) { |
| fprintf(stderr, "bpf_object__find_map_fd_by_name failed\n"); |
| return err; |
| } |
| mark_cpus_unavailable(); |
| |
| cpu = malloc(n_cpus * sizeof(int)); |
| if (!cpu) { |
| fprintf(stderr, "failed to allocate cpu array\n"); |
| return err; |
| } |
| memset(cpu, 0, n_cpus * sizeof(int)); |
| |
| /* Parse commands line args */ |
| while ((opt = getopt_long(argc, argv, "hSd:s:p:q:c:xzFf:e:r:m:n", |
| long_options, &longindex)) != -1) { |
| switch (opt) { |
| case 'd': |
| if (strlen(optarg) >= IF_NAMESIZE) { |
| fprintf(stderr, "ERR: --dev name too long\n"); |
| goto error; |
| } |
| ifname = (char *)&ifname_buf; |
| strncpy(ifname, optarg, IF_NAMESIZE); |
| ifindex = if_nametoindex(ifname); |
| if (ifindex == 0) { |
| fprintf(stderr, |
| "ERR: --dev name unknown err(%d):%s\n", |
| errno, strerror(errno)); |
| goto error; |
| } |
| break; |
| case 's': |
| interval = atoi(optarg); |
| break; |
| case 'S': |
| xdp_flags |= XDP_FLAGS_SKB_MODE; |
| break; |
| case 'x': |
| stress_mode = true; |
| break; |
| case 'z': |
| use_separators = false; |
| break; |
| case 'p': |
| /* Selecting eBPF prog to load */ |
| prog_name = optarg; |
| break; |
| case 'n': |
| mprog_disable = true; |
| break; |
| case 'f': |
| mprog_filename = optarg; |
| break; |
| case 'e': |
| mprog_name = optarg; |
| break; |
| case 'r': |
| redir_interface = optarg; |
| break; |
| case 'm': |
| redir_map = optarg; |
| break; |
| case 'c': |
| /* Add multiple CPUs */ |
| add_cpu = strtoul(optarg, NULL, 0); |
| if (add_cpu >= n_cpus) { |
| fprintf(stderr, |
| "--cpu nr too large for cpumap err(%d):%s\n", |
| errno, strerror(errno)); |
| goto error; |
| } |
| cpu[added_cpus++] = add_cpu; |
| break; |
| case 'q': |
| qsize = atoi(optarg); |
| break; |
| case 'F': |
| xdp_flags &= ~XDP_FLAGS_UPDATE_IF_NOEXIST; |
| break; |
| case 'h': |
| error: |
| default: |
| free(cpu); |
| usage(argv, obj); |
| return EXIT_FAIL_OPTION; |
| } |
| } |
| |
| if (!(xdp_flags & XDP_FLAGS_SKB_MODE)) |
| xdp_flags |= XDP_FLAGS_DRV_MODE; |
| |
| /* Required option */ |
| if (ifindex == -1) { |
| fprintf(stderr, "ERR: required option --dev missing\n"); |
| usage(argv, obj); |
| err = EXIT_FAIL_OPTION; |
| goto out; |
| } |
| /* Required option */ |
| if (add_cpu == -1) { |
| fprintf(stderr, "ERR: required option --cpu missing\n"); |
| fprintf(stderr, " Specify multiple --cpu option to add more\n"); |
| usage(argv, obj); |
| err = EXIT_FAIL_OPTION; |
| goto out; |
| } |
| |
| value.bpf_prog.fd = 0; |
| if (!mprog_disable) |
| value.bpf_prog.fd = load_cpumap_prog(mprog_filename, mprog_name, |
| redir_interface, redir_map); |
| if (value.bpf_prog.fd < 0) { |
| err = value.bpf_prog.fd; |
| goto out; |
| } |
| value.qsize = qsize; |
| |
| for (i = 0; i < added_cpus; i++) |
| create_cpu_entry(cpu[i], &value, i, true); |
| |
| /* Remove XDP program when program is interrupted or killed */ |
| signal(SIGINT, int_exit); |
| signal(SIGTERM, int_exit); |
| |
| prog = bpf_object__find_program_by_title(obj, prog_name); |
| if (!prog) { |
| fprintf(stderr, "bpf_object__find_program_by_title failed\n"); |
| goto out; |
| } |
| |
| prog_fd = bpf_program__fd(prog); |
| if (prog_fd < 0) { |
| fprintf(stderr, "bpf_program__fd failed\n"); |
| goto out; |
| } |
| |
| if (bpf_set_link_xdp_fd(ifindex, prog_fd, xdp_flags) < 0) { |
| fprintf(stderr, "link set xdp fd failed\n"); |
| err = EXIT_FAIL_XDP; |
| goto out; |
| } |
| |
| err = bpf_obj_get_info_by_fd(prog_fd, &info, &info_len); |
| if (err) { |
| printf("can't get prog info - %s\n", strerror(errno)); |
| goto out; |
| } |
| prog_id = info.id; |
| |
| stats_poll(interval, use_separators, prog_name, mprog_name, |
| &value, stress_mode); |
| |
| err = EXIT_OK; |
| out: |
| free(cpu); |
| return err; |
| } |