// SPDX-License-Identifier: GPL-2.0 /* GPLv2, Copyright(c) 2017 Jesper Dangaard Brouer, Red Hat, Inc. */ #ifndef _XDP_SAMPLE_COMMON_BPF_H #define _XDP_SAMPLE_COMMON_BPF_H #include "xdp_sample.bpf.h" #include #include #include #include #include #include array_map rx_cnt SEC(".maps"); array_map redir_err_cnt SEC(".maps"); array_map cpumap_enqueue_cnt SEC(".maps"); array_map cpumap_kthread_cnt SEC(".maps"); array_map exception_cnt SEC(".maps"); array_map devmap_xmit_cnt SEC(".maps"); array_map rxq_cnt SEC(".maps"); struct { __uint(type, BPF_MAP_TYPE_PERCPU_HASH); __uint(max_entries, 32 * 32); __type(key, __u64); __type(value, struct datarec); } devmap_xmit_cnt_multi SEC(".maps"); const volatile int nr_cpus = 0; /* These can be set before loading so that redundant comparisons can be DCE'd by * the verifier, and only actual matches are tried after loading tp_btf program. * This allows sample to filter tracepoint stats based on net_device. */ const volatile int from_match[32] = {}; const volatile int to_match[32] = {}; int cpumap_map_id = 0; /* Find if b is part of set a, but if a is empty set then evaluate to true */ #define IN_SET(a, b) \ ({ \ bool __res = !(a)[0]; \ for (int i = 0; i < ARRAY_SIZE(a) && (a)[i]; i++) { \ __res = (a)[i] == (b); \ if (__res) \ break; \ } \ __res; \ }) static __always_inline __u32 xdp_get_err_key(int err) { switch (err) { case 0: return 0; case -EINVAL: return 2; case -ENETDOWN: return 3; case -EMSGSIZE: return 4; case -EOPNOTSUPP: return 5; case -ENOSPC: return 6; default: return 1; } } static __always_inline int xdp_redirect_collect_stat(int from, int err) { __u32 cpu = bpf_get_smp_processor_id(); __u32 key = XDP_REDIRECT_ERROR; struct datarec *rec; __u32 idx; if (!IN_SET(from_match, from)) return 0; key = xdp_get_err_key(err); idx = key * nr_cpus + cpu; rec = bpf_map_lookup_elem(&redir_err_cnt, &idx); if (!rec) return 0; if (key) NO_TEAR_INC(rec->dropped); else NO_TEAR_INC(rec->processed); return 0; /* Indicate event was filtered (no further processing)*/ /* * Returning 1 here would allow e.g. a perf-record tracepoint * to see and record these events, but it doesn't work well * in-practice as stopping perf-record also unload this * bpf_prog. Plus, there is additional overhead of doing so. */ } SEC("tp_btf/xdp_redirect_err") int BPF_PROG(tp_xdp_redirect_err, const struct net_device *dev, const struct bpf_prog *xdp, const void *tgt, int err, const struct bpf_map *map, __u32 index) { return xdp_redirect_collect_stat(dev->ifindex, err); } SEC("tp_btf/xdp_redirect_map_err") int BPF_PROG(tp_xdp_redirect_map_err, const struct net_device *dev, const struct bpf_prog *xdp, const void *tgt, int err, const struct bpf_map *map, __u32 index) { return xdp_redirect_collect_stat(dev->ifindex, err); } SEC("tp_btf/xdp_redirect") int BPF_PROG(tp_xdp_redirect, const struct net_device *dev, const struct bpf_prog *xdp, const void *tgt, int err, const struct bpf_map *map, __u32 index) { return xdp_redirect_collect_stat(dev->ifindex, err); } SEC("tp_btf/xdp_redirect_map") int BPF_PROG(tp_xdp_redirect_map, const struct net_device *dev, const struct bpf_prog *xdp, const void *tgt, int err, const struct bpf_map *map, __u32 index) { return xdp_redirect_collect_stat(dev->ifindex, err); } SEC("tp_btf/xdp_cpumap_enqueue") int BPF_PROG(tp_xdp_cpumap_enqueue, int map_id, unsigned int processed, unsigned int drops, int to_cpu) { __u32 cpu = bpf_get_smp_processor_id(); struct datarec *rec; __u32 idx; if (cpumap_map_id && cpumap_map_id != map_id) return 0; idx = to_cpu * nr_cpus + cpu; rec = bpf_map_lookup_elem(&cpumap_enqueue_cnt, &idx); if (!rec) return 0; NO_TEAR_ADD(rec->processed, processed); NO_TEAR_ADD(rec->dropped, drops); /* Record bulk events, then userspace can calc average bulk size */ if (processed > 0) NO_TEAR_INC(rec->issue); /* Inception: It's possible to detect overload situations, via * this tracepoint. This can be used for creating a feedback * loop to XDP, which can take appropriate actions to mitigate * this overload situation. */ return 0; } SEC("tp_btf/xdp_cpumap_kthread") int BPF_PROG(tp_xdp_cpumap_kthread, int map_id, unsigned int processed, unsigned int drops, int sched, struct xdp_cpumap_stats *xdp_stats) { struct datarec *rec; __u32 cpu; if (cpumap_map_id && cpumap_map_id != map_id) return 0; cpu = bpf_get_smp_processor_id(); rec = bpf_map_lookup_elem(&cpumap_kthread_cnt, &cpu); if (!rec) return 0; NO_TEAR_ADD(rec->processed, processed); NO_TEAR_ADD(rec->dropped, drops); NO_TEAR_ADD(rec->xdp_pass, xdp_stats->pass); NO_TEAR_ADD(rec->xdp_drop, xdp_stats->drop); NO_TEAR_ADD(rec->xdp_redirect, xdp_stats->redirect); /* Count times kthread yielded CPU via schedule call */ if (sched) NO_TEAR_INC(rec->issue); return 0; } SEC("tp_btf/xdp_cpumap_kthread") int BPF_PROG(tp_xdp_cpumap_compat, int map_id, unsigned int processed, unsigned int drops, int sched) { struct datarec *rec; __u32 cpu; if (cpumap_map_id && cpumap_map_id != map_id) return 0; cpu = bpf_get_smp_processor_id(); rec = bpf_map_lookup_elem(&cpumap_kthread_cnt, &cpu); if (!rec) return 0; NO_TEAR_ADD(rec->processed, processed); NO_TEAR_ADD(rec->dropped, drops); /* Count times kthread yielded CPU via schedule call */ if (sched) NO_TEAR_INC(rec->issue); return 0; } SEC("tp_btf/xdp_exception") int BPF_PROG(tp_xdp_exception, const struct net_device *dev, const struct bpf_prog *xdp, __u32 act) { __u32 cpu = bpf_get_smp_processor_id(); struct datarec *rec; __u32 key = act, idx; if (!IN_SET(from_match, dev->ifindex)) return 0; if (!IN_SET(to_match, dev->ifindex)) return 0; if (key > XDP_REDIRECT) key = XDP_REDIRECT + 1; idx = key * nr_cpus + cpu; rec = bpf_map_lookup_elem(&exception_cnt, &idx); if (!rec) return 0; NO_TEAR_INC(rec->dropped); return 0; } SEC("tp_btf/xdp_devmap_xmit") int BPF_PROG(tp_xdp_devmap_xmit, const struct net_device *from_dev, const struct net_device *to_dev, int sent, int drops, int err) { struct datarec *rec; int idx_in, idx_out; __u32 cpu; idx_in = from_dev->ifindex; idx_out = to_dev->ifindex; if (!IN_SET(from_match, idx_in)) return 0; if (!IN_SET(to_match, idx_out)) return 0; cpu = bpf_get_smp_processor_id(); rec = bpf_map_lookup_elem(&devmap_xmit_cnt, &cpu); if (!rec) return 0; NO_TEAR_ADD(rec->processed, sent); NO_TEAR_ADD(rec->dropped, drops); /* Record bulk events, then userspace can calc average bulk size */ NO_TEAR_INC(rec->info); /* Record error cases, where no frame were sent */ /* Catch API error of drv ndo_xdp_xmit sent more than count */ if (err || drops < 0) NO_TEAR_INC(rec->issue); return 0; } SEC("tp_btf/xdp_devmap_xmit") int BPF_PROG(tp_xdp_devmap_xmit_multi, const struct net_device *from_dev, const struct net_device *to_dev, int sent, int drops, int err) { struct datarec empty = {}; struct datarec *rec; int idx_in, idx_out; __u64 idx; idx_in = from_dev->ifindex; idx_out = to_dev->ifindex; idx = idx_in; idx = idx << 32 | idx_out; if (!IN_SET(from_match, idx_in)) return 0; if (!IN_SET(to_match, idx_out)) return 0; bpf_map_update_elem(&devmap_xmit_cnt_multi, &idx, &empty, BPF_NOEXIST); rec = bpf_map_lookup_elem(&devmap_xmit_cnt_multi, &idx); if (!rec) return 0; NO_TEAR_ADD(rec->processed, sent); NO_TEAR_ADD(rec->dropped, drops); NO_TEAR_INC(rec->info); if (err || drops < 0) NO_TEAR_INC(rec->issue); return 0; } #endif