|  | // SPDX-License-Identifier: GPL-2.0 | 
|  | #include <vmlinux.h> | 
|  | #include <bpf/bpf_tracing.h> | 
|  | #include <bpf/bpf_helpers.h> | 
|  | #include <bpf/bpf_core_read.h> | 
|  | #include "bpf_experimental.h" | 
|  | #include "bpf_misc.h" | 
|  |  | 
|  | struct node_data { | 
|  | long key; | 
|  | long data; | 
|  | struct bpf_rb_node node; | 
|  | }; | 
|  |  | 
|  | #define private(name) SEC(".data." #name) __hidden __attribute__((aligned(8))) | 
|  | private(A) struct bpf_spin_lock glock; | 
|  | private(A) struct bpf_rb_root groot __contains(node_data, node); | 
|  | private(A) struct bpf_rb_root groot2 __contains(node_data, node); | 
|  |  | 
|  | static bool less(struct bpf_rb_node *a, const struct bpf_rb_node *b) | 
|  | { | 
|  | struct node_data *node_a; | 
|  | struct node_data *node_b; | 
|  |  | 
|  | node_a = container_of(a, struct node_data, node); | 
|  | node_b = container_of(b, struct node_data, node); | 
|  |  | 
|  | return node_a->key < node_b->key; | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("bpf_spin_lock at off=16 must be held for bpf_rb_root") | 
|  | long rbtree_api_nolock_add(void *ctx) | 
|  | { | 
|  | struct node_data *n; | 
|  |  | 
|  | n = bpf_obj_new(typeof(*n)); | 
|  | if (!n) | 
|  | return 1; | 
|  |  | 
|  | bpf_rbtree_add(&groot, &n->node, less); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("bpf_spin_lock at off=16 must be held for bpf_rb_root") | 
|  | long rbtree_api_nolock_remove(void *ctx) | 
|  | { | 
|  | struct node_data *n; | 
|  |  | 
|  | n = bpf_obj_new(typeof(*n)); | 
|  | if (!n) | 
|  | return 1; | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  | bpf_rbtree_add(&groot, &n->node, less); | 
|  | bpf_spin_unlock(&glock); | 
|  |  | 
|  | bpf_rbtree_remove(&groot, &n->node); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("bpf_spin_lock at off=16 must be held for bpf_rb_root") | 
|  | long rbtree_api_nolock_first(void *ctx) | 
|  | { | 
|  | bpf_rbtree_first(&groot); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("rbtree_remove node input must be non-owning ref") | 
|  | long rbtree_api_remove_unadded_node(void *ctx) | 
|  | { | 
|  | struct node_data *n, *m; | 
|  | struct bpf_rb_node *res; | 
|  |  | 
|  | n = bpf_obj_new(typeof(*n)); | 
|  | if (!n) | 
|  | return 1; | 
|  |  | 
|  | m = bpf_obj_new(typeof(*m)); | 
|  | if (!m) { | 
|  | bpf_obj_drop(n); | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  | bpf_rbtree_add(&groot, &n->node, less); | 
|  |  | 
|  | /* This remove should pass verifier */ | 
|  | res = bpf_rbtree_remove(&groot, &n->node); | 
|  | n = container_of(res, struct node_data, node); | 
|  |  | 
|  | /* This remove shouldn't, m isn't in an rbtree */ | 
|  | res = bpf_rbtree_remove(&groot, &m->node); | 
|  | m = container_of(res, struct node_data, node); | 
|  | bpf_spin_unlock(&glock); | 
|  |  | 
|  | if (n) | 
|  | bpf_obj_drop(n); | 
|  | if (m) | 
|  | bpf_obj_drop(m); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("Unreleased reference id=3 alloc_insn=10") | 
|  | long rbtree_api_remove_no_drop(void *ctx) | 
|  | { | 
|  | struct bpf_rb_node *res; | 
|  | struct node_data *n; | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  | res = bpf_rbtree_first(&groot); | 
|  | if (!res) | 
|  | goto unlock_err; | 
|  |  | 
|  | res = bpf_rbtree_remove(&groot, res); | 
|  |  | 
|  | if (res) { | 
|  | n = container_of(res, struct node_data, node); | 
|  | __sink(n); | 
|  | } | 
|  | bpf_spin_unlock(&glock); | 
|  |  | 
|  | /* if (res) { bpf_obj_drop(n); } is missing here */ | 
|  | return 0; | 
|  |  | 
|  | unlock_err: | 
|  | bpf_spin_unlock(&glock); | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("arg#1 expected pointer to allocated object") | 
|  | long rbtree_api_add_to_multiple_trees(void *ctx) | 
|  | { | 
|  | struct node_data *n; | 
|  |  | 
|  | n = bpf_obj_new(typeof(*n)); | 
|  | if (!n) | 
|  | return 1; | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  | bpf_rbtree_add(&groot, &n->node, less); | 
|  |  | 
|  | /* This add should fail since n already in groot's tree */ | 
|  | bpf_rbtree_add(&groot2, &n->node, less); | 
|  | bpf_spin_unlock(&glock); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("dereference of modified ptr_or_null_ ptr R2 off=16 disallowed") | 
|  | long rbtree_api_use_unchecked_remove_retval(void *ctx) | 
|  | { | 
|  | struct bpf_rb_node *res; | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  |  | 
|  | res = bpf_rbtree_first(&groot); | 
|  | if (!res) | 
|  | goto err_out; | 
|  | res = bpf_rbtree_remove(&groot, res); | 
|  |  | 
|  | bpf_spin_unlock(&glock); | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  | /* Must check res for NULL before using in rbtree_add below */ | 
|  | bpf_rbtree_add(&groot, res, less); | 
|  | bpf_spin_unlock(&glock); | 
|  | return 0; | 
|  |  | 
|  | err_out: | 
|  | bpf_spin_unlock(&glock); | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("rbtree_remove node input must be non-owning ref") | 
|  | long rbtree_api_add_release_unlock_escape(void *ctx) | 
|  | { | 
|  | struct node_data *n; | 
|  |  | 
|  | n = bpf_obj_new(typeof(*n)); | 
|  | if (!n) | 
|  | return 1; | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  | bpf_rbtree_add(&groot, &n->node, less); | 
|  | bpf_spin_unlock(&glock); | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  | /* After add() in previous critical section, n should be | 
|  | * release_on_unlock and released after previous spin_unlock, | 
|  | * so should not be possible to use it here | 
|  | */ | 
|  | bpf_rbtree_remove(&groot, &n->node); | 
|  | bpf_spin_unlock(&glock); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("rbtree_remove node input must be non-owning ref") | 
|  | long rbtree_api_first_release_unlock_escape(void *ctx) | 
|  | { | 
|  | struct bpf_rb_node *res; | 
|  | struct node_data *n; | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  | res = bpf_rbtree_first(&groot); | 
|  | if (!res) { | 
|  | bpf_spin_unlock(&glock); | 
|  | return 1; | 
|  | } | 
|  | n = container_of(res, struct node_data, node); | 
|  | bpf_spin_unlock(&glock); | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  | /* After first() in previous critical section, n should be | 
|  | * release_on_unlock and released after previous spin_unlock, | 
|  | * so should not be possible to use it here | 
|  | */ | 
|  | bpf_rbtree_remove(&groot, &n->node); | 
|  | bpf_spin_unlock(&glock); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static bool less__bad_fn_call_add(struct bpf_rb_node *a, const struct bpf_rb_node *b) | 
|  | { | 
|  | struct node_data *node_a; | 
|  | struct node_data *node_b; | 
|  |  | 
|  | node_a = container_of(a, struct node_data, node); | 
|  | node_b = container_of(b, struct node_data, node); | 
|  | bpf_rbtree_add(&groot, &node_a->node, less); | 
|  |  | 
|  | return node_a->key < node_b->key; | 
|  | } | 
|  |  | 
|  | static bool less__bad_fn_call_remove(struct bpf_rb_node *a, const struct bpf_rb_node *b) | 
|  | { | 
|  | struct node_data *node_a; | 
|  | struct node_data *node_b; | 
|  |  | 
|  | node_a = container_of(a, struct node_data, node); | 
|  | node_b = container_of(b, struct node_data, node); | 
|  | bpf_rbtree_remove(&groot, &node_a->node); | 
|  |  | 
|  | return node_a->key < node_b->key; | 
|  | } | 
|  |  | 
|  | static bool less__bad_fn_call_first_unlock_after(struct bpf_rb_node *a, const struct bpf_rb_node *b) | 
|  | { | 
|  | struct node_data *node_a; | 
|  | struct node_data *node_b; | 
|  |  | 
|  | node_a = container_of(a, struct node_data, node); | 
|  | node_b = container_of(b, struct node_data, node); | 
|  | bpf_rbtree_first(&groot); | 
|  | bpf_spin_unlock(&glock); | 
|  |  | 
|  | return node_a->key < node_b->key; | 
|  | } | 
|  |  | 
|  | static __always_inline | 
|  | long add_with_cb(bool (cb)(struct bpf_rb_node *a, const struct bpf_rb_node *b)) | 
|  | { | 
|  | struct node_data *n; | 
|  |  | 
|  | n = bpf_obj_new(typeof(*n)); | 
|  | if (!n) | 
|  | return 1; | 
|  |  | 
|  | bpf_spin_lock(&glock); | 
|  | bpf_rbtree_add(&groot, &n->node, cb); | 
|  | bpf_spin_unlock(&glock); | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("arg#1 expected pointer to allocated object") | 
|  | long rbtree_api_add_bad_cb_bad_fn_call_add(void *ctx) | 
|  | { | 
|  | return add_with_cb(less__bad_fn_call_add); | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("rbtree_remove not allowed in rbtree cb") | 
|  | long rbtree_api_add_bad_cb_bad_fn_call_remove(void *ctx) | 
|  | { | 
|  | return add_with_cb(less__bad_fn_call_remove); | 
|  | } | 
|  |  | 
|  | SEC("?tc") | 
|  | __failure __msg("can't spin_{lock,unlock} in rbtree cb") | 
|  | long rbtree_api_add_bad_cb_bad_fn_call_first_unlock_after(void *ctx) | 
|  | { | 
|  | return add_with_cb(less__bad_fn_call_first_unlock_after); | 
|  | } | 
|  |  | 
|  | char _license[] SEC("license") = "GPL"; |