bpf-fixes

-----BEGIN PGP SIGNATURE-----
 
 iQIzBAABCAAdFiEE+soXsSLHKoYyzcli6rmadz2vbToFAmjykZUACgkQ6rmadz2v
 bTppcA/+LKOzgJVJd9q3tbPeXb+xJOqiCGPDuv3tfmglHxD+rR5n2lsoTkryL4PZ
 k5yJhA95Z5RVeV3cevl8n4HlTJipm795k+fKz1YRbe9gB49w2SqqDf5s7LYuABOm
 YqdUSCWaLyoBNd+qi9SuyOVXSg0mSbJk0bEsXsgTp/5rUt9v6cz+36BE1Pkcp3Aa
 d6y2I2MBRGCADtEmsXh7+DI0aUvgMi2zlX8veQdfAJZjsQFwbr1ZxSxHYqtS6gux
 wueEiihzipakhONACJskQbRv80NwT5/VmrAI/ZRVzIhsywQhDGdXtQVNs6700/oq
 QmIZtgAL17Y0SAyhzQsQuGhJGKdWKhf3hzDKEDPslmyJ6OCpMAs/ttPHTJ6s/MmG
 6arSwZD/GAIoWhvWYP/zxTdmjqKX13uradvaNTv55hOhCLTTnZQKRSLk1NabHl7e
 V0f7NlZaVPnLerW/90+pn5pZFSrhk0Nno9to+yXaQ9TYJlK4e6dcB9y+rButQNrr
 7bTRyQ55fQlyS+NnaD85wL41IIX4WmJ3ATdrKZIMvGMJaZxjzXRvW4AjGHJ6hbbt
 GATdtISkYqZ4AdlSf2Vj9BysZkf7tS83SyRlM6WDm3IaAS3v5E/e1Ky2Kn6gWu70
 MNcSW/O0DSqXRkcDkY/tSOlYJBZJYo6ZuWhNdQAERA3OxldBSFM=
 =p8bI
 -----END PGP SIGNATURE-----

Merge tag 'bpf-fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/bpf/bpf

Pull bpf fixes from Alexei Starovoitov:

 - Replace bpf_map_kmalloc_node() with kmalloc_nolock() to fix kmemleak
   imbalance in tracking of bpf_async_cb structures (Alexei Starovoitov)

 - Make selftests/bpf arg_parsing.c more robust to errors (Andrii
   Nakryiko)

 - Fix redefinition of 'off' as different kind of symbol when I40E
   driver is builtin (Brahmajit Das)

 - Do not disable preemption in bpf_test_run (Sahil Chandna)

 - Fix memory leak in __lookup_instance error path (Shardul Bankar)

 - Ensure test data is flushed to disk before reading it (Xing Guo)

* tag 'bpf-fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/bpf/bpf:
  selftests/bpf: Fix redefinition of 'off' as different kind of symbol
  bpf: Do not disable preemption in bpf_test_run().
  bpf: Fix memory leak in __lookup_instance error path
  selftests: arg_parsing: Ensure data is flushed to disk before reading.
  bpf: Replace bpf_map_kmalloc_node() with kmalloc_nolock() to allocate bpf_async_cb structures.
  selftests/bpf: make arg_parsing.c more robust to crashes
  bpf: test_run: Fix ctx leak in bpf_prog_test_run_xdp error path
This commit is contained in:
Linus Torvalds 2025-10-18 08:00:43 -10:00
commit d303caf5ca
7 changed files with 59 additions and 40 deletions

View File

@ -2499,6 +2499,8 @@ int bpf_map_alloc_pages(const struct bpf_map *map, int nid,
#ifdef CONFIG_MEMCG #ifdef CONFIG_MEMCG
void *bpf_map_kmalloc_node(const struct bpf_map *map, size_t size, gfp_t flags, void *bpf_map_kmalloc_node(const struct bpf_map *map, size_t size, gfp_t flags,
int node); int node);
void *bpf_map_kmalloc_nolock(const struct bpf_map *map, size_t size, gfp_t flags,
int node);
void *bpf_map_kzalloc(const struct bpf_map *map, size_t size, gfp_t flags); void *bpf_map_kzalloc(const struct bpf_map *map, size_t size, gfp_t flags);
void *bpf_map_kvcalloc(struct bpf_map *map, size_t n, size_t size, void *bpf_map_kvcalloc(struct bpf_map *map, size_t n, size_t size,
gfp_t flags); gfp_t flags);
@ -2511,6 +2513,8 @@ void __percpu *bpf_map_alloc_percpu(const struct bpf_map *map, size_t size,
*/ */
#define bpf_map_kmalloc_node(_map, _size, _flags, _node) \ #define bpf_map_kmalloc_node(_map, _size, _flags, _node) \
kmalloc_node(_size, _flags, _node) kmalloc_node(_size, _flags, _node)
#define bpf_map_kmalloc_nolock(_map, _size, _flags, _node) \
kmalloc_nolock(_size, _flags, _node)
#define bpf_map_kzalloc(_map, _size, _flags) \ #define bpf_map_kzalloc(_map, _size, _flags) \
kzalloc(_size, _flags) kzalloc(_size, _flags)
#define bpf_map_kvcalloc(_map, _n, _size, _flags) \ #define bpf_map_kvcalloc(_map, _n, _size, _flags) \

View File

@ -1215,13 +1215,20 @@ static void bpf_wq_work(struct work_struct *work)
rcu_read_unlock_trace(); rcu_read_unlock_trace();
} }
static void bpf_async_cb_rcu_free(struct rcu_head *rcu)
{
struct bpf_async_cb *cb = container_of(rcu, struct bpf_async_cb, rcu);
kfree_nolock(cb);
}
static void bpf_wq_delete_work(struct work_struct *work) static void bpf_wq_delete_work(struct work_struct *work)
{ {
struct bpf_work *w = container_of(work, struct bpf_work, delete_work); struct bpf_work *w = container_of(work, struct bpf_work, delete_work);
cancel_work_sync(&w->work); cancel_work_sync(&w->work);
kfree_rcu(w, cb.rcu); call_rcu(&w->cb.rcu, bpf_async_cb_rcu_free);
} }
static void bpf_timer_delete_work(struct work_struct *work) static void bpf_timer_delete_work(struct work_struct *work)
@ -1230,13 +1237,13 @@ static void bpf_timer_delete_work(struct work_struct *work)
/* Cancel the timer and wait for callback to complete if it was running. /* Cancel the timer and wait for callback to complete if it was running.
* If hrtimer_cancel() can be safely called it's safe to call * If hrtimer_cancel() can be safely called it's safe to call
* kfree_rcu(t) right after for both preallocated and non-preallocated * call_rcu() right after for both preallocated and non-preallocated
* maps. The async->cb = NULL was already done and no code path can see * maps. The async->cb = NULL was already done and no code path can see
* address 't' anymore. Timer if armed for existing bpf_hrtimer before * address 't' anymore. Timer if armed for existing bpf_hrtimer before
* bpf_timer_cancel_and_free will have been cancelled. * bpf_timer_cancel_and_free will have been cancelled.
*/ */
hrtimer_cancel(&t->timer); hrtimer_cancel(&t->timer);
kfree_rcu(t, cb.rcu); call_rcu(&t->cb.rcu, bpf_async_cb_rcu_free);
} }
static int __bpf_async_init(struct bpf_async_kern *async, struct bpf_map *map, u64 flags, static int __bpf_async_init(struct bpf_async_kern *async, struct bpf_map *map, u64 flags,
@ -1270,11 +1277,7 @@ static int __bpf_async_init(struct bpf_async_kern *async, struct bpf_map *map, u
goto out; goto out;
} }
/* Allocate via bpf_map_kmalloc_node() for memcg accounting. Until cb = bpf_map_kmalloc_nolock(map, size, 0, map->numa_node);
* kmalloc_nolock() is available, avoid locking issues by using
* __GFP_HIGH (GFP_ATOMIC & ~__GFP_RECLAIM).
*/
cb = bpf_map_kmalloc_node(map, size, __GFP_HIGH, map->numa_node);
if (!cb) { if (!cb) {
ret = -ENOMEM; ret = -ENOMEM;
goto out; goto out;
@ -1315,7 +1318,7 @@ static int __bpf_async_init(struct bpf_async_kern *async, struct bpf_map *map, u
* or pinned in bpffs. * or pinned in bpffs.
*/ */
WRITE_ONCE(async->cb, NULL); WRITE_ONCE(async->cb, NULL);
kfree(cb); kfree_nolock(cb);
ret = -EPERM; ret = -EPERM;
} }
out: out:
@ -1580,7 +1583,7 @@ void bpf_timer_cancel_and_free(void *val)
* timer _before_ calling us, such that failing to cancel it here will * timer _before_ calling us, such that failing to cancel it here will
* cause it to possibly use struct hrtimer after freeing bpf_hrtimer. * cause it to possibly use struct hrtimer after freeing bpf_hrtimer.
* Therefore, we _need_ to cancel any outstanding timers before we do * Therefore, we _need_ to cancel any outstanding timers before we do
* kfree_rcu, even though no more timers can be armed. * call_rcu, even though no more timers can be armed.
* *
* Moreover, we need to schedule work even if timer does not belong to * Moreover, we need to schedule work even if timer does not belong to
* the calling callback_fn, as on two different CPUs, we can end up in a * the calling callback_fn, as on two different CPUs, we can end up in a
@ -1607,7 +1610,7 @@ void bpf_timer_cancel_and_free(void *val)
* completion. * completion.
*/ */
if (hrtimer_try_to_cancel(&t->timer) >= 0) if (hrtimer_try_to_cancel(&t->timer) >= 0)
kfree_rcu(t, cb.rcu); call_rcu(&t->cb.rcu, bpf_async_cb_rcu_free);
else else
queue_work(system_dfl_wq, &t->cb.delete_work); queue_work(system_dfl_wq, &t->cb.delete_work);
} else { } else {

View File

@ -195,8 +195,10 @@ static struct func_instance *__lookup_instance(struct bpf_verifier_env *env,
return ERR_PTR(-ENOMEM); return ERR_PTR(-ENOMEM);
result->must_write_set = kvcalloc(subprog_sz, sizeof(*result->must_write_set), result->must_write_set = kvcalloc(subprog_sz, sizeof(*result->must_write_set),
GFP_KERNEL_ACCOUNT); GFP_KERNEL_ACCOUNT);
if (!result->must_write_set) if (!result->must_write_set) {
kvfree(result);
return ERR_PTR(-ENOMEM); return ERR_PTR(-ENOMEM);
}
memcpy(&result->callchain, callchain, sizeof(*callchain)); memcpy(&result->callchain, callchain, sizeof(*callchain));
result->insn_cnt = subprog_sz; result->insn_cnt = subprog_sz;
hash_add(liveness->func_instances, &result->hl_node, key); hash_add(liveness->func_instances, &result->hl_node, key);

View File

@ -520,6 +520,21 @@ void *bpf_map_kmalloc_node(const struct bpf_map *map, size_t size, gfp_t flags,
return ptr; return ptr;
} }
void *bpf_map_kmalloc_nolock(const struct bpf_map *map, size_t size, gfp_t flags,
int node)
{
struct mem_cgroup *memcg, *old_memcg;
void *ptr;
memcg = bpf_map_get_memcg(map);
old_memcg = set_active_memcg(memcg);
ptr = kmalloc_nolock(size, flags | __GFP_ACCOUNT, node);
set_active_memcg(old_memcg);
mem_cgroup_put(memcg);
return ptr;
}
void *bpf_map_kzalloc(const struct bpf_map *map, size_t size, gfp_t flags) void *bpf_map_kzalloc(const struct bpf_map *map, size_t size, gfp_t flags)
{ {
struct mem_cgroup *memcg, *old_memcg; struct mem_cgroup *memcg, *old_memcg;

View File

@ -29,7 +29,6 @@
#include <trace/events/bpf_test_run.h> #include <trace/events/bpf_test_run.h>
struct bpf_test_timer { struct bpf_test_timer {
enum { NO_PREEMPT, NO_MIGRATE } mode;
u32 i; u32 i;
u64 time_start, time_spent; u64 time_start, time_spent;
}; };
@ -37,12 +36,7 @@ struct bpf_test_timer {
static void bpf_test_timer_enter(struct bpf_test_timer *t) static void bpf_test_timer_enter(struct bpf_test_timer *t)
__acquires(rcu) __acquires(rcu)
{ {
rcu_read_lock(); rcu_read_lock_dont_migrate();
if (t->mode == NO_PREEMPT)
preempt_disable();
else
migrate_disable();
t->time_start = ktime_get_ns(); t->time_start = ktime_get_ns();
} }
@ -50,12 +44,7 @@ static void bpf_test_timer_leave(struct bpf_test_timer *t)
__releases(rcu) __releases(rcu)
{ {
t->time_start = 0; t->time_start = 0;
rcu_read_unlock_migrate();
if (t->mode == NO_PREEMPT)
preempt_enable();
else
migrate_enable();
rcu_read_unlock();
} }
static bool bpf_test_timer_continue(struct bpf_test_timer *t, int iterations, static bool bpf_test_timer_continue(struct bpf_test_timer *t, int iterations,
@ -374,7 +363,7 @@ static int bpf_test_run_xdp_live(struct bpf_prog *prog, struct xdp_buff *ctx,
{ {
struct xdp_test_data xdp = { .batch_size = batch_size }; struct xdp_test_data xdp = { .batch_size = batch_size };
struct bpf_test_timer t = { .mode = NO_MIGRATE }; struct bpf_test_timer t = {};
int ret; int ret;
if (!repeat) if (!repeat)
@ -404,7 +393,7 @@ static int bpf_test_run(struct bpf_prog *prog, void *ctx, u32 repeat,
struct bpf_prog_array_item item = {.prog = prog}; struct bpf_prog_array_item item = {.prog = prog};
struct bpf_run_ctx *old_ctx; struct bpf_run_ctx *old_ctx;
struct bpf_cg_run_ctx run_ctx; struct bpf_cg_run_ctx run_ctx;
struct bpf_test_timer t = { NO_MIGRATE }; struct bpf_test_timer t = {};
enum bpf_cgroup_storage_type stype; enum bpf_cgroup_storage_type stype;
int ret; int ret;
@ -1269,7 +1258,7 @@ int bpf_prog_test_run_xdp(struct bpf_prog *prog, const union bpf_attr *kattr,
goto free_ctx; goto free_ctx;
if (kattr->test.data_size_in - meta_sz < ETH_HLEN) if (kattr->test.data_size_in - meta_sz < ETH_HLEN)
return -EINVAL; goto free_ctx;
data = bpf_test_init(kattr, linear_sz, max_linear_sz, headroom, tailroom); data = bpf_test_init(kattr, linear_sz, max_linear_sz, headroom, tailroom);
if (IS_ERR(data)) { if (IS_ERR(data)) {
@ -1377,7 +1366,7 @@ int bpf_prog_test_run_flow_dissector(struct bpf_prog *prog,
const union bpf_attr *kattr, const union bpf_attr *kattr,
union bpf_attr __user *uattr) union bpf_attr __user *uattr)
{ {
struct bpf_test_timer t = { NO_PREEMPT }; struct bpf_test_timer t = {};
u32 size = kattr->test.data_size_in; u32 size = kattr->test.data_size_in;
struct bpf_flow_dissector ctx = {}; struct bpf_flow_dissector ctx = {};
u32 repeat = kattr->test.repeat; u32 repeat = kattr->test.repeat;
@ -1445,7 +1434,7 @@ out:
int bpf_prog_test_run_sk_lookup(struct bpf_prog *prog, const union bpf_attr *kattr, int bpf_prog_test_run_sk_lookup(struct bpf_prog *prog, const union bpf_attr *kattr,
union bpf_attr __user *uattr) union bpf_attr __user *uattr)
{ {
struct bpf_test_timer t = { NO_PREEMPT }; struct bpf_test_timer t = {};
struct bpf_prog_array *progs = NULL; struct bpf_prog_array *progs = NULL;
struct bpf_sk_lookup_kern ctx = {}; struct bpf_sk_lookup_kern ctx = {};
u32 repeat = kattr->test.repeat; u32 repeat = kattr->test.repeat;

View File

@ -144,11 +144,17 @@ static void test_parse_test_list_file(void)
if (!ASSERT_OK(ferror(fp), "prepare tmp")) if (!ASSERT_OK(ferror(fp), "prepare tmp"))
goto out_fclose; goto out_fclose;
if (!ASSERT_OK(fsync(fileno(fp)), "fsync tmp"))
goto out_fclose;
init_test_filter_set(&set); init_test_filter_set(&set);
ASSERT_OK(parse_test_list_file(tmpfile, &set, true), "parse file"); if (!ASSERT_OK(parse_test_list_file(tmpfile, &set, true), "parse file"))
goto out_fclose;
if (!ASSERT_EQ(set.cnt, 4, "test count"))
goto out_free_set;
ASSERT_EQ(set.cnt, 4, "test count");
ASSERT_OK(strcmp("test_with_spaces", set.tests[0].name), "test 0 name"); ASSERT_OK(strcmp("test_with_spaces", set.tests[0].name), "test 0 name");
ASSERT_EQ(set.tests[0].subtest_cnt, 0, "test 0 subtest count"); ASSERT_EQ(set.tests[0].subtest_cnt, 0, "test 0 subtest count");
ASSERT_OK(strcmp("testA", set.tests[1].name), "test 1 name"); ASSERT_OK(strcmp("testA", set.tests[1].name), "test 1 name");
@ -158,8 +164,8 @@ static void test_parse_test_list_file(void)
ASSERT_OK(strcmp("testB", set.tests[2].name), "test 2 name"); ASSERT_OK(strcmp("testB", set.tests[2].name), "test 2 name");
ASSERT_OK(strcmp("testC_no_eof_newline", set.tests[3].name), "test 3 name"); ASSERT_OK(strcmp("testC_no_eof_newline", set.tests[3].name), "test 3 name");
out_free_set:
free_test_filter_set(&set); free_test_filter_set(&set);
out_fclose: out_fclose:
fclose(fp); fclose(fp);
out_remove: out_remove:

View File

@ -225,7 +225,7 @@ int trusted_to_untrusted(void *ctx)
} }
char mem[16]; char mem[16];
u32 off; u32 offset;
SEC("tp_btf/sys_enter") SEC("tp_btf/sys_enter")
__success __success
@ -240,9 +240,9 @@ int anything_to_untrusted(void *ctx)
/* scalar to untrusted */ /* scalar to untrusted */
subprog_untrusted(0); subprog_untrusted(0);
/* variable offset to untrusted (map) */ /* variable offset to untrusted (map) */
subprog_untrusted((void *)mem + off); subprog_untrusted((void *)mem + offset);
/* variable offset to untrusted (trusted) */ /* variable offset to untrusted (trusted) */
subprog_untrusted((void *)bpf_get_current_task_btf() + off); subprog_untrusted((void *)bpf_get_current_task_btf() + offset);
return 0; return 0;
} }
@ -298,12 +298,12 @@ int anything_to_untrusted_mem(void *ctx)
/* scalar to untrusted mem */ /* scalar to untrusted mem */
subprog_void_untrusted(0); subprog_void_untrusted(0);
/* variable offset to untrusted mem (map) */ /* variable offset to untrusted mem (map) */
subprog_void_untrusted((void *)mem + off); subprog_void_untrusted((void *)mem + offset);
/* variable offset to untrusted mem (trusted) */ /* variable offset to untrusted mem (trusted) */
subprog_void_untrusted(bpf_get_current_task_btf() + off); subprog_void_untrusted(bpf_get_current_task_btf() + offset);
/* variable offset to untrusted char/enum (map) */ /* variable offset to untrusted char/enum (map) */
subprog_char_untrusted(mem + off); subprog_char_untrusted(mem + offset);
subprog_enum_untrusted((void *)mem + off); subprog_enum_untrusted((void *)mem + offset);
return 0; return 0;
} }