> On Nov 6, 2020, at 2:07 PM, Martin KaFai Lau <kafai@xxxxxx> wrote: > > sk_storage_charge() is the only user of omem_charge(). > This patch simplifies it by folding omem_charge() into > sk_storage_charge(). > > Signed-off-by: Martin KaFai Lau <kafai@xxxxxx> Acked-by: Song Liu <songliubraving@xxxxxx> > --- > net/core/bpf_sk_storage.c | 23 ++++++++++------------- > 1 file changed, 10 insertions(+), 13 deletions(-) > > diff --git a/net/core/bpf_sk_storage.c b/net/core/bpf_sk_storage.c > index c907f0dc7f87..001eac65e40f 100644 > --- a/net/core/bpf_sk_storage.c > +++ b/net/core/bpf_sk_storage.c > @@ -15,18 +15,6 @@ > > DEFINE_BPF_STORAGE_CACHE(sk_cache); > > -static int omem_charge(struct sock *sk, unsigned int size) > -{ > - /* same check as in sock_kmalloc() */ > - if (size <= sysctl_optmem_max && > - atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) { > - atomic_add(size, &sk->sk_omem_alloc); > - return 0; > - } > - > - return -ENOMEM; > -} > - > static struct bpf_local_storage_data * > sk_storage_lookup(struct sock *sk, struct bpf_map *map, bool cacheit_lockit) > { > @@ -316,7 +304,16 @@ BPF_CALL_2(bpf_sk_storage_delete, struct bpf_map *, map, struct sock *, sk) > static int sk_storage_charge(struct bpf_local_storage_map *smap, > void *owner, u32 size) > { > - return omem_charge(owner, size); > + struct sock *sk = (struct sock *)owner; > + > + /* same check as in sock_kmalloc() */ > + if (size <= sysctl_optmem_max && > + atomic_read(&sk->sk_omem_alloc) + size < sysctl_optmem_max) { > + atomic_add(size, &sk->sk_omem_alloc); > + return 0; > + } > + > + return -ENOMEM; > } > > static void sk_storage_uncharge(struct bpf_local_storage_map *smap, > -- > 2.24.1 >