Re: [PATCH bpf-next 4/5] bpf/benchs: Add benchmark test for bloom filter maps

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Aug 31, 2021 at 3:52 PM Joanne Koong <joannekoong@xxxxxx> wrote:
>
> This patch adds benchmark tests for the throughput and false positive
> rate of bloom filter map lookups for a given number of entries and a
> given number of hash functions.
>
> These benchmarks show that as the number of hash functions increases,
> the throughput and the false positive rate of the bloom filter map
> decreases. From the benchmark data, the approximate average
> false-positive rates are roughly as follows:
>
> 1 hash function = ~30%
> 2 hash functions = ~15%
> 3 hash functions = ~5%
> 4 hash functions = ~2.5%
> 5 hash functions = ~1%
> 6 hash functions = 0.5%
> 7 hash functions  = ~0.35%
> 8 hash functions = ~0.15%
> 9 hash functions = ~0.1%
> 10 hash functions = ~0%
>
> Signed-off-by: Joanne Koong <joannekoong@xxxxxx>
> ---
>  tools/testing/selftests/bpf/Makefile          |   4 +-
>  tools/testing/selftests/bpf/bench.c           |  35 ++
>  tools/testing/selftests/bpf/bench.h           |   3 +
>  .../bpf/benchs/bench_bloom_filter_map.c       | 344 ++++++++++++++++++
>  .../bpf/benchs/run_bench_bloom_filter_map.sh  |  28 ++
>  .../bpf/benchs/run_bench_ringbufs.sh          |  30 +-
>  .../selftests/bpf/benchs/run_common.sh        |  48 +++
>  .../selftests/bpf/progs/bloom_filter_map.c    |  74 ++++
>  8 files changed, 537 insertions(+), 29 deletions(-)
>  create mode 100644 tools/testing/selftests/bpf/benchs/bench_bloom_filter_map.c
>  create mode 100755 tools/testing/selftests/bpf/benchs/run_bench_bloom_filter_map.sh
>  create mode 100644 tools/testing/selftests/bpf/benchs/run_common.sh
>

[...]

> diff --git a/tools/testing/selftests/bpf/progs/bloom_filter_map.c b/tools/testing/selftests/bpf/progs/bloom_filter_map.c
> index 2d9c43a30246..1b139689219e 100644
> --- a/tools/testing/selftests/bpf/progs/bloom_filter_map.c
> +++ b/tools/testing/selftests/bpf/progs/bloom_filter_map.c
> @@ -1,7 +1,9 @@
>  // SPDX-License-Identifier: GPL-3.0
>  /* Copyright (c) 2021 Facebook */
>
> +#include <errno.h>
>  #include <linux/bpf.h>
> +#include <stdbool.h>
>  #include <bpf/bpf_helpers.h>
>
>  char _license[] SEC("license") = "GPL";
> @@ -23,8 +25,38 @@ struct {
>         __uint(nr_hashes, 2);
>  } map_bloom_filter SEC(".maps");
>
> +/* Tracks the number of hits, drops, and false hits */
> +struct {
> +       __uint(type, BPF_MAP_TYPE_PERCPU_ARRAY);
> +       __uint(max_entries, 3);
> +       __type(key, __u32);
> +       __type(value, __u64);
> +} percpu_array SEC(".maps");
> +
> +struct {
> +       __uint(type, BPF_MAP_TYPE_HASH);
> +       __uint(max_entries, 1000);
> +       __type(key, __u64);
> +       __type(value, __u64);
> +} hashmap SEC(".maps");
> +
> +const __u32 hit_key  = 0;
> +const __u32 drop_key  = 1;
> +const __u32 false_hit_key = 2;
> +
> +bool hashmap_use_bloom_filter = true;
> +
>  int error = 0;
>
> +static __always_inline void log_result(__u32 key)
> +{
> +       __u64 *count;
> +
> +       count = bpf_map_lookup_elem(&percpu_array, &key);
> +       if (count)
> +               *count += 1;

it will be actually more performant to have a global array with some
fixed number of elements (e.g., 256, to support up to 256 CPUs), one
for each CPU, instead of BPF_MAP_TYPE_PERCPU_ARRAY. Don't know how
much impact that has on benchmark, but doing one extra per-cpu map
lookup for each Bloom filter lookup might be a significant portion of
spent CPU.

> +}
> +
>  static __u64
>  check_elem(struct bpf_map *map, __u32 *key, __u64 *val,
>            void *data)
> @@ -37,6 +69,8 @@ check_elem(struct bpf_map *map, __u32 *key, __u64 *val,
>                 return 1; /* stop the iteration */
>         }
>
> +       log_result(hit_key);
> +
>         return 0;
>  }
>
> @@ -47,3 +81,43 @@ int prog_bloom_filter(void *ctx)
>
>         return 0;
>  }
> +
> +SEC("fentry/__x64_sys_getpgid")
> +int prog_bloom_filter_hashmap_lookup(void *ctx)
> +{
> +       __u64 *result;
> +       int i, err;
> +
> +       union {
> +               __u64 data64;
> +               __u32 data32[2];
> +       } val;
> +
> +       for (i = 0; i < 512; i++) {
> +               val.data32[0] = bpf_get_prandom_u32();
> +               val.data32[1] = bpf_get_prandom_u32();
> +
> +               if (hashmap_use_bloom_filter) {
> +                       err = bpf_map_peek_elem(&map_bloom_filter, &val);
> +                       if (err) {
> +                               if (err != -ENOENT) {
> +                                       error |= 2;
> +                                       return 0;
> +                               }
> +                               log_result(drop_key);
> +                               continue;
> +                       }
> +               }
> +
> +               result = bpf_map_lookup_elem(&hashmap, &val);
> +               if (result) {
> +                       log_result(hit_key);
> +               } else {
> +                       if (hashmap_use_bloom_filter)
> +                               log_result(false_hit_key);
> +                       log_result(drop_key);
> +               }
> +       }
> +
> +       return 0;
> +}
> --
> 2.30.2
>



[Index of Archives]     [Linux Samsung SoC]     [Linux Rockchip SoC]     [Linux Actions SoC]     [Linux for Synopsys ARC Processors]     [Linux NFS]     [Linux NILFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]


  Powered by Linux