On Sun, Sep 12, 2021 at 11:00 PM Marco Elver <elver@xxxxxxxxxx> wrote: > > On Fri, 10 Sept 2021 at 23:17, Andrey Konovalov <andreyknvl@xxxxxxxxx> wrote: > > > > On Fri, Sep 10, 2021 at 11:14 PM Peter Collingbourne <pcc@xxxxxxxxxx> wrote: > > > > > > With HW tag-based KASAN, error checks are performed implicitly by the > > > load and store instructions in the memcpy implementation. A failed check > > > results in tag checks being disabled and execution will keep going. As a > > > result, under HW tag-based KASAN, prior to commit 1b0668be62cf ("kasan: > > > test: disable kmalloc_memmove_invalid_size for HW_TAGS"), this memcpy > > > would end up corrupting memory until it hits an inaccessible page and > > > causes a kernel panic. > > > > > > This is a pre-existing issue that was revealed by commit 285133040e6c > > > ("arm64: Import latest memcpy()/memmove() implementation") which changed > > > the memcpy implementation from using signed comparisons (incorrectly, > > > resulting in the memcpy being terminated early for negative sizes) > > > to using unsigned comparisons. > > > > > > It is unclear how this could be handled by memcpy itself in a reasonable > > > way. One possibility would be to add an exception handler that would force > > > memcpy to return if a tag check fault is detected -- this would make the > > > behavior roughly similar to generic and SW tag-based KASAN. However, > > > this wouldn't solve the problem for asynchronous mode and also makes > > > memcpy behavior inconsistent with manually copying data. > > > > > > This test was added as a part of a series that taught KASAN to detect > > > negative sizes in memory operations, see commit 8cceeff48f23 ("kasan: > > > detect negative size in memory operation function"). Therefore we > > > should keep testing for negative sizes with generic and SW tag-based > > > KASAN. But there is some value in testing small memcpy overflows, so > > > let's add another test with memcpy that does not destabilize the kernel > > > by performing out-of-bounds writes, and run it in all modes. > > > > > > Link: https://linux-review.googlesource.com/id/I048d1e6a9aff766c4a53f989fb0c83de68923882 > > > Signed-off-by: Peter Collingbourne <pcc@xxxxxxxxxx> > > > --- > > > lib/test_kasan.c | 18 +++++++++++++++++- > > > 1 file changed, 17 insertions(+), 1 deletion(-) > > > > > > diff --git a/lib/test_kasan.c b/lib/test_kasan.c > > > index 8835e0784578..aa8e42250219 100644 > > > --- a/lib/test_kasan.c > > > +++ b/lib/test_kasan.c > > > @@ -493,7 +493,7 @@ static void kmalloc_oob_in_memset(struct kunit *test) > > > kfree(ptr); > > > } > > > > > > -static void kmalloc_memmove_invalid_size(struct kunit *test) > > > +static void kmalloc_memmove_negative_size(struct kunit *test) > > > { > > > char *ptr; > > > size_t size = 64; > > > @@ -515,6 +515,21 @@ static void kmalloc_memmove_invalid_size(struct kunit *test) > > > kfree(ptr); > > > } > > > > > > +static void kmalloc_memmove_invalid_size(struct kunit *test) > > > +{ > > > + char *ptr; > > > + size_t size = 64; > > > + volatile size_t invalid_size = size; > > > + > > > + ptr = kmalloc(size, GFP_KERNEL); > > > + KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr); > > > + > > > + memset((char *)ptr, 0, 64); > > > + KUNIT_EXPECT_KASAN_FAIL(test, > > > + memmove((char *)ptr, (char *)ptr + 4, invalid_size)); > > > + kfree(ptr); > > > +} > > > + > > > static void kmalloc_uaf(struct kunit *test) > > > { > > > char *ptr; > > > @@ -1129,6 +1144,7 @@ static struct kunit_case kasan_kunit_test_cases[] = { > > > KUNIT_CASE(kmalloc_oob_memset_4), > > > KUNIT_CASE(kmalloc_oob_memset_8), > > > KUNIT_CASE(kmalloc_oob_memset_16), > > > + KUNIT_CASE(kmalloc_memmove_negative_size), > > > KUNIT_CASE(kmalloc_memmove_invalid_size), > > > KUNIT_CASE(kmalloc_uaf), > > > KUNIT_CASE(kmalloc_uaf_memset), > > > -- > > > 2.33.0.309.g3052b89438-goog > > > > > > > Reviewed-by: Andrey Konovalov <andreyknvl@xxxxxxxxx> > > Acked-by: Marco Elver <elver@xxxxxxxxxx> > > Do you intend this patch to go through the arm64 or mm tree? Let's take it through the mm tree. Peter