lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <aJtyR3hCW5fG+niV@e129823.arm.com>
Date: Tue, 12 Aug 2025 17:56:39 +0100
From: Yeoreum Yun <yeoreum.yun@....com>
To: Andrey Konovalov <andreyknvl@...il.com>
Cc: ryabinin.a.a@...il.com, glider@...gle.com, dvyukov@...gle.com,
	vincenzo.frascino@....com, corbet@....net, catalin.marinas@....com,
	will@...nel.org, akpm@...ux-foundation.org,
	scott@...amperecomputing.com, jhubbard@...dia.com,
	pankaj.gupta@....com, leitao@...ian.org, kaleshsingh@...gle.com,
	maz@...nel.org, broonie@...nel.org, oliver.upton@...ux.dev,
	james.morse@....com, ardb@...nel.org,
	hardevsinh.palaniya@...iconsignals.io, david@...hat.com,
	yang@...amperecomputing.com, kasan-dev@...glegroups.com,
	workflows@...r.kernel.org, linux-doc@...r.kernel.org,
	linux-kernel@...r.kernel.org, linux-arm-kernel@...ts.infradead.org,
	linux-mm@...ck.org
Subject: Re: [PATCH 2/2] kasan: apply store-only mode in kasan kunit testcases

Hi Andrey,

> >
> > When KASAN is configured in store-only mode,
> > fetch/load operations do not trigger tag check faults.
> > As a result, the outcome of some test cases may differ
> > compared to when KASAN is configured without store-only mode.
> >
> > To address this:
> >   1. Replace fetch/load expressions that would
> >      normally trigger tag check faults with store operation
> >      when running under store-only and sync mode.
> >      In case of async/asymm mode, skip the store operation triggering
> >      tag check fault since it corrupts memory.
> >
> >   2. Skip some testcases affected by initial value
> >      (i.e) atomic_cmpxchg() testcase maybe successd if
> >      it passes valid atomic_t address and invalid oldaval address.
> >      In this case, if invalid atomic_t doesn't have the same oldval,
> >      it won't trigger store operation so the test will pass.
> >
> > Signed-off-by: Yeoreum Yun <yeoreum.yun@....com>
> > ---
> >  mm/kasan/kasan_test_c.c | 423 ++++++++++++++++++++++++++++++++--------
> >  1 file changed, 341 insertions(+), 82 deletions(-)
> >
> > diff --git a/mm/kasan/kasan_test_c.c b/mm/kasan/kasan_test_c.c
> > index 2aa12dfa427a..22d5d6d6cd9f 100644
> > --- a/mm/kasan/kasan_test_c.c
> > +++ b/mm/kasan/kasan_test_c.c
> > @@ -94,11 +94,13 @@ static void kasan_test_exit(struct kunit *test)
> >  }
> >
> >  /**
> > - * KUNIT_EXPECT_KASAN_FAIL - check that the executed expression produces a
> > - * KASAN report; causes a KUnit test failure otherwise.
> > + * _KUNIT_EXPECT_KASAN_TEMPLATE - check that the executed expression produces
> > + * a KASAN report or not; a KUnit test failure when it's different from @produce.
> >   *
> >   * @test: Currently executing KUnit test.
> > - * @expression: Expression that must produce a KASAN report.
> > + * @expr: Expression produce a KASAN report or not.
> > + * @expr_str: Expression string
> > + * @produce: expression should produce a KASAN report.
> >   *
> >   * For hardware tag-based KASAN, when a synchronous tag fault happens, tag
> >   * checking is auto-disabled. When this happens, this test handler reenables
> > @@ -110,25 +112,29 @@ static void kasan_test_exit(struct kunit *test)
> >   * Use READ/WRITE_ONCE() for the accesses and compiler barriers around the
> >   * expression to prevent that.
> >   *
> > - * In between KUNIT_EXPECT_KASAN_FAIL checks, test_status.report_found is kept
> > + * In between _KUNIT_EXPECT_KASAN_TEMPLATE checks, test_status.report_found is kept
> >   * as false. This allows detecting KASAN reports that happen outside of the
> >   * checks by asserting !test_status.report_found at the start of
> > - * KUNIT_EXPECT_KASAN_FAIL and in kasan_test_exit.
> > + * _KUNIT_EXPECT_KASAN_TEMPLATE and in kasan_test_exit.
> >   */
> > -#define KUNIT_EXPECT_KASAN_FAIL(test, expression) do {                 \
> > +#define _KUNIT_EXPECT_KASAN_TEMPLATE(test, expr, expr_str, produce)    \
> > +do {                                                                   \
> >         if (IS_ENABLED(CONFIG_KASAN_HW_TAGS) &&                         \
> >             kasan_sync_fault_possible())                                \
> >                 migrate_disable();                                      \
> >         KUNIT_EXPECT_FALSE(test, READ_ONCE(test_status.report_found));  \
> >         barrier();                                                      \
> > -       expression;                                                     \
> > +       expr;                                                           \
> >         barrier();                                                      \
> >         if (kasan_async_fault_possible())                               \
> >                 kasan_force_async_fault();                              \
> > -       if (!READ_ONCE(test_status.report_found)) {                     \
> > -               KUNIT_FAIL(test, KUNIT_SUBTEST_INDENT "KASAN failure "  \
> > -                               "expected in \"" #expression            \
> > -                                "\", but none occurred");              \
> > +       if (READ_ONCE(test_status.report_found) != produce) {           \
> > +               KUNIT_FAIL(test, KUNIT_SUBTEST_INDENT "KASAN %s "       \
> > +                               "expected in \"" expr_str               \
> > +                                "\", but %soccurred",                  \
> > +                               (produce ? "failure" : "success"),      \
> > +                               (test_status.report_found ?             \
> > +                                "" : "none "));                        \
> >         }                                                               \
> >         if (IS_ENABLED(CONFIG_KASAN_HW_TAGS) &&                         \
> >             kasan_sync_fault_possible()) {                              \
> > @@ -141,6 +147,26 @@ static void kasan_test_exit(struct kunit *test)
> >         WRITE_ONCE(test_status.async_fault, false);                     \
> >  } while (0)
> >
> > +/*
> > + * KUNIT_EXPECT_KASAN_FAIL - check that the executed expression produces a
> > + * KASAN report; causes a KUnit test failure otherwise.
> > + *
> > + * @test: Currently executing KUnit test.
> > + * @expr: Expression produce a KASAN report.
> > + */
> > +#define KUNIT_EXPECT_KASAN_FAIL(test, expr)                    \
> > +       _KUNIT_EXPECT_KASAN_TEMPLATE(test, expr, #expr, true)
> > +
> > +/*
> > + * KUNIT_EXPECT_KASAN_SUCCESS - check that the executed expression doesn't
> > + * produces a KASAN report; causes a KUnit test failure otherwise.
> > + *
> > + * @test: Currently executing KUnit test.
> > + * @expr: Expression doesn't produce a KASAN report.
> > + */
> > +#define KUNIT_EXPECT_KASAN_SUCCESS(test, expr)                 \
> > +       _KUNIT_EXPECT_KASAN_TEMPLATE(test, expr, #expr, false)
> > +
> >  #define KASAN_TEST_NEEDS_CONFIG_ON(test, config) do {                  \
> >         if (!IS_ENABLED(config))                                        \
> >                 kunit_skip((test), "Test requires " #config "=y");      \
> > @@ -183,8 +209,15 @@ static void kmalloc_oob_right(struct kunit *test)
> >         KUNIT_EXPECT_KASAN_FAIL(test, ptr[size + 5] = 'y');
> >
> >         /* Out-of-bounds access past the aligned kmalloc object. */
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] =
> > -                                       ptr[size + KASAN_GRANULE_SIZE + 5]);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ptr[0] =
> > +                                               ptr[size + KASAN_GRANULE_SIZE + 5]);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test,
> > +                                       ptr[size + KASAN_GRANULE_SIZE + 5] = ptr[0]);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] =
> > +                                               ptr[size + KASAN_GRANULE_SIZE + 5]);
> >
> >         kfree(ptr);
> >  }
> > @@ -198,7 +231,13 @@ static void kmalloc_oob_left(struct kunit *test)
> >         KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr);
> >
> >         OPTIMIZER_HIDE_VAR(ptr);
> > -       KUNIT_EXPECT_KASAN_FAIL(test, *ptr = *(ptr - 1));
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, *ptr = *(ptr - 1));
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, *(ptr - 1) = *(ptr));
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, *ptr = *(ptr - 1));
> > +
> >         kfree(ptr);
> >  }
> >
> > @@ -211,7 +250,13 @@ static void kmalloc_node_oob_right(struct kunit *test)
> >         KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr);
> >
> >         OPTIMIZER_HIDE_VAR(ptr);
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] = ptr[size]);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ptr[0] = ptr[size]);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ptr[size] = ptr[0]);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] = ptr[size]);
> > +
> >         kfree(ptr);
> >  }
> >
> > @@ -291,7 +336,12 @@ static void kmalloc_large_uaf(struct kunit *test)
> >         KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr);
> >         kfree(ptr);
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[0]);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0] = 0);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]);
> >  }
> >
> >  static void kmalloc_large_invalid_free(struct kunit *test)
> > @@ -323,7 +373,13 @@ static void page_alloc_oob_right(struct kunit *test)
> >         ptr = page_address(pages);
> >         KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr);
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] = ptr[size]);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ptr[0] = ptr[size]);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ptr[size] = ptr[0]);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ptr[0] = ptr[size]);
> > +
> >         free_pages((unsigned long)ptr, order);
> >  }
> >
> > @@ -338,7 +394,12 @@ static void page_alloc_uaf(struct kunit *test)
> >         KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr);
> >         free_pages((unsigned long)ptr, order);
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[0]);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0] = 0);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]);
> >  }
> >
> >  static void krealloc_more_oob_helper(struct kunit *test,
> > @@ -455,10 +516,15 @@ static void krealloc_uaf(struct kunit *test)
> >         ptr1 = kmalloc(size1, GFP_KERNEL);
> >         KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr1);
> >         kfree(ptr1);
> > -
> >         KUNIT_EXPECT_KASAN_FAIL(test, ptr2 = krealloc(ptr1, size2, GFP_KERNEL));
> >         KUNIT_ASSERT_NULL(test, ptr2);
> > -       KUNIT_EXPECT_KASAN_FAIL(test, *(volatile char *)ptr1);
> > +
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, *(volatile char *)ptr1);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, *(volatile char *)ptr1 = 0);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, *(volatile char *)ptr1);
> >  }
> >
> >  static void kmalloc_oob_16(struct kunit *test)
> > @@ -501,7 +567,13 @@ static void kmalloc_uaf_16(struct kunit *test)
> >         KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr2);
> >         kfree(ptr2);
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, *ptr1 = *ptr2);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, *ptr1 = *ptr2);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, *ptr2 = *ptr1);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, *ptr1 = *ptr2);
> > +
> >         kfree(ptr1);
> >  }
> >
> > @@ -640,8 +712,17 @@ static void kmalloc_memmove_invalid_size(struct kunit *test)
> >         memset((char *)ptr, 0, 64);
> >         OPTIMIZER_HIDE_VAR(ptr);
> >         OPTIMIZER_HIDE_VAR(invalid_size);
> > -       KUNIT_EXPECT_KASAN_FAIL(test,
> > -               memmove((char *)ptr, (char *)ptr + 4, invalid_size));
> > +
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test,
> > +                       memmove((char *)ptr, (char *)ptr + 4, invalid_size));
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test,
> > +                               memmove((char *)ptr + 4, (char *)ptr, invalid_size));
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test,
> > +                       memmove((char *)ptr, (char *)ptr + 4, invalid_size));
> > +
> >         kfree(ptr);
> >  }
> >
> > @@ -654,7 +735,13 @@ static void kmalloc_uaf(struct kunit *test)
> >         KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr);
> >
> >         kfree(ptr);
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[8]);
> > +
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[8]);
> > +               if (!kasan_sync_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[8] = 0);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[8]);
> >  }
> >
> >  static void kmalloc_uaf_memset(struct kunit *test)
> > @@ -701,7 +788,13 @@ static void kmalloc_uaf2(struct kunit *test)
> >                 goto again;
> >         }
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[40]);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr1)[40]);
> > +               if (!kasan_sync_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[40] = 0);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[40]);
> > +
> >         KUNIT_EXPECT_PTR_NE(test, ptr1, ptr2);
> >
> >         kfree(ptr2);
> > @@ -727,19 +820,35 @@ static void kmalloc_uaf3(struct kunit *test)
> >         KUNIT_ASSERT_NOT_ERR_OR_NULL(test, ptr2);
> >         kfree(ptr2);
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[8]);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr1)[8]);
> > +               if (!kasan_sync_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[8] = 0);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr1)[8]);
> >  }
> >
> >  static void kasan_atomics_helper(struct kunit *test, void *unsafe, void *safe)
> >  {
> >         int *i_unsafe = unsafe;
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, READ_ONCE(*i_unsafe));
> > +       if (kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, READ_ONCE(*i_unsafe));
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, READ_ONCE(*i_unsafe));
> > +
> >         KUNIT_EXPECT_KASAN_FAIL(test, WRITE_ONCE(*i_unsafe, 42));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, smp_load_acquire(i_unsafe));
> > +       if (kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, smp_load_acquire(i_unsafe));
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, smp_load_acquire(i_unsafe));
> >         KUNIT_EXPECT_KASAN_FAIL(test, smp_store_release(i_unsafe, 42));
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_read(unsafe));
> > +       if (kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, atomic_read(unsafe));
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_read(unsafe));
> > +
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_set(unsafe, 42));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_add(42, unsafe));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_sub(42, unsafe));
> > @@ -752,18 +861,38 @@ static void kasan_atomics_helper(struct kunit *test, void *unsafe, void *safe)
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_xchg(unsafe, 42));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_cmpxchg(unsafe, 21, 42));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_try_cmpxchg(unsafe, safe, 42));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_try_cmpxchg(safe, unsafe, 42));
> > +
> > +       /*
> > +        * The result of the test below may vary due to garbage values of unsafe in
> > +        * store-only mode. Therefore, skip this test when KASAN is configured
> > +        * in store-only mode.
> > +        */
> > +       if (!kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_try_cmpxchg(safe, unsafe, 42));
> > +
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_sub_and_test(42, unsafe));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_dec_and_test(unsafe));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_and_test(unsafe));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_add_negative(42, unsafe));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_add_unless(unsafe, 21, 42));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_not_zero(unsafe));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_unless_negative(unsafe));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_dec_unless_positive(unsafe));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_dec_if_positive(unsafe));
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_read(unsafe));
> > +       /*
> > +        * The result of the test below may vary due to garbage values of unsafe in
> > +        * store-only mode. Therefore, skip this test when KASAN is configured
> > +        * in store-only mode.
> > +        */
> > +       if (!kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_add_unless(unsafe, 21, 42));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_not_zero(unsafe));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_inc_unless_negative(unsafe));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_dec_unless_positive(unsafe));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_dec_if_positive(unsafe));
> > +       }
> > +
> > +       if (kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, atomic_long_read(unsafe));
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_read(unsafe));
> > +
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_set(unsafe, 42));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_add(42, unsafe));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_sub(42, unsafe));
> > @@ -776,16 +905,32 @@ static void kasan_atomics_helper(struct kunit *test, void *unsafe, void *safe)
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_xchg(unsafe, 42));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_cmpxchg(unsafe, 21, 42));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_try_cmpxchg(unsafe, safe, 42));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_try_cmpxchg(safe, unsafe, 42));
> > +
> > +       /*
> > +        * The result of the test below may vary due to garbage values in
> > +        * store-only mode. Therefore, skip this test when KASAN is configured
> > +        * in store-only mode.
> > +        */
> > +       if (!kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_try_cmpxchg(safe, unsafe, 42));
> > +
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_sub_and_test(42, unsafe));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_dec_and_test(unsafe));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_inc_and_test(unsafe));
> >         KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_add_negative(42, unsafe));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_add_unless(unsafe, 21, 42));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_inc_not_zero(unsafe));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_inc_unless_negative(unsafe));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_dec_unless_positive(unsafe));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_dec_if_positive(unsafe));
> > +
> > +       /*
> > +        * The result of the test below may vary due to garbage values in
> > +        * store-only mode. Therefore, skip this test when KASAN is configured
> > +        * in store-only mode.
> > +        */
> > +       if (!kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_add_unless(unsafe, 21, 42));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_inc_not_zero(unsafe));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_inc_unless_negative(unsafe));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_dec_unless_positive(unsafe));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, atomic_long_dec_if_positive(unsafe));
> > +       }
> >  }
> >
> >  static void kasan_atomics(struct kunit *test)
> > @@ -842,8 +987,18 @@ static void ksize_unpoisons_memory(struct kunit *test)
> >         /* These must trigger a KASAN report. */
> >         if (IS_ENABLED(CONFIG_KASAN_GENERIC))
> >                 KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size]);
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size + 5]);
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[real_size - 1]);
> > +
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[size + 5]);
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[real_size - 1]);
> > +               if (!kasan_sync_fault_possible()) {
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size + 5] = 0);
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[real_size - 1] = 0);
> > +               }
> > +       } else {
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size + 5]);
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[real_size - 1]);
> > +       }
> >
> >         kfree(ptr);
> >  }
> > @@ -863,8 +1018,17 @@ static void ksize_uaf(struct kunit *test)
> >
> >         OPTIMIZER_HIDE_VAR(ptr);
> >         KUNIT_EXPECT_KASAN_FAIL(test, ksize(ptr));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]);
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size]);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[0]);
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[size]);
> > +               if (!kasan_sync_fault_possible()) {
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0] = 0);
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size] = 0);
> > +               }
> > +       } else {
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]);
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[size]);
> > +       }
> >  }
> >
> >  /*
> > @@ -886,7 +1050,11 @@ static void rcu_uaf_reclaim(struct rcu_head *rp)
> >                 container_of(rp, struct kasan_rcu_info, rcu);
> >
> >         kfree(fp);
> > -       ((volatile struct kasan_rcu_info *)fp)->i;
> > +
> > +       if (kasan_stonly_enabled() && !kasan_async_fault_possible())
> > +               ((volatile struct kasan_rcu_info *)fp)->i = 0;
> > +       else
> > +               ((volatile struct kasan_rcu_info *)fp)->i;
> >  }
> >
> >  static void rcu_uaf(struct kunit *test)
> > @@ -899,9 +1067,14 @@ static void rcu_uaf(struct kunit *test)
> >         global_rcu_ptr = rcu_dereference_protected(
> >                                 (struct kasan_rcu_info __rcu *)ptr, NULL);
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test,
> > -               call_rcu(&global_rcu_ptr->rcu, rcu_uaf_reclaim);
> > -               rcu_barrier());
> > +       if (kasan_stonly_enabled() && kasan_async_fault_possible())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test,
> > +                       call_rcu(&global_rcu_ptr->rcu, rcu_uaf_reclaim);
> > +                       rcu_barrier());
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test,
> > +                       call_rcu(&global_rcu_ptr->rcu, rcu_uaf_reclaim);
> > +                       rcu_barrier());
> >  }
> >
> >  static void workqueue_uaf_work(struct work_struct *work)
> > @@ -924,8 +1097,12 @@ static void workqueue_uaf(struct kunit *test)
> >         queue_work(workqueue, work);
> >         destroy_workqueue(workqueue);
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test,
> > -               ((volatile struct work_struct *)work)->data);
> > +       if (kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test,
> > +                       ((volatile struct work_struct *)work)->data);
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test,
> > +                       ((volatile struct work_struct *)work)->data);
> >  }
> >
> >  static void kfree_via_page(struct kunit *test)
> > @@ -972,7 +1149,12 @@ static void kmem_cache_oob(struct kunit *test)
> >                 return;
> >         }
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, *p = p[size + OOB_TAG_OFF]);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, *p = p[size + OOB_TAG_OFF]);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, p[size + OOB_TAG_OFF] = *p);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, *p = p[size + OOB_TAG_OFF]);
> >
> >         kmem_cache_free(cache, p);
> >         kmem_cache_destroy(cache);
> > @@ -1068,7 +1250,12 @@ static void kmem_cache_rcu_uaf(struct kunit *test)
> >          */
> >         rcu_barrier();
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, READ_ONCE(*p));
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, READ_ONCE(*p));
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, WRITE_ONCE(*p, 0));
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, READ_ONCE(*p));
> >
> >         kmem_cache_destroy(cache);
> >  }
> > @@ -1206,7 +1393,13 @@ static void mempool_oob_right_helper(struct kunit *test, mempool_t *pool, size_t
> >         if (IS_ENABLED(CONFIG_KASAN_GENERIC))
> >                 KUNIT_EXPECT_KASAN_FAIL(test,
> >                         ((volatile char *)&elem[size])[0]);
> > -       else
> > +       else if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test,
> > +                       ((volatile char *)&elem[round_up(size, KASAN_GRANULE_SIZE)])[0]);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test,
> > +                               ((volatile char *)&elem[round_up(size, KASAN_GRANULE_SIZE)])[0] = 0);
> > +       } else
> >                 KUNIT_EXPECT_KASAN_FAIL(test,
> >                         ((volatile char *)&elem[round_up(size, KASAN_GRANULE_SIZE)])[0]);
> >
> > @@ -1273,7 +1466,13 @@ static void mempool_uaf_helper(struct kunit *test, mempool_t *pool, bool page)
> >         mempool_free(elem, pool);
> >
> >         ptr = page ? page_address((struct page *)elem) : elem;
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]);
> > +
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)ptr)[0]);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0] = 0);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)ptr)[0]);
> >  }
> >
> >  static void mempool_kmalloc_uaf(struct kunit *test)
> > @@ -1532,8 +1731,13 @@ static void kasan_memchr(struct kunit *test)
> >
> >         OPTIMIZER_HIDE_VAR(ptr);
> >         OPTIMIZER_HIDE_VAR(size);
> > -       KUNIT_EXPECT_KASAN_FAIL(test,
> > -               kasan_ptr_result = memchr(ptr, '1', size + 1));
> > +
> > +       if (kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test,
> > +                       kasan_ptr_result = memchr(ptr, '1', size + 1));
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test,
> > +                       kasan_ptr_result = memchr(ptr, '1', size + 1));
> >
> >         kfree(ptr);
> >  }
> > @@ -1559,8 +1763,14 @@ static void kasan_memcmp(struct kunit *test)
> >
> >         OPTIMIZER_HIDE_VAR(ptr);
> >         OPTIMIZER_HIDE_VAR(size);
> > -       KUNIT_EXPECT_KASAN_FAIL(test,
> > -               kasan_int_result = memcmp(ptr, arr, size+1));
> > +
> > +       if (kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test,
> > +                       kasan_int_result = memcmp(ptr, arr, size+1));
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test,
> > +                       kasan_int_result = memcmp(ptr, arr, size+1));
> > +
> >         kfree(ptr);
> >  }
> >
> > @@ -1593,9 +1803,16 @@ static void kasan_strings(struct kunit *test)
> >         KUNIT_EXPECT_EQ(test, KASAN_GRANULE_SIZE - 2,
> >                         strscpy(ptr, src + 1, KASAN_GRANULE_SIZE));
> >
> > -       /* strscpy should fail if the first byte is unreadable. */
> > -       KUNIT_EXPECT_KASAN_FAIL(test, strscpy(ptr, src + KASAN_GRANULE_SIZE,
> > -                                             KASAN_GRANULE_SIZE));
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, strscpy(ptr, src + KASAN_GRANULE_SIZE,
> > +                                                     KASAN_GRANULE_SIZE));
> > +               if (!kasan_async_fault_possible())
> > +                       /* strscpy should fail when the first byte is to be written. */
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, strscpy(ptr + size, src, KASAN_GRANULE_SIZE));
> > +       } else
> > +               /* strscpy should fail if the first byte is unreadable. */
> > +               KUNIT_EXPECT_KASAN_FAIL(test, strscpy(ptr, src + KASAN_GRANULE_SIZE,
> > +                                                     KASAN_GRANULE_SIZE));
> >
> >         kfree(src);
> >         kfree(ptr);
> > @@ -1607,17 +1824,22 @@ static void kasan_strings(struct kunit *test)
> >          * will likely point to zeroed byte.
> >          */
> >         ptr += 16;
> > -       KUNIT_EXPECT_KASAN_FAIL(test, kasan_ptr_result = strchr(ptr, '1'));
> >
> > -       KUNIT_EXPECT_KASAN_FAIL(test, kasan_ptr_result = strrchr(ptr, '1'));
> > -
> > -       KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strcmp(ptr, "2"));
> > -
> > -       KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strncmp(ptr, "2", 1));
> > -
> > -       KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strlen(ptr));
> > -
> > -       KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strnlen(ptr, 1));
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_ptr_result = strchr(ptr, '1'));
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_ptr_result = strrchr(ptr, '1'));
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_int_result = strcmp(ptr, "2"));
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_int_result = strncmp(ptr, "2", 1));
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_int_result = strlen(ptr));
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_int_result = strnlen(ptr, 1));
> > +       } else {
> > +               KUNIT_EXPECT_KASAN_FAIL(test, kasan_ptr_result = strchr(ptr, '1'));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, kasan_ptr_result = strrchr(ptr, '1'));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strcmp(ptr, "2"));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strncmp(ptr, "2", 1));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strlen(ptr));
> > +               KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = strnlen(ptr, 1));
> > +       }
> >  }
> >
> >  static void kasan_bitops_modify(struct kunit *test, int nr, void *addr)
> > @@ -1636,12 +1858,27 @@ static void kasan_bitops_test_and_modify(struct kunit *test, int nr, void *addr)
> >  {
> >         KUNIT_EXPECT_KASAN_FAIL(test, test_and_set_bit(nr, addr));
> >         KUNIT_EXPECT_KASAN_FAIL(test, __test_and_set_bit(nr, addr));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, test_and_set_bit_lock(nr, addr));
> > +
> > +       /*
> > +        * When KASAN is running in store-only mode,
> > +        * a fault won't occur even if the bit is set.
> > +        * Therefore, skip the test_and_set_bit_lock test in store-only mode.
> > +        */
> > +       if (!kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_FAIL(test, test_and_set_bit_lock(nr, addr));
> > +
> >         KUNIT_EXPECT_KASAN_FAIL(test, test_and_clear_bit(nr, addr));
> >         KUNIT_EXPECT_KASAN_FAIL(test, __test_and_clear_bit(nr, addr));
> >         KUNIT_EXPECT_KASAN_FAIL(test, test_and_change_bit(nr, addr));
> >         KUNIT_EXPECT_KASAN_FAIL(test, __test_and_change_bit(nr, addr));
> > -       KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = test_bit(nr, addr));
> > +
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, kasan_int_result = test_bit(nr, addr));
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, set_bit(nr, addr));
> > +  } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result = test_bit(nr, addr));
> > +
> >         if (nr < 7)
> >                 KUNIT_EXPECT_KASAN_FAIL(test, kasan_int_result =
> >                                 xor_unlock_is_negative_byte(1 << nr, addr));
> > @@ -1765,7 +2002,12 @@ static void vmalloc_oob(struct kunit *test)
> >                 KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)v_ptr)[size]);
> >
> >         /* An aligned access into the first out-of-bounds granule. */
> > -       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)v_ptr)[size + 5]);
> > +       if (kasan_stonly_enabled()) {
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test, ((volatile char *)v_ptr)[size + 5]);
> > +               if (!kasan_async_fault_possible())
> > +                       KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)v_ptr)[size + 5] = 0);
> > +       } else
> > +               KUNIT_EXPECT_KASAN_FAIL(test, ((volatile char *)v_ptr)[size + 5]);
> >
> >         /* Check that in-bounds accesses to the physical page are valid. */
> >         page = vmalloc_to_page(v_ptr);
> > @@ -2042,16 +2284,33 @@ static void copy_user_test_oob(struct kunit *test)
> >
> >         KUNIT_EXPECT_KASAN_FAIL(test,
> >                 unused = copy_from_user(kmem, usermem, size + 1));
> > -       KUNIT_EXPECT_KASAN_FAIL(test,
> > -               unused = copy_to_user(usermem, kmem, size + 1));
> > +
> > +       if (kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test,
> > +                       unused = copy_to_user(usermem, kmem, size + 1));
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test,
> > +                       unused = copy_to_user(usermem, kmem, size + 1));
> > +
> >         KUNIT_EXPECT_KASAN_FAIL(test,
> >                 unused = __copy_from_user(kmem, usermem, size + 1));
> > -       KUNIT_EXPECT_KASAN_FAIL(test,
> > -               unused = __copy_to_user(usermem, kmem, size + 1));
> > +
> > +       if (kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test,
> > +                       unused = __copy_to_user(usermem, kmem, size + 1));
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test,
> > +                       unused = __copy_to_user(usermem, kmem, size + 1));
> > +
> >         KUNIT_EXPECT_KASAN_FAIL(test,
> >                 unused = __copy_from_user_inatomic(kmem, usermem, size + 1));
> > -       KUNIT_EXPECT_KASAN_FAIL(test,
> > -               unused = __copy_to_user_inatomic(usermem, kmem, size + 1));
> > +
> > +       if (kasan_stonly_enabled())
> > +               KUNIT_EXPECT_KASAN_SUCCESS(test,
> > +                       unused = __copy_to_user_inatomic(usermem, kmem, size + 1));
> > +       else
> > +               KUNIT_EXPECT_KASAN_FAIL(test,
> > +                       unused = __copy_to_user_inatomic(usermem, kmem, size + 1));
> >
> >         /*
> >         * Prepare a long string in usermem to avoid the strncpy_from_user test
> > --
> > LEVI:{C3F47F37-75D8-414A-A8BA-3980EC8A46D7}
> >
>
> This patch does not look good.
>
> Right now, KASAN tests are crafted to avoid/self-contain harmful
> memory corruptions that they do (e.g. make sure that OOB write
> accesses land in in-object kmalloc training space, etc.). If you turn
> read accesses in tests into write accesses, memory corruptions caused
> by the earlier tests will crash the kernel or the latter tests.

That's why I run the store-only test when this mode is "sync"
In case of "async/asymm" as you mention since it reports "after",
there will be memory corruption.

But in case of sync, when the MTE fault happens, it doesn't
write to memory so, I think it's fine.

>
> The easiest thing to do for now is to disable the tests that check bad
> read accesses when store-only is enabled.
>
> If we want to convert tests into doing write accesses instead of
> reads, this needs to be done separately for each test (i.e. via a
> separate patch) with an explanation why doing this is safe (and
> adjustments whenever it's not). And we need a better way to code this
> instead of the horrifying number of if/else checks.
>
> Thank you!

Hmm, as I mention above, the testcase with store-only/sync mode seems to fine.
But, If the "testcase" is failed, as you mention it makes a memory
corruption.

If success case is fine,
Please let me make all related story-only case be seperated
to each function (but almost simliar to pre-exist testcase) with
sync mode otherwise, let me seperate them just checking it whether
it success when it accesses to invalid memory with read/fetch.

Thanks :)

--
Sincerely,
Yeoreum Yun

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ