Re: [PATCH] timerfd: Protect the might cancel mechanism proper

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jan 31, 2017 at 3:24 PM, Thomas Gleixner <tglx@xxxxxxxxxxxxx> wrote:
> The handling of the might_cancel queueing is not properly protected, so
> parallel operations on the file descriptor can race with each other and
> lead to list corruptions or use after free.
>
> Protect the context for these operations with a seperate lock.
>
> The wait queue lock cannot be reused for this because that would create a
> lock inversion scenario vs. the cancel lock. Replacing might_cancel with an
> atomic (atomic_t or atomic bit) does not help either because it still can
> race vs. the actual list operation.
>
> Reported-by: Dmitry Vyukov <dvyukov@xxxxxxxxxx>
> Signed-off-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
> ---
>  fs/timerfd.c |   17 ++++++++++++++---
>  1 file changed, 14 insertions(+), 3 deletions(-)
>
> --- a/fs/timerfd.c
> +++ b/fs/timerfd.c
> @@ -40,6 +40,7 @@ struct timerfd_ctx {
>         short unsigned settime_flags;   /* to show in fdinfo */
>         struct rcu_head rcu;
>         struct list_head clist;
> +       spinlock_t cancel_lock;
>         bool might_cancel;
>  };
>
> @@ -112,7 +113,7 @@ void timerfd_clock_was_set(void)
>         rcu_read_unlock();
>  }
>
> -static void timerfd_remove_cancel(struct timerfd_ctx *ctx)
> +static void __timerfd_remove_cancel(struct timerfd_ctx *ctx)
>  {
>         if (ctx->might_cancel) {
>                 ctx->might_cancel = false;
> @@ -122,6 +123,13 @@ static void timerfd_remove_cancel(struct
>         }
>  }
>
> +static void timerfd_remove_cancel(struct timerfd_ctx *ctx)
> +{
> +       spin_lock(&ctx->cancel_lock);
> +       __timerfd_remove_cancel(ctx);
> +       spin_unlock(&ctx->cancel_lock);
> +}
> +
>  static bool timerfd_canceled(struct timerfd_ctx *ctx)
>  {
>         if (!ctx->might_cancel || ctx->moffs != KTIME_MAX)
> @@ -132,6 +140,7 @@ static bool timerfd_canceled(struct time
>
>  static void timerfd_setup_cancel(struct timerfd_ctx *ctx, int flags)
>  {
> +       spin_lock(&ctx->cancel_lock);
>         if ((ctx->clockid == CLOCK_REALTIME ||
>              ctx->clockid == CLOCK_REALTIME_ALARM) &&
>             (flags & TFD_TIMER_ABSTIME) && (flags & TFD_TIMER_CANCEL_ON_SET)) {
> @@ -141,9 +150,10 @@ static void timerfd_setup_cancel(struct
>                         list_add_rcu(&ctx->clist, &cancel_list);
>                         spin_unlock(&cancel_lock);
>                 }
> -       } else if (ctx->might_cancel) {
> -               timerfd_remove_cancel(ctx);
> +       } else {
> +               __timerfd_remove_cancel(ctx);
>         }
> +       spin_unlock(&ctx->cancel_lock);
>  }
>
>  static ktime_t timerfd_get_remaining(struct timerfd_ctx *ctx)
> @@ -400,6 +410,7 @@ SYSCALL_DEFINE2(timerfd_create, int, clo
>                 return -ENOMEM;
>
>         init_waitqueue_head(&ctx->wqh);
> +       spin_lock_init(&ctx->cancel_lock);
>         ctx->clockid = clockid;
>
>         if (isalarm(ctx))


Can't we still end up with an inconsistently setup timer?
do_timerfd_settime executes timerfd_setup_cancel and timerfd_setup as
two separate non-atomic actions. So if there are 2 concurrent
timerfd_settime calls, one that needs cancel and another that does not
need cancel, can't we end up with inconsistent setup? E.g. setup timer
that needs cancel, but it won't be in cancel_list. Or vice versa.



[Index of Archives]     [Linux Ext4 Filesystem]     [Union Filesystem]     [Filesystem Testing]     [Ceph Users]     [Ecryptfs]     [AutoFS]     [Kernel Newbies]     [Share Photos]     [Security]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux Cachefs]     [Reiser Filesystem]     [Linux RAID]     [Samba]     [Device Mapper]     [CEPH Development]
  Powered by Linux