On Tue 09-01-18 12:00:14, Reshetova, Elena wrote: > > > On Wed 29-11-17 13:22:20, Elena Reshetova wrote: > > > atomic_t variables are currently used to implement reference > > > counters with the following properties: > > > - counter is initialized to 1 using atomic_set() > > > - a resource is freed upon counter reaching zero > > > - once counter reaches zero, its further > > > increments aren't allowed > > > - counter schema uses basic atomic operations > > > (set, inc, inc_not_zero, dec_and_test, etc.) > > > > > > Such atomic variables should be converted to a newly provided > > > refcount_t type and API that prevents accidental counter overflows > > > and underflows. This is important since overflows and underflows > > > can lead to use-after-free situation and be exploitable. > > > > > > The variable mb_cache_entry.e_refcnt is used as pure reference counter. > > > Convert it to refcount_t and fix up the operations. > > > > > > **Important note for maintainers: > > > > > > Some functions from refcount_t API defined in lib/refcount.c > > > have different memory ordering guarantees than their atomic > > > counterparts. > > > The full comparison can be seen in > > > https://lkml.org/lkml/2017/11/15/57 and it is hopefully soon > > > in state to be merged to the documentation tree. > > > Normally the differences should not matter since refcount_t provides > > > enough guarantees to satisfy the refcounting use cases, but in > > > some rare cases it might matter. > > > Please double check that you don't have some undocumented > > > memory guarantees for this variable usage. > > > > > > For the mb_cache_entry.e_refcnt it might make a difference > > > in following places: > > > - mb_cache_destroy(), mb_cache_shrink() and mb_cache_entry_delete(): > > > decrement in refcount_dec() provides RELEASE ordering vs. fully > > > unordered atomic counterpart. Since the change is for better, it > > > should not matter for these cases. > > > - mb_cache_entry_put(): decrement in refcount_dec_and_test() only > > > provides RELEASE ordering and control dependency on success > > > vs. fully ordered atomic counterpart > > > > > > Suggested-by: Kees Cook <keescook@xxxxxxxxxxxx> > > > Reviewed-by: David Windsor <dwindsor@xxxxxxxxx> > > > Reviewed-by: Hans Liljestrand <ishkamiel@xxxxxxxxx> > > > Signed-off-by: Elena Reshetova <elena.reshetova@xxxxxxxxx> > > > > The patch looks good to me. You can add: > > > > Reviewed-by: Jan Kara <jack@xxxxxxx> > > Thank you Jan! > Would you be able to propagate the patch further into the respective tree? > These patches are best to merge using usual paths. Ted usually merges mbcache patches. Ted? Honza > > > --- > > > fs/mbcache.c | 16 ++++++++-------- > > > include/linux/mbcache.h | 6 +++--- > > > 2 files changed, 11 insertions(+), 11 deletions(-) > > > > > > diff --git a/fs/mbcache.c b/fs/mbcache.c > > > index d818fd2..8bdb52b 100644 > > > --- a/fs/mbcache.c > > > +++ b/fs/mbcache.c > > > @@ -90,7 +90,7 @@ int mb_cache_entry_create(struct mb_cache *cache, gfp_t > > mask, u32 key, > > > > > > INIT_LIST_HEAD(&entry->e_list); > > > /* One ref for hash, one ref returned */ > > > - atomic_set(&entry->e_refcnt, 1); > > > + refcount_set(&entry->e_refcnt, 1); > > > entry->e_key = key; > > > entry->e_value = value; > > > entry->e_reusable = reusable; > > > @@ -109,7 +109,7 @@ int mb_cache_entry_create(struct mb_cache *cache, > > gfp_t mask, u32 key, > > > spin_lock(&cache->c_list_lock); > > > list_add_tail(&entry->e_list, &cache->c_list); > > > /* Grab ref for LRU list */ > > > - atomic_inc(&entry->e_refcnt); > > > + refcount_inc(&entry->e_refcnt); > > > cache->c_entry_count++; > > > spin_unlock(&cache->c_list_lock); > > > > > > @@ -141,7 +141,7 @@ static struct mb_cache_entry *__entry_find(struct > > mb_cache *cache, > > > entry = hlist_bl_entry(node, struct mb_cache_entry, > > > e_hash_list); > > > if (entry->e_key == key && entry->e_reusable) { > > > - atomic_inc(&entry->e_refcnt); > > > + refcount_inc(&entry->e_refcnt); > > > goto out; > > > } > > > node = node->next; > > > @@ -204,7 +204,7 @@ struct mb_cache_entry *mb_cache_entry_get(struct > > mb_cache *cache, u32 key, > > > hlist_bl_lock(head); > > > hlist_bl_for_each_entry(entry, node, head, e_hash_list) { > > > if (entry->e_key == key && entry->e_value == value) { > > > - atomic_inc(&entry->e_refcnt); > > > + refcount_inc(&entry->e_refcnt); > > > goto out; > > > } > > > } > > > @@ -239,7 +239,7 @@ void mb_cache_entry_delete(struct mb_cache *cache, > > u32 key, u64 value) > > > if (!list_empty(&entry->e_list)) { > > > list_del_init(&entry- > > >e_list); > > > cache->c_entry_count--; > > > - atomic_dec(&entry- > > >e_refcnt); > > > + refcount_dec(&entry- > > >e_refcnt); > > > } > > > spin_unlock(&cache->c_list_lock); > > > mb_cache_entry_put(cache, entry); > > > @@ -300,7 +300,7 @@ static unsigned long mb_cache_shrink(struct mb_cache > > *cache, > > > hlist_bl_lock(head); > > > if (!hlist_bl_unhashed(&entry->e_hash_list)) { > > > hlist_bl_del_init(&entry->e_hash_list); > > > - atomic_dec(&entry->e_refcnt); > > > + refcount_dec(&entry->e_refcnt); > > > } > > > hlist_bl_unlock(head); > > > if (mb_cache_entry_put(cache, entry)) > > > @@ -397,11 +397,11 @@ void mb_cache_destroy(struct mb_cache *cache) > > > list_for_each_entry_safe(entry, next, &cache->c_list, e_list) { > > > if (!hlist_bl_unhashed(&entry->e_hash_list)) { > > > hlist_bl_del_init(&entry->e_hash_list); > > > - atomic_dec(&entry->e_refcnt); > > > + refcount_dec(&entry->e_refcnt); > > > } else > > > WARN_ON(1); > > > list_del(&entry->e_list); > > > - WARN_ON(atomic_read(&entry->e_refcnt) != 1); > > > + WARN_ON(refcount_read(&entry->e_refcnt) != 1); > > > mb_cache_entry_put(cache, entry); > > > } > > > kfree(cache->c_hash); > > > diff --git a/include/linux/mbcache.h b/include/linux/mbcache.h > > > index 20f1e3f..0bf2af6 100644 > > > --- a/include/linux/mbcache.h > > > +++ b/include/linux/mbcache.h > > > @@ -5,7 +5,7 @@ > > > #include <linux/hash.h> > > > #include <linux/list_bl.h> > > > #include <linux/list.h> > > > -#include <linux/atomic.h> > > > +#include <linux/refcount.h> > > > #include <linux/fs.h> > > > > > > struct mb_cache; > > > @@ -15,7 +15,7 @@ struct mb_cache_entry { > > > struct list_head e_list; > > > /* Hash table list - protected by hash chain bitlock */ > > > struct hlist_bl_node e_hash_list; > > > - atomic_t e_refcnt; > > > + refcount_t e_refcnt; > > > /* Key in hash - stable during lifetime of the entry */ > > > u32 e_key; > > > u32 e_referenced:1; > > > @@ -33,7 +33,7 @@ void __mb_cache_entry_free(struct mb_cache_entry > > *entry); > > > static inline int mb_cache_entry_put(struct mb_cache *cache, > > > struct > > mb_cache_entry *entry) > > > { > > > - if (!atomic_dec_and_test(&entry->e_refcnt)) > > > + if (!refcount_dec_and_test(&entry->e_refcnt)) > > > return 0; > > > __mb_cache_entry_free(entry); > > > return 1; > > > -- > > > 2.7.4 > > > > > -- > > Jan Kara <jack@xxxxxxxx> > > SUSE Labs, CR -- Jan Kara <jack@xxxxxxxx> SUSE Labs, CR