Sounds good Damien. Thanks for reviewing! On Thu, Aug 23, 2018 at 6:12 PM, Damien Le Moal <Damien.LeMoal@xxxxxxx> wrote: > John, > > On 2018/08/23 10:37, John Pittman wrote: >> The API surrounding refcount_t should be used in place of atomic_t >> when variables are being used as reference counters. This API can >> prevent issues such as counter overflows and use-after-free >> conditions. Within the dm zoned metadata stack, the atomic_t API >> is used for mblk->ref and zone->refcount. Change these to use >> refcount_t, avoiding the issues mentioned. >> >> Signed-off-by: John Pittman <jpittman@xxxxxxxxxx> >> --- >> drivers/md/dm-zoned-metadata.c | 25 +++++++++++++------------ >> drivers/md/dm-zoned.h | 2 +- >> 2 files changed, 14 insertions(+), 13 deletions(-) >> >> diff --git a/drivers/md/dm-zoned-metadata.c b/drivers/md/dm-zoned-metadata.c >> index 969954915566..92e635749414 100644 >> --- a/drivers/md/dm-zoned-metadata.c >> +++ b/drivers/md/dm-zoned-metadata.c >> @@ -99,7 +99,7 @@ struct dmz_mblock { >> struct rb_node node; >> struct list_head link; >> sector_t no; >> - atomic_t ref; >> + refcount_t ref; > > While reviewing your patch, I realized that this ref is always manipulated under > the zmd->mblk_lock spinlock. So there is no need for it to be an atomic or a > refcount. An unsigned int would do as well and be faster. My fault. > > I will send a patch to go on top of yours to fix that. > > Otherwise: > > Reviewed-by: Damien Le Moal <damien.lemoal@xxxxxxx> > Tested-by: Damien Le Moal <damien.lemoal@xxxxxxx> > > Thanks ! > > >> unsigned long state; >> struct page *page; >> void *data; >> @@ -296,7 +296,7 @@ static struct dmz_mblock *dmz_alloc_mblock(struct dmz_metadata *zmd, >> >> RB_CLEAR_NODE(&mblk->node); >> INIT_LIST_HEAD(&mblk->link); >> - atomic_set(&mblk->ref, 0); >> + refcount_set(&mblk->ref, 0); >> mblk->state = 0; >> mblk->no = mblk_no; >> mblk->data = page_address(mblk->page); >> @@ -397,7 +397,7 @@ static struct dmz_mblock *dmz_fetch_mblock(struct dmz_metadata *zmd, >> return NULL; >> >> spin_lock(&zmd->mblk_lock); >> - atomic_inc(&mblk->ref); >> + refcount_inc(&mblk->ref); >> set_bit(DMZ_META_READING, &mblk->state); >> dmz_insert_mblock(zmd, mblk); >> spin_unlock(&zmd->mblk_lock); >> @@ -484,7 +484,7 @@ static void dmz_release_mblock(struct dmz_metadata *zmd, >> >> spin_lock(&zmd->mblk_lock); >> >> - if (atomic_dec_and_test(&mblk->ref)) { >> + if (refcount_dec_and_test(&mblk->ref)) { >> if (test_bit(DMZ_META_ERROR, &mblk->state)) { >> rb_erase(&mblk->node, &zmd->mblk_rbtree); >> dmz_free_mblock(zmd, mblk); >> @@ -511,7 +511,8 @@ static struct dmz_mblock *dmz_get_mblock(struct dmz_metadata *zmd, >> mblk = dmz_lookup_mblock(zmd, mblk_no); >> if (mblk) { >> /* Cache hit: remove block from LRU list */ >> - if (atomic_inc_return(&mblk->ref) == 1 && >> + refcount_inc(&mblk->ref); >> + if (refcount_read(&mblk->ref) == 1 && >> !test_bit(DMZ_META_DIRTY, &mblk->state)) >> list_del_init(&mblk->link); >> } >> @@ -753,7 +754,7 @@ int dmz_flush_metadata(struct dmz_metadata *zmd) >> >> spin_lock(&zmd->mblk_lock); >> clear_bit(DMZ_META_DIRTY, &mblk->state); >> - if (atomic_read(&mblk->ref) == 0) >> + if (refcount_read(&mblk->ref) == 0) >> list_add_tail(&mblk->link, &zmd->mblk_lru_list); >> spin_unlock(&zmd->mblk_lock); >> } >> @@ -1048,7 +1049,7 @@ static int dmz_init_zone(struct dmz_metadata *zmd, struct dm_zone *zone, >> } >> >> INIT_LIST_HEAD(&zone->link); >> - atomic_set(&zone->refcount, 0); >> + refcount_set(&zone->refcount, 0); >> zone->chunk = DMZ_MAP_UNMAPPED; >> >> if (blkz->type == BLK_ZONE_TYPE_CONVENTIONAL) { >> @@ -1574,7 +1575,7 @@ struct dm_zone *dmz_get_zone_for_reclaim(struct dmz_metadata *zmd) >> void dmz_activate_zone(struct dm_zone *zone) >> { >> set_bit(DMZ_ACTIVE, &zone->flags); >> - atomic_inc(&zone->refcount); >> + refcount_inc(&zone->refcount); >> } >> >> /* >> @@ -1585,7 +1586,7 @@ void dmz_activate_zone(struct dm_zone *zone) >> */ >> void dmz_deactivate_zone(struct dm_zone *zone) >> { >> - if (atomic_dec_and_test(&zone->refcount)) { >> + if (refcount_dec_and_test(&zone->refcount)) { >> WARN_ON(!test_bit(DMZ_ACTIVE, &zone->flags)); >> clear_bit_unlock(DMZ_ACTIVE, &zone->flags); >> smp_mb__after_atomic(); >> @@ -2308,7 +2309,7 @@ static void dmz_cleanup_metadata(struct dmz_metadata *zmd) >> mblk = list_first_entry(&zmd->mblk_dirty_list, >> struct dmz_mblock, link); >> dmz_dev_warn(zmd->dev, "mblock %llu still in dirty list (ref %u)", >> - (u64)mblk->no, atomic_read(&mblk->ref)); >> + (u64)mblk->no, refcount_read(&mblk->ref)); >> list_del_init(&mblk->link); >> rb_erase(&mblk->node, &zmd->mblk_rbtree); >> dmz_free_mblock(zmd, mblk); >> @@ -2326,8 +2327,8 @@ static void dmz_cleanup_metadata(struct dmz_metadata *zmd) >> root = &zmd->mblk_rbtree; >> rbtree_postorder_for_each_entry_safe(mblk, next, root, node) { >> dmz_dev_warn(zmd->dev, "mblock %llu ref %u still in rbtree", >> - (u64)mblk->no, atomic_read(&mblk->ref)); >> - atomic_set(&mblk->ref, 0); >> + (u64)mblk->no, refcount_read(&mblk->ref)); >> + refcount_set(&mblk->ref, 0); >> dmz_free_mblock(zmd, mblk); >> } >> >> diff --git a/drivers/md/dm-zoned.h b/drivers/md/dm-zoned.h >> index 12419f0bfe78..b7829a615d26 100644 >> --- a/drivers/md/dm-zoned.h >> +++ b/drivers/md/dm-zoned.h >> @@ -78,7 +78,7 @@ struct dm_zone { >> unsigned long flags; >> >> /* Zone activation reference count */ >> - atomic_t refcount; >> + refcount_t refcount; >> >> /* Zone write pointer block (relative to the zone start block) */ >> unsigned int wp_block; >> > > > -- > Damien Le Moal > Western Digital Research -- dm-devel mailing list dm-devel@xxxxxxxxxx https://www.redhat.com/mailman/listinfo/dm-devel