Re: [PATCH 2/3] dm zoned: metadata: use refcount_t for dm zoned reference counters

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Sounds good Damien.  Thanks for reviewing!

On Thu, Aug 23, 2018 at 6:12 PM, Damien Le Moal <Damien.LeMoal@xxxxxxx> wrote:
> John,
>
> On 2018/08/23 10:37, John Pittman wrote:
>> The API surrounding refcount_t should be used in place of atomic_t
>> when variables are being used as reference counters.  This API can
>> prevent issues such as counter overflows and use-after-free
>> conditions.  Within the dm zoned metadata stack, the atomic_t API
>> is used for mblk->ref and zone->refcount.  Change these to use
>> refcount_t, avoiding the issues mentioned.
>>
>> Signed-off-by: John Pittman <jpittman@xxxxxxxxxx>
>> ---
>>  drivers/md/dm-zoned-metadata.c | 25 +++++++++++++------------
>>  drivers/md/dm-zoned.h          |  2 +-
>>  2 files changed, 14 insertions(+), 13 deletions(-)
>>
>> diff --git a/drivers/md/dm-zoned-metadata.c b/drivers/md/dm-zoned-metadata.c
>> index 969954915566..92e635749414 100644
>> --- a/drivers/md/dm-zoned-metadata.c
>> +++ b/drivers/md/dm-zoned-metadata.c
>> @@ -99,7 +99,7 @@ struct dmz_mblock {
>>       struct rb_node          node;
>>       struct list_head        link;
>>       sector_t                no;
>> -     atomic_t                ref;
>> +     refcount_t              ref;
>
> While reviewing your patch, I realized that this ref is always manipulated under
> the zmd->mblk_lock spinlock. So there is no need for it to be an atomic or a
> refcount. An unsigned int would do as well and be faster. My fault.
>
> I will send a patch to go on top of yours to fix that.
>
> Otherwise:
>
> Reviewed-by: Damien Le Moal <damien.lemoal@xxxxxxx>
> Tested-by: Damien Le Moal <damien.lemoal@xxxxxxx>
>
> Thanks !
>
>
>>       unsigned long           state;
>>       struct page             *page;
>>       void                    *data;
>> @@ -296,7 +296,7 @@ static struct dmz_mblock *dmz_alloc_mblock(struct dmz_metadata *zmd,
>>
>>       RB_CLEAR_NODE(&mblk->node);
>>       INIT_LIST_HEAD(&mblk->link);
>> -     atomic_set(&mblk->ref, 0);
>> +     refcount_set(&mblk->ref, 0);
>>       mblk->state = 0;
>>       mblk->no = mblk_no;
>>       mblk->data = page_address(mblk->page);
>> @@ -397,7 +397,7 @@ static struct dmz_mblock *dmz_fetch_mblock(struct dmz_metadata *zmd,
>>               return NULL;
>>
>>       spin_lock(&zmd->mblk_lock);
>> -     atomic_inc(&mblk->ref);
>> +     refcount_inc(&mblk->ref);
>>       set_bit(DMZ_META_READING, &mblk->state);
>>       dmz_insert_mblock(zmd, mblk);
>>       spin_unlock(&zmd->mblk_lock);
>> @@ -484,7 +484,7 @@ static void dmz_release_mblock(struct dmz_metadata *zmd,
>>
>>       spin_lock(&zmd->mblk_lock);
>>
>> -     if (atomic_dec_and_test(&mblk->ref)) {
>> +     if (refcount_dec_and_test(&mblk->ref)) {
>>               if (test_bit(DMZ_META_ERROR, &mblk->state)) {
>>                       rb_erase(&mblk->node, &zmd->mblk_rbtree);
>>                       dmz_free_mblock(zmd, mblk);
>> @@ -511,7 +511,8 @@ static struct dmz_mblock *dmz_get_mblock(struct dmz_metadata *zmd,
>>       mblk = dmz_lookup_mblock(zmd, mblk_no);
>>       if (mblk) {
>>               /* Cache hit: remove block from LRU list */
>> -             if (atomic_inc_return(&mblk->ref) == 1 &&
>> +             refcount_inc(&mblk->ref);
>> +             if (refcount_read(&mblk->ref) == 1 &&
>>                   !test_bit(DMZ_META_DIRTY, &mblk->state))
>>                       list_del_init(&mblk->link);
>>       }
>> @@ -753,7 +754,7 @@ int dmz_flush_metadata(struct dmz_metadata *zmd)
>>
>>               spin_lock(&zmd->mblk_lock);
>>               clear_bit(DMZ_META_DIRTY, &mblk->state);
>> -             if (atomic_read(&mblk->ref) == 0)
>> +             if (refcount_read(&mblk->ref) == 0)
>>                       list_add_tail(&mblk->link, &zmd->mblk_lru_list);
>>               spin_unlock(&zmd->mblk_lock);
>>       }
>> @@ -1048,7 +1049,7 @@ static int dmz_init_zone(struct dmz_metadata *zmd, struct dm_zone *zone,
>>       }
>>
>>       INIT_LIST_HEAD(&zone->link);
>> -     atomic_set(&zone->refcount, 0);
>> +     refcount_set(&zone->refcount, 0);
>>       zone->chunk = DMZ_MAP_UNMAPPED;
>>
>>       if (blkz->type == BLK_ZONE_TYPE_CONVENTIONAL) {
>> @@ -1574,7 +1575,7 @@ struct dm_zone *dmz_get_zone_for_reclaim(struct dmz_metadata *zmd)
>>  void dmz_activate_zone(struct dm_zone *zone)
>>  {
>>       set_bit(DMZ_ACTIVE, &zone->flags);
>> -     atomic_inc(&zone->refcount);
>> +     refcount_inc(&zone->refcount);
>>  }
>>
>>  /*
>> @@ -1585,7 +1586,7 @@ void dmz_activate_zone(struct dm_zone *zone)
>>   */
>>  void dmz_deactivate_zone(struct dm_zone *zone)
>>  {
>> -     if (atomic_dec_and_test(&zone->refcount)) {
>> +     if (refcount_dec_and_test(&zone->refcount)) {
>>               WARN_ON(!test_bit(DMZ_ACTIVE, &zone->flags));
>>               clear_bit_unlock(DMZ_ACTIVE, &zone->flags);
>>               smp_mb__after_atomic();
>> @@ -2308,7 +2309,7 @@ static void dmz_cleanup_metadata(struct dmz_metadata *zmd)
>>               mblk = list_first_entry(&zmd->mblk_dirty_list,
>>                                       struct dmz_mblock, link);
>>               dmz_dev_warn(zmd->dev, "mblock %llu still in dirty list (ref %u)",
>> -                          (u64)mblk->no, atomic_read(&mblk->ref));
>> +                          (u64)mblk->no, refcount_read(&mblk->ref));
>>               list_del_init(&mblk->link);
>>               rb_erase(&mblk->node, &zmd->mblk_rbtree);
>>               dmz_free_mblock(zmd, mblk);
>> @@ -2326,8 +2327,8 @@ static void dmz_cleanup_metadata(struct dmz_metadata *zmd)
>>       root = &zmd->mblk_rbtree;
>>       rbtree_postorder_for_each_entry_safe(mblk, next, root, node) {
>>               dmz_dev_warn(zmd->dev, "mblock %llu ref %u still in rbtree",
>> -                          (u64)mblk->no, atomic_read(&mblk->ref));
>> -             atomic_set(&mblk->ref, 0);
>> +                          (u64)mblk->no, refcount_read(&mblk->ref));
>> +             refcount_set(&mblk->ref, 0);
>>               dmz_free_mblock(zmd, mblk);
>>       }
>>
>> diff --git a/drivers/md/dm-zoned.h b/drivers/md/dm-zoned.h
>> index 12419f0bfe78..b7829a615d26 100644
>> --- a/drivers/md/dm-zoned.h
>> +++ b/drivers/md/dm-zoned.h
>> @@ -78,7 +78,7 @@ struct dm_zone {
>>       unsigned long           flags;
>>
>>       /* Zone activation reference count */
>> -     atomic_t                refcount;
>> +     refcount_t              refcount;
>>
>>       /* Zone write pointer block (relative to the zone start block) */
>>       unsigned int            wp_block;
>>
>
>
> --
> Damien Le Moal
> Western Digital Research

--
dm-devel mailing list
dm-devel@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/dm-devel



[Index of Archives]     [DM Crypt]     [Fedora Desktop]     [ATA RAID]     [Fedora Marketing]     [Fedora Packaging]     [Fedora SELinux]     [Yosemite Discussion]     [KDE Users]     [Fedora Docs]

  Powered by Linux