Brian Foster <bfoster@xxxxxxxxxx> writes: > On Mon, May 08, 2023 at 12:57:57AM +0530, Ritesh Harjani (IBM) wrote: >> This patch moves up and combine the definitions of two functions >> (iomap_iop_set_range_uptodate() & iomap_set_range_uptodate()) into >> iop_set_range_uptodate() & refactors it's arguments a bit. >> >> No functionality change in this patch. >> >> Signed-off-by: Ritesh Harjani (IBM) <ritesh.list@xxxxxxxxx> >> --- > > Hi Ritesh, > > I just have a few random and nitty comments/questions on the series.. > Thanks a lot for reviewing. >> fs/iomap/buffered-io.c | 57 ++++++++++++++++++++---------------------- >> 1 file changed, 27 insertions(+), 30 deletions(-) >> >> diff --git a/fs/iomap/buffered-io.c b/fs/iomap/buffered-io.c >> index cbd945d96584..e732581dc2d4 100644 >> --- a/fs/iomap/buffered-io.c >> +++ b/fs/iomap/buffered-io.c >> @@ -43,6 +43,27 @@ static inline struct iomap_page *to_iomap_page(struct folio *folio) >> >> static struct bio_set iomap_ioend_bioset; >> >> +static void iop_set_range_uptodate(struct inode *inode, struct folio *folio, >> + size_t off, size_t len) >> +{ > > Any particular reason this now takes the inode as a param now instead of > continuing to use the folio? > Mainly for blkbits (or blocksize) and blocks_per_folio. Earlier we were explicitly passing this info, but with recent comments, I felt the api make sense to have inode and folio (given that we want to work on blocksize, blocks_per_folio and folio). -ritesh > Brian > >> + struct iomap_page *iop = to_iomap_page(folio); >> + unsigned int first_blk = off >> inode->i_blkbits; >> + unsigned int last_blk = (off + len - 1) >> inode->i_blkbits; >> + unsigned int nr_blks = last_blk - first_blk + 1; >> + unsigned long flags; >> + >> + if (iop) { >> + spin_lock_irqsave(&iop->uptodate_lock, flags); >> + bitmap_set(iop->uptodate, first_blk, nr_blks); >> + if (bitmap_full(iop->uptodate, >> + i_blocks_per_folio(inode, folio))) >> + folio_mark_uptodate(folio); >> + spin_unlock_irqrestore(&iop->uptodate_lock, flags); >> + } else { >> + folio_mark_uptodate(folio); >> + } >> +} >> + >> static struct iomap_page *iop_alloc(struct inode *inode, struct folio *folio, >> unsigned int flags) >> { >> @@ -145,30 +166,6 @@ static void iomap_adjust_read_range(struct inode *inode, struct folio *folio, >> *lenp = plen; >> } >> >> -static void iomap_iop_set_range_uptodate(struct folio *folio, >> - struct iomap_page *iop, size_t off, size_t len) >> -{ >> - struct inode *inode = folio->mapping->host; >> - unsigned first = off >> inode->i_blkbits; >> - unsigned last = (off + len - 1) >> inode->i_blkbits; >> - unsigned long flags; >> - >> - spin_lock_irqsave(&iop->uptodate_lock, flags); >> - bitmap_set(iop->uptodate, first, last - first + 1); >> - if (bitmap_full(iop->uptodate, i_blocks_per_folio(inode, folio))) >> - folio_mark_uptodate(folio); >> - spin_unlock_irqrestore(&iop->uptodate_lock, flags); >> -} >> - >> -static void iomap_set_range_uptodate(struct folio *folio, >> - struct iomap_page *iop, size_t off, size_t len) >> -{ >> - if (iop) >> - iomap_iop_set_range_uptodate(folio, iop, off, len); >> - else >> - folio_mark_uptodate(folio); >> -} >> - >> static void iomap_finish_folio_read(struct folio *folio, size_t offset, >> size_t len, int error) >> { >> @@ -178,7 +175,8 @@ static void iomap_finish_folio_read(struct folio *folio, size_t offset, >> folio_clear_uptodate(folio); >> folio_set_error(folio); >> } else { >> - iomap_set_range_uptodate(folio, iop, offset, len); >> + iop_set_range_uptodate(folio->mapping->host, folio, offset, >> + len); >> } >> >> if (!iop || atomic_sub_and_test(len, &iop->read_bytes_pending)) >> @@ -240,7 +238,7 @@ static int iomap_read_inline_data(const struct iomap_iter *iter, >> memcpy(addr, iomap->inline_data, size); >> memset(addr + size, 0, PAGE_SIZE - poff - size); >> kunmap_local(addr); >> - iomap_set_range_uptodate(folio, iop, offset, PAGE_SIZE - poff); >> + iop_set_range_uptodate(iter->inode, folio, offset, PAGE_SIZE - poff); >> return 0; >> } >> >> @@ -277,7 +275,7 @@ static loff_t iomap_readpage_iter(const struct iomap_iter *iter, >> >> if (iomap_block_needs_zeroing(iter, pos)) { >> folio_zero_range(folio, poff, plen); >> - iomap_set_range_uptodate(folio, iop, poff, plen); >> + iop_set_range_uptodate(iter->inode, folio, poff, plen); >> goto done; >> } >> >> @@ -598,7 +596,7 @@ static int __iomap_write_begin(const struct iomap_iter *iter, loff_t pos, >> if (status) >> return status; >> } >> - iomap_set_range_uptodate(folio, iop, poff, plen); >> + iop_set_range_uptodate(iter->inode, folio, poff, plen); >> } while ((block_start += plen) < block_end); >> >> return 0; >> @@ -705,7 +703,6 @@ static int iomap_write_begin(struct iomap_iter *iter, loff_t pos, >> static size_t __iomap_write_end(struct inode *inode, loff_t pos, size_t len, >> size_t copied, struct folio *folio) >> { >> - struct iomap_page *iop = to_iomap_page(folio); >> flush_dcache_folio(folio); >> >> /* >> @@ -721,7 +718,7 @@ static size_t __iomap_write_end(struct inode *inode, loff_t pos, size_t len, >> */ >> if (unlikely(copied < len && !folio_test_uptodate(folio))) >> return 0; >> - iomap_set_range_uptodate(folio, iop, offset_in_folio(folio, pos), len); >> + iop_set_range_uptodate(inode, folio, offset_in_folio(folio, pos), len); >> filemap_dirty_folio(inode->i_mapping, folio); >> return copied; >> } >> -- >> 2.39.2 >>