erofs: clean up .read_folio() and .readahead() in fscache mode

The implementation of these two functions in fscache mode is almost the
same. Extract the same part as a generic helper to remove the code
duplication.

Signed-off-by: Jingbo Xu <jefflexu@linux.alibaba.com>
Reviewed-by: Jia Zhu <zhujia.zj@bytedance.com>
Link: https://lore.kernel.org/r/20220922062414.20437-1-jefflexu@linux.alibaba.com
Signed-off-by: Gao Xiang <hsiangkao@linux.alibaba.com>
This commit is contained in:
Jingbo Xu 2022-09-22 14:24:14 +08:00 committed by Gao Xiang
parent 2ef1644141
commit 1ae9470c3e

View file

@ -240,113 +240,111 @@ static int erofs_fscache_meta_read_folio(struct file *data, struct folio *folio)
return ret; return ret;
} }
static int erofs_fscache_read_folio_inline(struct folio *folio, /*
struct erofs_map_blocks *map) * Read into page cache in the range described by (@pos, @len).
*
* On return, the caller is responsible for page unlocking if the output @unlock
* is true, or the callee will take this responsibility through netfs_io_request
* interface.
*
* The return value is the number of bytes successfully handled, or negative
* error code on failure. The only exception is that, the length of the range
* instead of the error code is returned on failure after netfs_io_request is
* allocated, so that .readahead() could advance rac accordingly.
*/
static int erofs_fscache_data_read(struct address_space *mapping,
loff_t pos, size_t len, bool *unlock)
{ {
struct super_block *sb = folio_mapping(folio)->host->i_sb; struct inode *inode = mapping->host;
struct erofs_buf buf = __EROFS_BUF_INITIALIZER;
erofs_blk_t blknr;
size_t offset, len;
void *src, *dst;
/* For tail packing layout, the offset may be non-zero. */
offset = erofs_blkoff(map->m_pa);
blknr = erofs_blknr(map->m_pa);
len = map->m_llen;
src = erofs_read_metabuf(&buf, sb, blknr, EROFS_KMAP);
if (IS_ERR(src))
return PTR_ERR(src);
dst = kmap_local_folio(folio, 0);
memcpy(dst, src + offset, len);
memset(dst + len, 0, PAGE_SIZE - len);
kunmap_local(dst);
erofs_put_metabuf(&buf);
return 0;
}
static int erofs_fscache_read_folio(struct file *file, struct folio *folio)
{
struct inode *inode = folio_mapping(folio)->host;
struct super_block *sb = inode->i_sb; struct super_block *sb = inode->i_sb;
struct netfs_io_request *rreq;
struct erofs_map_blocks map; struct erofs_map_blocks map;
struct erofs_map_dev mdev; struct erofs_map_dev mdev;
struct netfs_io_request *rreq; struct iov_iter iter;
erofs_off_t pos; size_t count;
loff_t pstart;
int ret; int ret;
DBG_BUGON(folio_size(folio) != EROFS_BLKSIZ); *unlock = true;
pos = folio_pos(folio);
map.m_la = pos; map.m_la = pos;
ret = erofs_map_blocks(inode, &map, EROFS_GET_BLOCKS_RAW); ret = erofs_map_blocks(inode, &map, EROFS_GET_BLOCKS_RAW);
if (ret) if (ret)
goto out_unlock; return ret;
if (!(map.m_flags & EROFS_MAP_MAPPED)) {
folio_zero_range(folio, 0, folio_size(folio));
goto out_uptodate;
}
if (map.m_flags & EROFS_MAP_META) { if (map.m_flags & EROFS_MAP_META) {
ret = erofs_fscache_read_folio_inline(folio, &map); struct erofs_buf buf = __EROFS_BUF_INITIALIZER;
goto out_uptodate; erofs_blk_t blknr;
size_t offset, size;
void *src;
/* For tail packing layout, the offset may be non-zero. */
offset = erofs_blkoff(map.m_pa);
blknr = erofs_blknr(map.m_pa);
size = map.m_llen;
src = erofs_read_metabuf(&buf, sb, blknr, EROFS_KMAP);
if (IS_ERR(src))
return PTR_ERR(src);
iov_iter_xarray(&iter, READ, &mapping->i_pages, pos, PAGE_SIZE);
if (copy_to_iter(src + offset, size, &iter) != size)
return -EFAULT;
iov_iter_zero(PAGE_SIZE - size, &iter);
erofs_put_metabuf(&buf);
return PAGE_SIZE;
}
count = min_t(size_t, map.m_llen - (pos - map.m_la), len);
DBG_BUGON(!count || count % PAGE_SIZE);
if (!(map.m_flags & EROFS_MAP_MAPPED)) {
iov_iter_xarray(&iter, READ, &mapping->i_pages, pos, count);
iov_iter_zero(count, &iter);
return count;
} }
mdev = (struct erofs_map_dev) { mdev = (struct erofs_map_dev) {
.m_deviceid = map.m_deviceid, .m_deviceid = map.m_deviceid,
.m_pa = map.m_pa, .m_pa = map.m_pa,
}; };
ret = erofs_map_dev(sb, &mdev); ret = erofs_map_dev(sb, &mdev);
if (ret) if (ret)
goto out_unlock; return ret;
rreq = erofs_fscache_alloc_request(mapping, pos, count);
if (IS_ERR(rreq))
return PTR_ERR(rreq);
rreq = erofs_fscache_alloc_request(folio_mapping(folio), *unlock = false;
folio_pos(folio), folio_size(folio)); erofs_fscache_read_folios_async(mdev.m_fscache->cookie,
if (IS_ERR(rreq)) { rreq, mdev.m_pa + (pos - map.m_la));
ret = PTR_ERR(rreq); return count;
goto out_unlock;
}
pstart = mdev.m_pa + (pos - map.m_la);
return erofs_fscache_read_folios_async(mdev.m_fscache->cookie,
rreq, pstart);
out_uptodate:
if (!ret)
folio_mark_uptodate(folio);
out_unlock:
folio_unlock(folio);
return ret;
} }
static void erofs_fscache_advance_folios(struct readahead_control *rac, static int erofs_fscache_read_folio(struct file *file, struct folio *folio)
size_t len, bool unlock)
{ {
while (len) { bool unlock;
struct folio *folio = readahead_folio(rac); int ret;
len -= folio_size(folio);
if (unlock) { DBG_BUGON(folio_size(folio) != EROFS_BLKSIZ);
ret = erofs_fscache_data_read(folio_mapping(folio), folio_pos(folio),
folio_size(folio), &unlock);
if (unlock) {
if (ret > 0)
folio_mark_uptodate(folio); folio_mark_uptodate(folio);
folio_unlock(folio); folio_unlock(folio);
}
} }
return ret < 0 ? ret : 0;
} }
static void erofs_fscache_readahead(struct readahead_control *rac) static void erofs_fscache_readahead(struct readahead_control *rac)
{ {
struct inode *inode = rac->mapping->host; struct folio *folio;
struct super_block *sb = inode->i_sb; size_t len, done = 0;
size_t len, count, done = 0; loff_t start, pos;
erofs_off_t pos; bool unlock;
loff_t start, offset; int ret, size;
int ret;
if (!readahead_count(rac)) if (!readahead_count(rac))
return; return;
@ -355,67 +353,22 @@ static void erofs_fscache_readahead(struct readahead_control *rac)
len = readahead_length(rac); len = readahead_length(rac);
do { do {
struct erofs_map_blocks map;
struct erofs_map_dev mdev;
struct netfs_io_request *rreq;
pos = start + done; pos = start + done;
map.m_la = pos; ret = erofs_fscache_data_read(rac->mapping, pos,
len - done, &unlock);
ret = erofs_map_blocks(inode, &map, EROFS_GET_BLOCKS_RAW); if (ret <= 0)
if (ret)
return; return;
offset = start + done; size = ret;
count = min_t(size_t, map.m_llen - (pos - map.m_la), while (size) {
len - done); folio = readahead_folio(rac);
size -= folio_size(folio);
if (!(map.m_flags & EROFS_MAP_MAPPED)) { if (unlock) {
struct iov_iter iter;
iov_iter_xarray(&iter, READ, &rac->mapping->i_pages,
offset, count);
iov_iter_zero(count, &iter);
erofs_fscache_advance_folios(rac, count, true);
ret = count;
continue;
}
if (map.m_flags & EROFS_MAP_META) {
struct folio *folio = readahead_folio(rac);
ret = erofs_fscache_read_folio_inline(folio, &map);
if (!ret) {
folio_mark_uptodate(folio); folio_mark_uptodate(folio);
ret = folio_size(folio); folio_unlock(folio);
} }
folio_unlock(folio);
continue;
} }
} while ((done += ret) < len);
mdev = (struct erofs_map_dev) {
.m_deviceid = map.m_deviceid,
.m_pa = map.m_pa,
};
ret = erofs_map_dev(sb, &mdev);
if (ret)
return;
rreq = erofs_fscache_alloc_request(rac->mapping, offset, count);
if (IS_ERR(rreq))
return;
/*
* Drop the ref of folios here. Unlock them in
* rreq_unlock_folios() when rreq complete.
*/
erofs_fscache_advance_folios(rac, count, false);
ret = erofs_fscache_read_folios_async(mdev.m_fscache->cookie,
rreq, mdev.m_pa + (pos - map.m_la));
if (!ret)
ret = count;
} while (ret > 0 && ((done += ret) < len));
} }
static const struct address_space_operations erofs_fscache_meta_aops = { static const struct address_space_operations erofs_fscache_meta_aops = {