fs: Convert iomap_readpage to iomap_read_folio
A straightforward conversion as iomap_readpage already worked in folios. Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
This commit is contained in:
Родитель
6c62371b7f
Коммит
7479c505b4
|
@ -337,9 +337,9 @@ int erofs_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo,
|
||||||
* since we dont have write or truncate flows, so no inode
|
* since we dont have write or truncate flows, so no inode
|
||||||
* locking needs to be held at the moment.
|
* locking needs to be held at the moment.
|
||||||
*/
|
*/
|
||||||
static int erofs_readpage(struct file *file, struct page *page)
|
static int erofs_read_folio(struct file *file, struct folio *folio)
|
||||||
{
|
{
|
||||||
return iomap_readpage(page, &erofs_iomap_ops);
|
return iomap_read_folio(folio, &erofs_iomap_ops);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void erofs_readahead(struct readahead_control *rac)
|
static void erofs_readahead(struct readahead_control *rac)
|
||||||
|
@ -394,7 +394,7 @@ static ssize_t erofs_file_read_iter(struct kiocb *iocb, struct iov_iter *to)
|
||||||
|
|
||||||
/* for uncompressed (aligned) files and raw access for other files */
|
/* for uncompressed (aligned) files and raw access for other files */
|
||||||
const struct address_space_operations erofs_raw_access_aops = {
|
const struct address_space_operations erofs_raw_access_aops = {
|
||||||
.readpage = erofs_readpage,
|
.read_folio = erofs_read_folio,
|
||||||
.readahead = erofs_readahead,
|
.readahead = erofs_readahead,
|
||||||
.bmap = erofs_bmap,
|
.bmap = erofs_bmap,
|
||||||
.direct_IO = noop_direct_IO,
|
.direct_IO = noop_direct_IO,
|
||||||
|
|
|
@ -467,6 +467,7 @@ static int stuffed_readpage(struct gfs2_inode *ip, struct page *page)
|
||||||
|
|
||||||
static int __gfs2_readpage(void *file, struct page *page)
|
static int __gfs2_readpage(void *file, struct page *page)
|
||||||
{
|
{
|
||||||
|
struct folio *folio = page_folio(page);
|
||||||
struct inode *inode = page->mapping->host;
|
struct inode *inode = page->mapping->host;
|
||||||
struct gfs2_inode *ip = GFS2_I(inode);
|
struct gfs2_inode *ip = GFS2_I(inode);
|
||||||
struct gfs2_sbd *sdp = GFS2_SB(inode);
|
struct gfs2_sbd *sdp = GFS2_SB(inode);
|
||||||
|
@ -474,7 +475,7 @@ static int __gfs2_readpage(void *file, struct page *page)
|
||||||
|
|
||||||
if (!gfs2_is_jdata(ip) ||
|
if (!gfs2_is_jdata(ip) ||
|
||||||
(i_blocksize(inode) == PAGE_SIZE && !page_has_buffers(page))) {
|
(i_blocksize(inode) == PAGE_SIZE && !page_has_buffers(page))) {
|
||||||
error = iomap_readpage(page, &gfs2_iomap_ops);
|
error = iomap_read_folio(folio, &gfs2_iomap_ops);
|
||||||
} else if (gfs2_is_stuffed(ip)) {
|
} else if (gfs2_is_stuffed(ip)) {
|
||||||
error = stuffed_readpage(ip, page);
|
error = stuffed_readpage(ip, page);
|
||||||
unlock_page(page);
|
unlock_page(page);
|
||||||
|
|
|
@ -320,10 +320,8 @@ done:
|
||||||
return pos - orig_pos + plen;
|
return pos - orig_pos + plen;
|
||||||
}
|
}
|
||||||
|
|
||||||
int
|
int iomap_read_folio(struct folio *folio, const struct iomap_ops *ops)
|
||||||
iomap_readpage(struct page *page, const struct iomap_ops *ops)
|
|
||||||
{
|
{
|
||||||
struct folio *folio = page_folio(page);
|
|
||||||
struct iomap_iter iter = {
|
struct iomap_iter iter = {
|
||||||
.inode = folio->mapping->host,
|
.inode = folio->mapping->host,
|
||||||
.pos = folio_pos(folio),
|
.pos = folio_pos(folio),
|
||||||
|
@ -352,12 +350,12 @@ iomap_readpage(struct page *page, const struct iomap_ops *ops)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Just like mpage_readahead and block_read_full_page, we always
|
* Just like mpage_readahead and block_read_full_page, we always
|
||||||
* return 0 and just mark the page as PageError on errors. This
|
* return 0 and just set the folio error flag on errors. This
|
||||||
* should be cleaned up throughout the stack eventually.
|
* should be cleaned up throughout the stack eventually.
|
||||||
*/
|
*/
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(iomap_readpage);
|
EXPORT_SYMBOL_GPL(iomap_read_folio);
|
||||||
|
|
||||||
static loff_t iomap_readahead_iter(const struct iomap_iter *iter,
|
static loff_t iomap_readahead_iter(const struct iomap_iter *iter,
|
||||||
struct iomap_readpage_ctx *ctx)
|
struct iomap_readpage_ctx *ctx)
|
||||||
|
@ -663,10 +661,10 @@ static size_t __iomap_write_end(struct inode *inode, loff_t pos, size_t len,
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* The blocks that were entirely written will now be uptodate, so we
|
* The blocks that were entirely written will now be uptodate, so we
|
||||||
* don't have to worry about a readpage reading them and overwriting a
|
* don't have to worry about a read_folio reading them and overwriting a
|
||||||
* partial write. However, if we've encountered a short write and only
|
* partial write. However, if we've encountered a short write and only
|
||||||
* partially written into a block, it will not be marked uptodate, so a
|
* partially written into a block, it will not be marked uptodate, so a
|
||||||
* readpage might come in and destroy our partial write.
|
* read_folio might come in and destroy our partial write.
|
||||||
*
|
*
|
||||||
* Do the simplest thing and just treat any short write to a
|
* Do the simplest thing and just treat any short write to a
|
||||||
* non-uptodate page as a zero-length write, and force the caller to
|
* non-uptodate page as a zero-length write, and force the caller to
|
||||||
|
|
|
@ -538,11 +538,11 @@ xfs_vm_bmap(
|
||||||
}
|
}
|
||||||
|
|
||||||
STATIC int
|
STATIC int
|
||||||
xfs_vm_readpage(
|
xfs_vm_read_folio(
|
||||||
struct file *unused,
|
struct file *unused,
|
||||||
struct page *page)
|
struct folio *folio)
|
||||||
{
|
{
|
||||||
return iomap_readpage(page, &xfs_read_iomap_ops);
|
return iomap_read_folio(folio, &xfs_read_iomap_ops);
|
||||||
}
|
}
|
||||||
|
|
||||||
STATIC void
|
STATIC void
|
||||||
|
@ -564,7 +564,7 @@ xfs_iomap_swapfile_activate(
|
||||||
}
|
}
|
||||||
|
|
||||||
const struct address_space_operations xfs_address_space_operations = {
|
const struct address_space_operations xfs_address_space_operations = {
|
||||||
.readpage = xfs_vm_readpage,
|
.read_folio = xfs_vm_read_folio,
|
||||||
.readahead = xfs_vm_readahead,
|
.readahead = xfs_vm_readahead,
|
||||||
.writepages = xfs_vm_writepages,
|
.writepages = xfs_vm_writepages,
|
||||||
.dirty_folio = filemap_dirty_folio,
|
.dirty_folio = filemap_dirty_folio,
|
||||||
|
|
|
@ -124,9 +124,9 @@ static const struct iomap_ops zonefs_iomap_ops = {
|
||||||
.iomap_begin = zonefs_iomap_begin,
|
.iomap_begin = zonefs_iomap_begin,
|
||||||
};
|
};
|
||||||
|
|
||||||
static int zonefs_readpage(struct file *unused, struct page *page)
|
static int zonefs_read_folio(struct file *unused, struct folio *folio)
|
||||||
{
|
{
|
||||||
return iomap_readpage(page, &zonefs_iomap_ops);
|
return iomap_read_folio(folio, &zonefs_iomap_ops);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void zonefs_readahead(struct readahead_control *rac)
|
static void zonefs_readahead(struct readahead_control *rac)
|
||||||
|
@ -192,7 +192,7 @@ static int zonefs_swap_activate(struct swap_info_struct *sis,
|
||||||
}
|
}
|
||||||
|
|
||||||
static const struct address_space_operations zonefs_file_aops = {
|
static const struct address_space_operations zonefs_file_aops = {
|
||||||
.readpage = zonefs_readpage,
|
.read_folio = zonefs_read_folio,
|
||||||
.readahead = zonefs_readahead,
|
.readahead = zonefs_readahead,
|
||||||
.writepage = zonefs_writepage,
|
.writepage = zonefs_writepage,
|
||||||
.writepages = zonefs_writepages,
|
.writepages = zonefs_writepages,
|
||||||
|
|
|
@ -225,7 +225,7 @@ static inline const struct iomap *iomap_iter_srcmap(const struct iomap_iter *i)
|
||||||
|
|
||||||
ssize_t iomap_file_buffered_write(struct kiocb *iocb, struct iov_iter *from,
|
ssize_t iomap_file_buffered_write(struct kiocb *iocb, struct iov_iter *from,
|
||||||
const struct iomap_ops *ops);
|
const struct iomap_ops *ops);
|
||||||
int iomap_readpage(struct page *page, const struct iomap_ops *ops);
|
int iomap_read_folio(struct folio *folio, const struct iomap_ops *ops);
|
||||||
void iomap_readahead(struct readahead_control *, const struct iomap_ops *ops);
|
void iomap_readahead(struct readahead_control *, const struct iomap_ops *ops);
|
||||||
bool iomap_is_partially_uptodate(struct folio *, size_t from, size_t count);
|
bool iomap_is_partially_uptodate(struct folio *, size_t from, size_t count);
|
||||||
int iomap_releasepage(struct page *page, gfp_t gfp_mask);
|
int iomap_releasepage(struct page *page, gfp_t gfp_mask);
|
||||||
|
|
Загрузка…
Ссылка в новой задаче