fs/block_dev.c: add bdev_read_page() and bdev_write_page()
A block device driver may choose to provide a rw_page operation. These will be called when the filesystem is attempting to do page sized I/O to page cache pages (ie not for direct I/O). This does preclude I/Os that are larger than page size, so this may only be a performance gain for some devices. Signed-off-by: Matthew Wilcox <matthew.r.wilcox@intel.com> Tested-by: Dheeraj Reddy <dheeraj.reddy@intel.com> Cc: Dave Chinner <david@fromorbit.com> Cc: Hugh Dickins <hughd@google.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
57d998456a
commit
47a191fd38
|
@ -363,6 +363,69 @@ int blkdev_fsync(struct file *filp, loff_t start, loff_t end, int datasync)
|
|||
}
|
||||
EXPORT_SYMBOL(blkdev_fsync);
|
||||
|
||||
/**
|
||||
* bdev_read_page() - Start reading a page from a block device
|
||||
* @bdev: The device to read the page from
|
||||
* @sector: The offset on the device to read the page to (need not be aligned)
|
||||
* @page: The page to read
|
||||
*
|
||||
* On entry, the page should be locked. It will be unlocked when the page
|
||||
* has been read. If the block driver implements rw_page synchronously,
|
||||
* that will be true on exit from this function, but it need not be.
|
||||
*
|
||||
* Errors returned by this function are usually "soft", eg out of memory, or
|
||||
* queue full; callers should try a different route to read this page rather
|
||||
* than propagate an error back up the stack.
|
||||
*
|
||||
* Return: negative errno if an error occurs, 0 if submission was successful.
|
||||
*/
|
||||
int bdev_read_page(struct block_device *bdev, sector_t sector,
|
||||
struct page *page)
|
||||
{
|
||||
const struct block_device_operations *ops = bdev->bd_disk->fops;
|
||||
if (!ops->rw_page)
|
||||
return -EOPNOTSUPP;
|
||||
return ops->rw_page(bdev, sector + get_start_sect(bdev), page, READ);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(bdev_read_page);
|
||||
|
||||
/**
|
||||
* bdev_write_page() - Start writing a page to a block device
|
||||
* @bdev: The device to write the page to
|
||||
* @sector: The offset on the device to write the page to (need not be aligned)
|
||||
* @page: The page to write
|
||||
* @wbc: The writeback_control for the write
|
||||
*
|
||||
* On entry, the page should be locked and not currently under writeback.
|
||||
* On exit, if the write started successfully, the page will be unlocked and
|
||||
* under writeback. If the write failed already (eg the driver failed to
|
||||
* queue the page to the device), the page will still be locked. If the
|
||||
* caller is a ->writepage implementation, it will need to unlock the page.
|
||||
*
|
||||
* Errors returned by this function are usually "soft", eg out of memory, or
|
||||
* queue full; callers should try a different route to write this page rather
|
||||
* than propagate an error back up the stack.
|
||||
*
|
||||
* Return: negative errno if an error occurs, 0 if submission was successful.
|
||||
*/
|
||||
int bdev_write_page(struct block_device *bdev, sector_t sector,
|
||||
struct page *page, struct writeback_control *wbc)
|
||||
{
|
||||
int result;
|
||||
int rw = (wbc->sync_mode == WB_SYNC_ALL) ? WRITE_SYNC : WRITE;
|
||||
const struct block_device_operations *ops = bdev->bd_disk->fops;
|
||||
if (!ops->rw_page)
|
||||
return -EOPNOTSUPP;
|
||||
set_page_writeback(page);
|
||||
result = ops->rw_page(bdev, sector + get_start_sect(bdev), page, rw);
|
||||
if (result)
|
||||
end_page_writeback(page);
|
||||
else
|
||||
unlock_page(page);
|
||||
return result;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(bdev_write_page);
|
||||
|
||||
/*
|
||||
* pseudo-fs
|
||||
*/
|
||||
|
|
12
fs/mpage.c
12
fs/mpage.c
|
@ -269,6 +269,11 @@ do_mpage_readpage(struct bio *bio, struct page *page, unsigned nr_pages,
|
|||
|
||||
alloc_new:
|
||||
if (bio == NULL) {
|
||||
if (first_hole == blocks_per_page) {
|
||||
if (!bdev_read_page(bdev, blocks[0] << (blkbits - 9),
|
||||
page))
|
||||
goto out;
|
||||
}
|
||||
bio = mpage_alloc(bdev, blocks[0] << (blkbits - 9),
|
||||
min_t(int, nr_pages, bio_get_nr_vecs(bdev)),
|
||||
GFP_KERNEL);
|
||||
|
@ -587,6 +592,13 @@ static int __mpage_writepage(struct page *page, struct writeback_control *wbc,
|
|||
|
||||
alloc_new:
|
||||
if (bio == NULL) {
|
||||
if (first_unmapped == blocks_per_page) {
|
||||
if (!bdev_write_page(bdev, blocks[0] << (blkbits - 9),
|
||||
page, wbc)) {
|
||||
clean_buffers(page, first_unmapped);
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
bio = mpage_alloc(bdev, blocks[0] << (blkbits - 9),
|
||||
bio_get_nr_vecs(bdev), GFP_NOFS|__GFP_HIGH);
|
||||
if (bio == NULL)
|
||||
|
|
|
@ -1588,6 +1588,7 @@ static inline bool blk_integrity_is_initialized(struct gendisk *g)
|
|||
struct block_device_operations {
|
||||
int (*open) (struct block_device *, fmode_t);
|
||||
void (*release) (struct gendisk *, fmode_t);
|
||||
int (*rw_page)(struct block_device *, sector_t, struct page *, int rw);
|
||||
int (*ioctl) (struct block_device *, fmode_t, unsigned, unsigned long);
|
||||
int (*compat_ioctl) (struct block_device *, fmode_t, unsigned, unsigned long);
|
||||
int (*direct_access) (struct block_device *, sector_t,
|
||||
|
@ -1606,6 +1607,9 @@ struct block_device_operations {
|
|||
|
||||
extern int __blkdev_driver_ioctl(struct block_device *, fmode_t, unsigned int,
|
||||
unsigned long);
|
||||
extern int bdev_read_page(struct block_device *, sector_t, struct page *);
|
||||
extern int bdev_write_page(struct block_device *, sector_t, struct page *,
|
||||
struct writeback_control *);
|
||||
#else /* CONFIG_BLOCK */
|
||||
|
||||
struct block_device;
|
||||
|
|
Loading…
Reference in New Issue
Block a user