mirror of https://gitee.com/openkylin/linux.git
mm: make snapshotting pages for stable writes a per-bio operation
Walking a bio's page mappings has proved problematic, so create a new bio flag to indicate that a bio's data needs to be snapshotted in order to guarantee stable pages during writeback. Next, for the one user (ext3/jbd) of snapshotting, hook all the places where writes can be initiated without PG_writeback set, and set BIO_SNAP_STABLE there. We must also flag journal "metadata" bios for stable writeout, since file data can be written through the journal. Finally, the MS_SNAP_STABLE mount flag (only used by ext3) is now superfluous, so get rid of it. [akpm@linux-foundation.org: rename _submit_bh()'s `flags' to `bio_flags', delobotomize the _submit_bh declaration] [akpm@linux-foundation.org: teeny cleanup] Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> Cc: Andy Lutomirski <luto@amacapital.net> Cc: Adrian Hunter <adrian.hunter@intel.com> Cc: Artem Bityutskiy <dedekind1@gmail.com> Reviewed-by: Jan Kara <jack@suse.cz> Cc: Jens Axboe <axboe@kernel.dk> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
106c992a5e
commit
7136851117
|
@ -2949,7 +2949,7 @@ static void guard_bh_eod(int rw, struct bio *bio, struct buffer_head *bh)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
int submit_bh(int rw, struct buffer_head * bh)
|
int _submit_bh(int rw, struct buffer_head *bh, unsigned long bio_flags)
|
||||||
{
|
{
|
||||||
struct bio *bio;
|
struct bio *bio;
|
||||||
int ret = 0;
|
int ret = 0;
|
||||||
|
@ -2984,6 +2984,7 @@ int submit_bh(int rw, struct buffer_head * bh)
|
||||||
|
|
||||||
bio->bi_end_io = end_bio_bh_io_sync;
|
bio->bi_end_io = end_bio_bh_io_sync;
|
||||||
bio->bi_private = bh;
|
bio->bi_private = bh;
|
||||||
|
bio->bi_flags |= bio_flags;
|
||||||
|
|
||||||
/* Take care of bh's that straddle the end of the device */
|
/* Take care of bh's that straddle the end of the device */
|
||||||
guard_bh_eod(rw, bio, bh);
|
guard_bh_eod(rw, bio, bh);
|
||||||
|
@ -2997,6 +2998,12 @@ int submit_bh(int rw, struct buffer_head * bh)
|
||||||
bio_put(bio);
|
bio_put(bio);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
EXPORT_SYMBOL_GPL(_submit_bh);
|
||||||
|
|
||||||
|
int submit_bh(int rw, struct buffer_head *bh)
|
||||||
|
{
|
||||||
|
return _submit_bh(rw, bh, 0);
|
||||||
|
}
|
||||||
EXPORT_SYMBOL(submit_bh);
|
EXPORT_SYMBOL(submit_bh);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -2067,7 +2067,6 @@ static int ext3_fill_super (struct super_block *sb, void *data, int silent)
|
||||||
test_opt(sb,DATA_FLAGS) == EXT3_MOUNT_JOURNAL_DATA ? "journal":
|
test_opt(sb,DATA_FLAGS) == EXT3_MOUNT_JOURNAL_DATA ? "journal":
|
||||||
test_opt(sb,DATA_FLAGS) == EXT3_MOUNT_ORDERED_DATA ? "ordered":
|
test_opt(sb,DATA_FLAGS) == EXT3_MOUNT_ORDERED_DATA ? "ordered":
|
||||||
"writeback");
|
"writeback");
|
||||||
sb->s_flags |= MS_SNAP_STABLE;
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
|
|
|
@ -162,8 +162,17 @@ static void journal_do_submit_data(struct buffer_head **wbuf, int bufs,
|
||||||
|
|
||||||
for (i = 0; i < bufs; i++) {
|
for (i = 0; i < bufs; i++) {
|
||||||
wbuf[i]->b_end_io = end_buffer_write_sync;
|
wbuf[i]->b_end_io = end_buffer_write_sync;
|
||||||
/* We use-up our safety reference in submit_bh() */
|
/*
|
||||||
submit_bh(write_op, wbuf[i]);
|
* Here we write back pagecache data that may be mmaped. Since
|
||||||
|
* we cannot afford to clean the page and set PageWriteback
|
||||||
|
* here due to lock ordering (page lock ranks above transaction
|
||||||
|
* start), the data can change while IO is in flight. Tell the
|
||||||
|
* block layer it should bounce the bio pages if stable data
|
||||||
|
* during write is required.
|
||||||
|
*
|
||||||
|
* We use up our safety reference in submit_bh().
|
||||||
|
*/
|
||||||
|
_submit_bh(write_op, wbuf[i], 1 << BIO_SNAP_STABLE);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -667,7 +676,17 @@ void journal_commit_transaction(journal_t *journal)
|
||||||
clear_buffer_dirty(bh);
|
clear_buffer_dirty(bh);
|
||||||
set_buffer_uptodate(bh);
|
set_buffer_uptodate(bh);
|
||||||
bh->b_end_io = journal_end_buffer_io_sync;
|
bh->b_end_io = journal_end_buffer_io_sync;
|
||||||
submit_bh(write_op, bh);
|
/*
|
||||||
|
* In data=journal mode, here we can end up
|
||||||
|
* writing pagecache data that might be
|
||||||
|
* mmapped. Since we can't afford to clean the
|
||||||
|
* page and set PageWriteback (see the comment
|
||||||
|
* near the other use of _submit_bh()), the
|
||||||
|
* data can change while the write is in
|
||||||
|
* flight. Tell the block layer to bounce the
|
||||||
|
* bio pages if stable pages are required.
|
||||||
|
*/
|
||||||
|
_submit_bh(write_op, bh, 1 << BIO_SNAP_STABLE);
|
||||||
}
|
}
|
||||||
cond_resched();
|
cond_resched();
|
||||||
|
|
||||||
|
|
|
@ -111,12 +111,13 @@ struct bio {
|
||||||
#define BIO_FS_INTEGRITY 9 /* fs owns integrity data, not block layer */
|
#define BIO_FS_INTEGRITY 9 /* fs owns integrity data, not block layer */
|
||||||
#define BIO_QUIET 10 /* Make BIO Quiet */
|
#define BIO_QUIET 10 /* Make BIO Quiet */
|
||||||
#define BIO_MAPPED_INTEGRITY 11/* integrity metadata has been remapped */
|
#define BIO_MAPPED_INTEGRITY 11/* integrity metadata has been remapped */
|
||||||
|
#define BIO_SNAP_STABLE 12 /* bio data must be snapshotted during write */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Flags starting here get preserved by bio_reset() - this includes
|
* Flags starting here get preserved by bio_reset() - this includes
|
||||||
* BIO_POOL_IDX()
|
* BIO_POOL_IDX()
|
||||||
*/
|
*/
|
||||||
#define BIO_RESET_BITS 12
|
#define BIO_RESET_BITS 13
|
||||||
|
|
||||||
#define bio_flagged(bio, flag) ((bio)->bi_flags & (1 << (flag)))
|
#define bio_flagged(bio, flag) ((bio)->bi_flags & (1 << (flag)))
|
||||||
|
|
||||||
|
|
|
@ -181,6 +181,7 @@ void ll_rw_block(int, int, struct buffer_head * bh[]);
|
||||||
int sync_dirty_buffer(struct buffer_head *bh);
|
int sync_dirty_buffer(struct buffer_head *bh);
|
||||||
int __sync_dirty_buffer(struct buffer_head *bh, int rw);
|
int __sync_dirty_buffer(struct buffer_head *bh, int rw);
|
||||||
void write_dirty_buffer(struct buffer_head *bh, int rw);
|
void write_dirty_buffer(struct buffer_head *bh, int rw);
|
||||||
|
int _submit_bh(int rw, struct buffer_head *bh, unsigned long bio_flags);
|
||||||
int submit_bh(int, struct buffer_head *);
|
int submit_bh(int, struct buffer_head *);
|
||||||
void write_boundary_block(struct block_device *bdev,
|
void write_boundary_block(struct block_device *bdev,
|
||||||
sector_t bblock, unsigned blocksize);
|
sector_t bblock, unsigned blocksize);
|
||||||
|
|
|
@ -88,7 +88,6 @@ struct inodes_stat_t {
|
||||||
#define MS_STRICTATIME (1<<24) /* Always perform atime updates */
|
#define MS_STRICTATIME (1<<24) /* Always perform atime updates */
|
||||||
|
|
||||||
/* These sb flags are internal to the kernel */
|
/* These sb flags are internal to the kernel */
|
||||||
#define MS_SNAP_STABLE (1<<27) /* Snapshot pages during writeback, if needed */
|
|
||||||
#define MS_NOSEC (1<<28)
|
#define MS_NOSEC (1<<28)
|
||||||
#define MS_BORN (1<<29)
|
#define MS_BORN (1<<29)
|
||||||
#define MS_ACTIVE (1<<30)
|
#define MS_ACTIVE (1<<30)
|
||||||
|
|
21
mm/bounce.c
21
mm/bounce.c
|
@ -181,32 +181,13 @@ static void bounce_end_io_read_isa(struct bio *bio, int err)
|
||||||
#ifdef CONFIG_NEED_BOUNCE_POOL
|
#ifdef CONFIG_NEED_BOUNCE_POOL
|
||||||
static int must_snapshot_stable_pages(struct request_queue *q, struct bio *bio)
|
static int must_snapshot_stable_pages(struct request_queue *q, struct bio *bio)
|
||||||
{
|
{
|
||||||
struct page *page;
|
|
||||||
struct backing_dev_info *bdi;
|
|
||||||
struct address_space *mapping;
|
|
||||||
struct bio_vec *from;
|
|
||||||
int i;
|
|
||||||
|
|
||||||
if (bio_data_dir(bio) != WRITE)
|
if (bio_data_dir(bio) != WRITE)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
if (!bdi_cap_stable_pages_required(&q->backing_dev_info))
|
if (!bdi_cap_stable_pages_required(&q->backing_dev_info))
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
/*
|
return test_bit(BIO_SNAP_STABLE, &bio->bi_flags);
|
||||||
* Based on the first page that has a valid mapping, decide whether or
|
|
||||||
* not we have to employ bounce buffering to guarantee stable pages.
|
|
||||||
*/
|
|
||||||
bio_for_each_segment(from, bio, i) {
|
|
||||||
page = from->bv_page;
|
|
||||||
mapping = page_mapping(page);
|
|
||||||
if (!mapping)
|
|
||||||
continue;
|
|
||||||
bdi = mapping->backing_dev_info;
|
|
||||||
return mapping->host->i_sb->s_flags & MS_SNAP_STABLE;
|
|
||||||
}
|
|
||||||
|
|
||||||
return 0;
|
|
||||||
}
|
}
|
||||||
#else
|
#else
|
||||||
static int must_snapshot_stable_pages(struct request_queue *q, struct bio *bio)
|
static int must_snapshot_stable_pages(struct request_queue *q, struct bio *bio)
|
||||||
|
|
|
@ -2311,10 +2311,6 @@ void wait_for_stable_page(struct page *page)
|
||||||
|
|
||||||
if (!bdi_cap_stable_pages_required(bdi))
|
if (!bdi_cap_stable_pages_required(bdi))
|
||||||
return;
|
return;
|
||||||
#ifdef CONFIG_NEED_BOUNCE_POOL
|
|
||||||
if (mapping->host->i_sb->s_flags & MS_SNAP_STABLE)
|
|
||||||
return;
|
|
||||||
#endif /* CONFIG_NEED_BOUNCE_POOL */
|
|
||||||
|
|
||||||
wait_on_page_writeback(page);
|
wait_on_page_writeback(page);
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue