mirror of
https://github.com/adulau/aha.git
synced 2024-12-28 11:46:19 +00:00
writeback: separate starting of sync vs opportunistic writeback
bdi_start_writeback() is currently split into two paths, one for WB_SYNC_NONE and one for WB_SYNC_ALL. Add bdi_sync_writeback() for WB_SYNC_ALL writeback and let bdi_start_writeback() handle only WB_SYNC_NONE. Push down the writeback_control allocation and only accept the parameters that make sense for each function. This cleans up the API considerably. Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
This commit is contained in:
parent
bcddc3f01c
commit
b6e51316da
5 changed files with 75 additions and 95 deletions
|
@ -74,14 +74,10 @@ static inline bool bdi_work_on_stack(struct bdi_work *work)
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline void bdi_work_init(struct bdi_work *work,
|
static inline void bdi_work_init(struct bdi_work *work,
|
||||||
struct writeback_control *wbc)
|
struct wb_writeback_args *args)
|
||||||
{
|
{
|
||||||
INIT_RCU_HEAD(&work->rcu_head);
|
INIT_RCU_HEAD(&work->rcu_head);
|
||||||
work->args.sb = wbc->sb;
|
work->args = *args;
|
||||||
work->args.nr_pages = wbc->nr_to_write;
|
|
||||||
work->args.sync_mode = wbc->sync_mode;
|
|
||||||
work->args.range_cyclic = wbc->range_cyclic;
|
|
||||||
work->args.for_kupdate = 0;
|
|
||||||
work->state = WS_USED;
|
work->state = WS_USED;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -194,7 +190,7 @@ static void bdi_wait_on_work_clear(struct bdi_work *work)
|
||||||
}
|
}
|
||||||
|
|
||||||
static void bdi_alloc_queue_work(struct backing_dev_info *bdi,
|
static void bdi_alloc_queue_work(struct backing_dev_info *bdi,
|
||||||
struct writeback_control *wbc)
|
struct wb_writeback_args *args)
|
||||||
{
|
{
|
||||||
struct bdi_work *work;
|
struct bdi_work *work;
|
||||||
|
|
||||||
|
@ -204,7 +200,7 @@ static void bdi_alloc_queue_work(struct backing_dev_info *bdi,
|
||||||
*/
|
*/
|
||||||
work = kmalloc(sizeof(*work), GFP_ATOMIC);
|
work = kmalloc(sizeof(*work), GFP_ATOMIC);
|
||||||
if (work) {
|
if (work) {
|
||||||
bdi_work_init(work, wbc);
|
bdi_work_init(work, args);
|
||||||
bdi_queue_work(bdi, work);
|
bdi_queue_work(bdi, work);
|
||||||
} else {
|
} else {
|
||||||
struct bdi_writeback *wb = &bdi->wb;
|
struct bdi_writeback *wb = &bdi->wb;
|
||||||
|
@ -214,24 +210,54 @@ static void bdi_alloc_queue_work(struct backing_dev_info *bdi,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
void bdi_start_writeback(struct writeback_control *wbc)
|
/**
|
||||||
|
* bdi_sync_writeback - start and wait for writeback
|
||||||
|
* @bdi: the backing device to write from
|
||||||
|
* @sb: write inodes from this super_block
|
||||||
|
*
|
||||||
|
* Description:
|
||||||
|
* This does WB_SYNC_ALL data integrity writeback and waits for the
|
||||||
|
* IO to complete. Callers must hold the sb s_umount semaphore for
|
||||||
|
* reading, to avoid having the super disappear before we are done.
|
||||||
|
*/
|
||||||
|
static void bdi_sync_writeback(struct backing_dev_info *bdi,
|
||||||
|
struct super_block *sb)
|
||||||
{
|
{
|
||||||
/*
|
struct wb_writeback_args args = {
|
||||||
* WB_SYNC_NONE is opportunistic writeback. If this allocation fails,
|
.sb = sb,
|
||||||
* bdi_queue_work() will wake up the thread and flush old data. This
|
.sync_mode = WB_SYNC_ALL,
|
||||||
* should ensure some amount of progress in freeing memory.
|
.nr_pages = LONG_MAX,
|
||||||
*/
|
.range_cyclic = 0,
|
||||||
if (wbc->sync_mode != WB_SYNC_ALL)
|
};
|
||||||
bdi_alloc_queue_work(wbc->bdi, wbc);
|
struct bdi_work work;
|
||||||
else {
|
|
||||||
struct bdi_work work;
|
|
||||||
|
|
||||||
bdi_work_init(&work, wbc);
|
bdi_work_init(&work, &args);
|
||||||
work.state |= WS_ONSTACK;
|
work.state |= WS_ONSTACK;
|
||||||
|
|
||||||
bdi_queue_work(wbc->bdi, &work);
|
bdi_queue_work(bdi, &work);
|
||||||
bdi_wait_on_work_clear(&work);
|
bdi_wait_on_work_clear(&work);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* bdi_start_writeback - start writeback
|
||||||
|
* @bdi: the backing device to write from
|
||||||
|
* @nr_pages: the number of pages to write
|
||||||
|
*
|
||||||
|
* Description:
|
||||||
|
* This does WB_SYNC_NONE opportunistic writeback. The IO is only
|
||||||
|
* started when this function returns, we make no guarentees on
|
||||||
|
* completion. Caller need not hold sb s_umount semaphore.
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
void bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages)
|
||||||
|
{
|
||||||
|
struct wb_writeback_args args = {
|
||||||
|
.sync_mode = WB_SYNC_NONE,
|
||||||
|
.nr_pages = nr_pages,
|
||||||
|
.range_cyclic = 1,
|
||||||
|
};
|
||||||
|
|
||||||
|
bdi_alloc_queue_work(bdi, &args);
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -863,23 +889,25 @@ int bdi_writeback_task(struct bdi_writeback *wb)
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Schedule writeback for all backing devices. Can only be used for
|
* Schedule writeback for all backing devices. This does WB_SYNC_NONE
|
||||||
* WB_SYNC_NONE writeback, WB_SYNC_ALL should use bdi_start_writeback()
|
* writeback, for integrity writeback see bdi_sync_writeback().
|
||||||
* and pass in the superblock.
|
|
||||||
*/
|
*/
|
||||||
static void bdi_writeback_all(struct writeback_control *wbc)
|
static void bdi_writeback_all(struct super_block *sb, long nr_pages)
|
||||||
{
|
{
|
||||||
|
struct wb_writeback_args args = {
|
||||||
|
.sb = sb,
|
||||||
|
.nr_pages = nr_pages,
|
||||||
|
.sync_mode = WB_SYNC_NONE,
|
||||||
|
};
|
||||||
struct backing_dev_info *bdi;
|
struct backing_dev_info *bdi;
|
||||||
|
|
||||||
WARN_ON(wbc->sync_mode == WB_SYNC_ALL);
|
|
||||||
|
|
||||||
rcu_read_lock();
|
rcu_read_lock();
|
||||||
|
|
||||||
list_for_each_entry_rcu(bdi, &bdi_list, bdi_list) {
|
list_for_each_entry_rcu(bdi, &bdi_list, bdi_list) {
|
||||||
if (!bdi_has_dirty_io(bdi))
|
if (!bdi_has_dirty_io(bdi))
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
bdi_alloc_queue_work(bdi, wbc);
|
bdi_alloc_queue_work(bdi, &args);
|
||||||
}
|
}
|
||||||
|
|
||||||
rcu_read_unlock();
|
rcu_read_unlock();
|
||||||
|
@ -891,17 +919,10 @@ static void bdi_writeback_all(struct writeback_control *wbc)
|
||||||
*/
|
*/
|
||||||
void wakeup_flusher_threads(long nr_pages)
|
void wakeup_flusher_threads(long nr_pages)
|
||||||
{
|
{
|
||||||
struct writeback_control wbc = {
|
|
||||||
.sync_mode = WB_SYNC_NONE,
|
|
||||||
.older_than_this = NULL,
|
|
||||||
.range_cyclic = 1,
|
|
||||||
};
|
|
||||||
|
|
||||||
if (nr_pages == 0)
|
if (nr_pages == 0)
|
||||||
nr_pages = global_page_state(NR_FILE_DIRTY) +
|
nr_pages = global_page_state(NR_FILE_DIRTY) +
|
||||||
global_page_state(NR_UNSTABLE_NFS);
|
global_page_state(NR_UNSTABLE_NFS);
|
||||||
wbc.nr_to_write = nr_pages;
|
bdi_writeback_all(NULL, nr_pages);
|
||||||
bdi_writeback_all(&wbc);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static noinline void block_dump___mark_inode_dirty(struct inode *inode)
|
static noinline void block_dump___mark_inode_dirty(struct inode *inode)
|
||||||
|
@ -1048,7 +1069,7 @@ EXPORT_SYMBOL(__mark_inode_dirty);
|
||||||
* on the writer throttling path, and we get decent balancing between many
|
* on the writer throttling path, and we get decent balancing between many
|
||||||
* throttled threads: we don't want them all piling up on inode_sync_wait.
|
* throttled threads: we don't want them all piling up on inode_sync_wait.
|
||||||
*/
|
*/
|
||||||
static void wait_sb_inodes(struct writeback_control *wbc)
|
static void wait_sb_inodes(struct super_block *sb)
|
||||||
{
|
{
|
||||||
struct inode *inode, *old_inode = NULL;
|
struct inode *inode, *old_inode = NULL;
|
||||||
|
|
||||||
|
@ -1056,7 +1077,7 @@ static void wait_sb_inodes(struct writeback_control *wbc)
|
||||||
* We need to be protected against the filesystem going from
|
* We need to be protected against the filesystem going from
|
||||||
* r/o to r/w or vice versa.
|
* r/o to r/w or vice versa.
|
||||||
*/
|
*/
|
||||||
WARN_ON(!rwsem_is_locked(&wbc->sb->s_umount));
|
WARN_ON(!rwsem_is_locked(&sb->s_umount));
|
||||||
|
|
||||||
spin_lock(&inode_lock);
|
spin_lock(&inode_lock);
|
||||||
|
|
||||||
|
@ -1067,7 +1088,7 @@ static void wait_sb_inodes(struct writeback_control *wbc)
|
||||||
* In which case, the inode may not be on the dirty list, but
|
* In which case, the inode may not be on the dirty list, but
|
||||||
* we still have to wait for that writeout.
|
* we still have to wait for that writeout.
|
||||||
*/
|
*/
|
||||||
list_for_each_entry(inode, &wbc->sb->s_inodes, i_sb_list) {
|
list_for_each_entry(inode, &sb->s_inodes, i_sb_list) {
|
||||||
struct address_space *mapping;
|
struct address_space *mapping;
|
||||||
|
|
||||||
if (inode->i_state & (I_FREEING|I_CLEAR|I_WILL_FREE|I_NEW))
|
if (inode->i_state & (I_FREEING|I_CLEAR|I_WILL_FREE|I_NEW))
|
||||||
|
@ -1107,14 +1128,8 @@ static void wait_sb_inodes(struct writeback_control *wbc)
|
||||||
* for IO completion of submitted IO. The number of pages submitted is
|
* for IO completion of submitted IO. The number of pages submitted is
|
||||||
* returned.
|
* returned.
|
||||||
*/
|
*/
|
||||||
long writeback_inodes_sb(struct super_block *sb)
|
void writeback_inodes_sb(struct super_block *sb)
|
||||||
{
|
{
|
||||||
struct writeback_control wbc = {
|
|
||||||
.sb = sb,
|
|
||||||
.sync_mode = WB_SYNC_NONE,
|
|
||||||
.range_start = 0,
|
|
||||||
.range_end = LLONG_MAX,
|
|
||||||
};
|
|
||||||
unsigned long nr_dirty = global_page_state(NR_FILE_DIRTY);
|
unsigned long nr_dirty = global_page_state(NR_FILE_DIRTY);
|
||||||
unsigned long nr_unstable = global_page_state(NR_UNSTABLE_NFS);
|
unsigned long nr_unstable = global_page_state(NR_UNSTABLE_NFS);
|
||||||
long nr_to_write;
|
long nr_to_write;
|
||||||
|
@ -1122,9 +1137,7 @@ long writeback_inodes_sb(struct super_block *sb)
|
||||||
nr_to_write = nr_dirty + nr_unstable +
|
nr_to_write = nr_dirty + nr_unstable +
|
||||||
(inodes_stat.nr_inodes - inodes_stat.nr_unused);
|
(inodes_stat.nr_inodes - inodes_stat.nr_unused);
|
||||||
|
|
||||||
wbc.nr_to_write = nr_to_write;
|
bdi_writeback_all(sb, nr_to_write);
|
||||||
bdi_writeback_all(&wbc);
|
|
||||||
return nr_to_write - wbc.nr_to_write;
|
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(writeback_inodes_sb);
|
EXPORT_SYMBOL(writeback_inodes_sb);
|
||||||
|
|
||||||
|
@ -1135,21 +1148,10 @@ EXPORT_SYMBOL(writeback_inodes_sb);
|
||||||
* This function writes and waits on any dirty inode belonging to this
|
* This function writes and waits on any dirty inode belonging to this
|
||||||
* super_block. The number of pages synced is returned.
|
* super_block. The number of pages synced is returned.
|
||||||
*/
|
*/
|
||||||
long sync_inodes_sb(struct super_block *sb)
|
void sync_inodes_sb(struct super_block *sb)
|
||||||
{
|
{
|
||||||
struct writeback_control wbc = {
|
bdi_sync_writeback(sb->s_bdi, sb);
|
||||||
.sb = sb,
|
wait_sb_inodes(sb);
|
||||||
.bdi = sb->s_bdi,
|
|
||||||
.sync_mode = WB_SYNC_ALL,
|
|
||||||
.range_start = 0,
|
|
||||||
.range_end = LLONG_MAX,
|
|
||||||
};
|
|
||||||
long nr_to_write = LONG_MAX; /* doesn't actually matter */
|
|
||||||
|
|
||||||
wbc.nr_to_write = nr_to_write;
|
|
||||||
bdi_start_writeback(&wbc);
|
|
||||||
wait_sb_inodes(&wbc);
|
|
||||||
return nr_to_write - wbc.nr_to_write;
|
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(sync_inodes_sb);
|
EXPORT_SYMBOL(sync_inodes_sb);
|
||||||
|
|
||||||
|
|
|
@ -54,29 +54,15 @@
|
||||||
* @nr_to_write: how many dirty pages to write-back
|
* @nr_to_write: how many dirty pages to write-back
|
||||||
*
|
*
|
||||||
* This function shrinks UBIFS liability by means of writing back some amount
|
* This function shrinks UBIFS liability by means of writing back some amount
|
||||||
* of dirty inodes and their pages. Returns the amount of pages which were
|
* of dirty inodes and their pages.
|
||||||
* written back. The returned value does not include dirty inodes which were
|
|
||||||
* synchronized.
|
|
||||||
*
|
*
|
||||||
* Note, this function synchronizes even VFS inodes which are locked
|
* Note, this function synchronizes even VFS inodes which are locked
|
||||||
* (@i_mutex) by the caller of the budgeting function, because write-back does
|
* (@i_mutex) by the caller of the budgeting function, because write-back does
|
||||||
* not touch @i_mutex.
|
* not touch @i_mutex.
|
||||||
*/
|
*/
|
||||||
static int shrink_liability(struct ubifs_info *c, int nr_to_write)
|
static void shrink_liability(struct ubifs_info *c, int nr_to_write)
|
||||||
{
|
{
|
||||||
int nr_written;
|
writeback_inodes_sb(c->vfs_sb);
|
||||||
|
|
||||||
nr_written = writeback_inodes_sb(c->vfs_sb);
|
|
||||||
if (!nr_written) {
|
|
||||||
/*
|
|
||||||
* Re-try again but wait on pages/inodes which are being
|
|
||||||
* written-back concurrently (e.g., by pdflush).
|
|
||||||
*/
|
|
||||||
nr_written = sync_inodes_sb(c->vfs_sb);
|
|
||||||
}
|
|
||||||
|
|
||||||
dbg_budg("%d pages were written back", nr_written);
|
|
||||||
return nr_written;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
|
|
@ -101,7 +101,7 @@ int bdi_register(struct backing_dev_info *bdi, struct device *parent,
|
||||||
const char *fmt, ...);
|
const char *fmt, ...);
|
||||||
int bdi_register_dev(struct backing_dev_info *bdi, dev_t dev);
|
int bdi_register_dev(struct backing_dev_info *bdi, dev_t dev);
|
||||||
void bdi_unregister(struct backing_dev_info *bdi);
|
void bdi_unregister(struct backing_dev_info *bdi);
|
||||||
void bdi_start_writeback(struct writeback_control *wbc);
|
void bdi_start_writeback(struct backing_dev_info *bdi, long nr_pages);
|
||||||
int bdi_writeback_task(struct bdi_writeback *wb);
|
int bdi_writeback_task(struct bdi_writeback *wb);
|
||||||
int bdi_has_dirty_io(struct backing_dev_info *bdi);
|
int bdi_has_dirty_io(struct backing_dev_info *bdi);
|
||||||
|
|
||||||
|
|
|
@ -68,8 +68,8 @@ struct writeback_control {
|
||||||
*/
|
*/
|
||||||
struct bdi_writeback;
|
struct bdi_writeback;
|
||||||
int inode_wait(void *);
|
int inode_wait(void *);
|
||||||
long writeback_inodes_sb(struct super_block *);
|
void writeback_inodes_sb(struct super_block *);
|
||||||
long sync_inodes_sb(struct super_block *);
|
void sync_inodes_sb(struct super_block *);
|
||||||
void writeback_inodes_wbc(struct writeback_control *wbc);
|
void writeback_inodes_wbc(struct writeback_control *wbc);
|
||||||
long wb_do_writeback(struct bdi_writeback *wb, int force_wait);
|
long wb_do_writeback(struct bdi_writeback *wb, int force_wait);
|
||||||
void wakeup_flusher_threads(long nr_pages);
|
void wakeup_flusher_threads(long nr_pages);
|
||||||
|
|
|
@ -582,16 +582,8 @@ static void balance_dirty_pages(struct address_space *mapping)
|
||||||
if ((laptop_mode && pages_written) ||
|
if ((laptop_mode && pages_written) ||
|
||||||
(!laptop_mode && ((nr_writeback = global_page_state(NR_FILE_DIRTY)
|
(!laptop_mode && ((nr_writeback = global_page_state(NR_FILE_DIRTY)
|
||||||
+ global_page_state(NR_UNSTABLE_NFS))
|
+ global_page_state(NR_UNSTABLE_NFS))
|
||||||
> background_thresh))) {
|
> background_thresh)))
|
||||||
struct writeback_control wbc = {
|
bdi_start_writeback(bdi, nr_writeback);
|
||||||
.bdi = bdi,
|
|
||||||
.sync_mode = WB_SYNC_NONE,
|
|
||||||
.nr_to_write = nr_writeback,
|
|
||||||
};
|
|
||||||
|
|
||||||
|
|
||||||
bdi_start_writeback(&wbc);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void set_page_dirty_balance(struct page *page, int page_mkwrite)
|
void set_page_dirty_balance(struct page *page, int page_mkwrite)
|
||||||
|
|
Loading…
Reference in a new issue