mirror of
https://github.com/adulau/aha.git
synced 2024-12-28 11:46:19 +00:00
prevent cifs_writepages() from skipping unwritten pages
Fixes a data corruption under heavy stress in which pages could be left dirty after all open instances of a inode have been closed. In order to write contiguous pages whenever possible, cifs_writepages() asks pagevec_lookup_tag() for more pages than it may write at one time. Normally, it then resets index just past the last page written before calling pagevec_lookup_tag() again. If cifs_writepages() can't write the first page returned, it wasn't resetting index, and the next call to pagevec_lookup_tag() resulted in skipping all of the pages it previously returned, even though cifs_writepages() did nothing with them. This can result in data loss when the file descriptor is about to be closed. This patch ensures that index gets set back to the next returned page so that none get skipped. Signed-off-by: Dave Kleikamp <shaggy@linux.vnet.ibm.com> Acked-by: Jeff Layton <jlayton@redhat.com> Cc: Shirish S Pargaonkar <shirishp@us.ibm.com> Signed-off-by: Steve French <sfrench@us.ibm.com>
This commit is contained in:
parent
2c55608f28
commit
b066a48c95
1 changed files with 4 additions and 1 deletions
|
@ -1404,7 +1404,10 @@ retry:
|
|||
if ((wbc->nr_to_write -= n_iov) <= 0)
|
||||
done = 1;
|
||||
index = next;
|
||||
}
|
||||
} else
|
||||
/* Need to re-find the pages we skipped */
|
||||
index = pvec.pages[0]->index + 1;
|
||||
|
||||
pagevec_release(&pvec);
|
||||
}
|
||||
if (!scanned && !done) {
|
||||
|
|
Loading…
Reference in a new issue