mirror of
https://github.com/adulau/aha.git
synced 2024-12-29 12:16:20 +00:00
08ca0db8aa
A read request outside i_size will be handled in do_generic_file_read(). So we just return 0 to avoid getting -EIO as normal reading, let do_generic_file_read do the rest. At the same time we need unlock the page to avoid system stuck. Fixes http://bugzilla.kernel.org/show_bug.cgi?id=10227 Signed-off-by: Dave Young <hidave.darkstar@gmail.com> Acked-by: Jan Kara <jack@suse.cz> Report-by: Christian Perle <chris@linuxinfotag.de> Cc: <stable@kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
343 lines
9.8 KiB
C
343 lines
9.8 KiB
C
/* -*- linux-c -*- ------------------------------------------------------- *
|
|
*
|
|
* Copyright 2001 H. Peter Anvin - All Rights Reserved
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation, Inc., 675 Mass Ave, Cambridge MA 02139,
|
|
* USA; either version 2 of the License, or (at your option) any later
|
|
* version; incorporated herein by reference.
|
|
*
|
|
* ----------------------------------------------------------------------- */
|
|
|
|
/*
|
|
* linux/fs/isofs/compress.c
|
|
*
|
|
* Transparent decompression of files on an iso9660 filesystem
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/zlib.h>
|
|
|
|
#include "isofs.h"
|
|
#include "zisofs.h"
|
|
|
|
/* This should probably be global. */
|
|
static char zisofs_sink_page[PAGE_CACHE_SIZE];
|
|
|
|
/*
|
|
* This contains the zlib memory allocation and the mutex for the
|
|
* allocation; this avoids failures at block-decompression time.
|
|
*/
|
|
static void *zisofs_zlib_workspace;
|
|
static DEFINE_MUTEX(zisofs_zlib_lock);
|
|
|
|
/*
|
|
* When decompressing, we typically obtain more than one page
|
|
* per reference. We inject the additional pages into the page
|
|
* cache as a form of readahead.
|
|
*/
|
|
static int zisofs_readpage(struct file *file, struct page *page)
|
|
{
|
|
struct inode *inode = file->f_path.dentry->d_inode;
|
|
struct address_space *mapping = inode->i_mapping;
|
|
unsigned int maxpage, xpage, fpage, blockindex;
|
|
unsigned long offset;
|
|
unsigned long blockptr, blockendptr, cstart, cend, csize;
|
|
struct buffer_head *bh, *ptrbh[2];
|
|
unsigned long bufsize = ISOFS_BUFFER_SIZE(inode);
|
|
unsigned int bufshift = ISOFS_BUFFER_BITS(inode);
|
|
unsigned long bufmask = bufsize - 1;
|
|
int err = -EIO;
|
|
int i;
|
|
unsigned int header_size = ISOFS_I(inode)->i_format_parm[0];
|
|
unsigned int zisofs_block_shift = ISOFS_I(inode)->i_format_parm[1];
|
|
/* unsigned long zisofs_block_size = 1UL << zisofs_block_shift; */
|
|
unsigned int zisofs_block_page_shift = zisofs_block_shift-PAGE_CACHE_SHIFT;
|
|
unsigned long zisofs_block_pages = 1UL << zisofs_block_page_shift;
|
|
unsigned long zisofs_block_page_mask = zisofs_block_pages-1;
|
|
struct page *pages[zisofs_block_pages];
|
|
unsigned long index = page->index;
|
|
int indexblocks;
|
|
|
|
/* We have already been given one page, this is the one
|
|
we must do. */
|
|
xpage = index & zisofs_block_page_mask;
|
|
pages[xpage] = page;
|
|
|
|
/* The remaining pages need to be allocated and inserted */
|
|
offset = index & ~zisofs_block_page_mask;
|
|
blockindex = offset >> zisofs_block_page_shift;
|
|
maxpage = (inode->i_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT;
|
|
|
|
/*
|
|
* If this page is wholly outside i_size we just return zero;
|
|
* do_generic_file_read() will handle this for us
|
|
*/
|
|
if (page->index >= maxpage) {
|
|
SetPageUptodate(page);
|
|
unlock_page(page);
|
|
return 0;
|
|
}
|
|
|
|
maxpage = min(zisofs_block_pages, maxpage-offset);
|
|
|
|
for ( i = 0 ; i < maxpage ; i++, offset++ ) {
|
|
if ( i != xpage ) {
|
|
pages[i] = grab_cache_page_nowait(mapping, offset);
|
|
}
|
|
page = pages[i];
|
|
if ( page ) {
|
|
ClearPageError(page);
|
|
kmap(page);
|
|
}
|
|
}
|
|
|
|
/* This is the last page filled, plus one; used in case of abort. */
|
|
fpage = 0;
|
|
|
|
/* Find the pointer to this specific chunk */
|
|
/* Note: we're not using isonum_731() here because the data is known aligned */
|
|
/* Note: header_size is in 32-bit words (4 bytes) */
|
|
blockptr = (header_size + blockindex) << 2;
|
|
blockendptr = blockptr + 4;
|
|
|
|
indexblocks = ((blockptr^blockendptr) >> bufshift) ? 2 : 1;
|
|
ptrbh[0] = ptrbh[1] = NULL;
|
|
|
|
if ( isofs_get_blocks(inode, blockptr >> bufshift, ptrbh, indexblocks) != indexblocks ) {
|
|
if ( ptrbh[0] ) brelse(ptrbh[0]);
|
|
printk(KERN_DEBUG "zisofs: Null buffer on reading block table, inode = %lu, block = %lu\n",
|
|
inode->i_ino, blockptr >> bufshift);
|
|
goto eio;
|
|
}
|
|
ll_rw_block(READ, indexblocks, ptrbh);
|
|
|
|
bh = ptrbh[0];
|
|
if ( !bh || (wait_on_buffer(bh), !buffer_uptodate(bh)) ) {
|
|
printk(KERN_DEBUG "zisofs: Failed to read block table, inode = %lu, block = %lu\n",
|
|
inode->i_ino, blockptr >> bufshift);
|
|
if ( ptrbh[1] )
|
|
brelse(ptrbh[1]);
|
|
goto eio;
|
|
}
|
|
cstart = le32_to_cpu(*(__le32 *)(bh->b_data + (blockptr & bufmask)));
|
|
|
|
if ( indexblocks == 2 ) {
|
|
/* We just crossed a block boundary. Switch to the next block */
|
|
brelse(bh);
|
|
bh = ptrbh[1];
|
|
if ( !bh || (wait_on_buffer(bh), !buffer_uptodate(bh)) ) {
|
|
printk(KERN_DEBUG "zisofs: Failed to read block table, inode = %lu, block = %lu\n",
|
|
inode->i_ino, blockendptr >> bufshift);
|
|
goto eio;
|
|
}
|
|
}
|
|
cend = le32_to_cpu(*(__le32 *)(bh->b_data + (blockendptr & bufmask)));
|
|
brelse(bh);
|
|
|
|
if (cstart > cend)
|
|
goto eio;
|
|
|
|
csize = cend-cstart;
|
|
|
|
if (csize > deflateBound(1UL << zisofs_block_shift))
|
|
goto eio;
|
|
|
|
/* Now page[] contains an array of pages, any of which can be NULL,
|
|
and the locks on which we hold. We should now read the data and
|
|
release the pages. If the pages are NULL the decompressed data
|
|
for that particular page should be discarded. */
|
|
|
|
if ( csize == 0 ) {
|
|
/* This data block is empty. */
|
|
|
|
for ( fpage = 0 ; fpage < maxpage ; fpage++ ) {
|
|
if ( (page = pages[fpage]) != NULL ) {
|
|
memset(page_address(page), 0, PAGE_CACHE_SIZE);
|
|
|
|
flush_dcache_page(page);
|
|
SetPageUptodate(page);
|
|
kunmap(page);
|
|
unlock_page(page);
|
|
if ( fpage == xpage )
|
|
err = 0; /* The critical page */
|
|
else
|
|
page_cache_release(page);
|
|
}
|
|
}
|
|
} else {
|
|
/* This data block is compressed. */
|
|
z_stream stream;
|
|
int bail = 0, left_out = -1;
|
|
int zerr;
|
|
int needblocks = (csize + (cstart & bufmask) + bufmask) >> bufshift;
|
|
int haveblocks;
|
|
struct buffer_head *bhs[needblocks+1];
|
|
struct buffer_head **bhptr;
|
|
|
|
/* Because zlib is not thread-safe, do all the I/O at the top. */
|
|
|
|
blockptr = cstart >> bufshift;
|
|
memset(bhs, 0, (needblocks+1)*sizeof(struct buffer_head *));
|
|
haveblocks = isofs_get_blocks(inode, blockptr, bhs, needblocks);
|
|
ll_rw_block(READ, haveblocks, bhs);
|
|
|
|
bhptr = &bhs[0];
|
|
bh = *bhptr++;
|
|
|
|
/* First block is special since it may be fractional.
|
|
We also wait for it before grabbing the zlib
|
|
mutex; odds are that the subsequent blocks are
|
|
going to come in in short order so we don't hold
|
|
the zlib mutex longer than necessary. */
|
|
|
|
if ( !bh || (wait_on_buffer(bh), !buffer_uptodate(bh)) ) {
|
|
printk(KERN_DEBUG "zisofs: Hit null buffer, fpage = %d, xpage = %d, csize = %ld\n",
|
|
fpage, xpage, csize);
|
|
goto b_eio;
|
|
}
|
|
stream.next_in = bh->b_data + (cstart & bufmask);
|
|
stream.avail_in = min(bufsize-(cstart & bufmask), csize);
|
|
csize -= stream.avail_in;
|
|
|
|
stream.workspace = zisofs_zlib_workspace;
|
|
mutex_lock(&zisofs_zlib_lock);
|
|
|
|
zerr = zlib_inflateInit(&stream);
|
|
if ( zerr != Z_OK ) {
|
|
if ( err && zerr == Z_MEM_ERROR )
|
|
err = -ENOMEM;
|
|
printk(KERN_DEBUG "zisofs: zisofs_inflateInit returned %d\n",
|
|
zerr);
|
|
goto z_eio;
|
|
}
|
|
|
|
while ( !bail && fpage < maxpage ) {
|
|
page = pages[fpage];
|
|
if ( page )
|
|
stream.next_out = page_address(page);
|
|
else
|
|
stream.next_out = (void *)&zisofs_sink_page;
|
|
stream.avail_out = PAGE_CACHE_SIZE;
|
|
|
|
while ( stream.avail_out ) {
|
|
int ao, ai;
|
|
if ( stream.avail_in == 0 && left_out ) {
|
|
if ( !csize ) {
|
|
printk(KERN_WARNING "zisofs: ZF read beyond end of input\n");
|
|
bail = 1;
|
|
break;
|
|
} else {
|
|
bh = *bhptr++;
|
|
if ( !bh ||
|
|
(wait_on_buffer(bh), !buffer_uptodate(bh)) ) {
|
|
/* Reached an EIO */
|
|
printk(KERN_DEBUG "zisofs: Hit null buffer, fpage = %d, xpage = %d, csize = %ld\n",
|
|
fpage, xpage, csize);
|
|
|
|
bail = 1;
|
|
break;
|
|
}
|
|
stream.next_in = bh->b_data;
|
|
stream.avail_in = min(csize,bufsize);
|
|
csize -= stream.avail_in;
|
|
}
|
|
}
|
|
ao = stream.avail_out; ai = stream.avail_in;
|
|
zerr = zlib_inflate(&stream, Z_SYNC_FLUSH);
|
|
left_out = stream.avail_out;
|
|
if ( zerr == Z_BUF_ERROR && stream.avail_in == 0 )
|
|
continue;
|
|
if ( zerr != Z_OK ) {
|
|
/* EOF, error, or trying to read beyond end of input */
|
|
if ( err && zerr == Z_MEM_ERROR )
|
|
err = -ENOMEM;
|
|
if ( zerr != Z_STREAM_END )
|
|
printk(KERN_DEBUG "zisofs: zisofs_inflate returned %d, inode = %lu, index = %lu, fpage = %d, xpage = %d, avail_in = %d, avail_out = %d, ai = %d, ao = %d\n",
|
|
zerr, inode->i_ino, index,
|
|
fpage, xpage,
|
|
stream.avail_in, stream.avail_out,
|
|
ai, ao);
|
|
bail = 1;
|
|
break;
|
|
}
|
|
}
|
|
|
|
if ( stream.avail_out && zerr == Z_STREAM_END ) {
|
|
/* Fractional page written before EOF. This may
|
|
be the last page in the file. */
|
|
memset(stream.next_out, 0, stream.avail_out);
|
|
stream.avail_out = 0;
|
|
}
|
|
|
|
if ( !stream.avail_out ) {
|
|
/* This page completed */
|
|
if ( page ) {
|
|
flush_dcache_page(page);
|
|
SetPageUptodate(page);
|
|
kunmap(page);
|
|
unlock_page(page);
|
|
if ( fpage == xpage )
|
|
err = 0; /* The critical page */
|
|
else
|
|
page_cache_release(page);
|
|
}
|
|
fpage++;
|
|
}
|
|
}
|
|
zlib_inflateEnd(&stream);
|
|
|
|
z_eio:
|
|
mutex_unlock(&zisofs_zlib_lock);
|
|
|
|
b_eio:
|
|
for ( i = 0 ; i < haveblocks ; i++ ) {
|
|
if ( bhs[i] )
|
|
brelse(bhs[i]);
|
|
}
|
|
}
|
|
|
|
eio:
|
|
|
|
/* Release any residual pages, do not SetPageUptodate */
|
|
while ( fpage < maxpage ) {
|
|
page = pages[fpage];
|
|
if ( page ) {
|
|
flush_dcache_page(page);
|
|
if ( fpage == xpage )
|
|
SetPageError(page);
|
|
kunmap(page);
|
|
unlock_page(page);
|
|
if ( fpage != xpage )
|
|
page_cache_release(page);
|
|
}
|
|
fpage++;
|
|
}
|
|
|
|
/* At this point, err contains 0 or -EIO depending on the "critical" page */
|
|
return err;
|
|
}
|
|
|
|
const struct address_space_operations zisofs_aops = {
|
|
.readpage = zisofs_readpage,
|
|
/* No sync_page operation supported? */
|
|
/* No bmap operation supported */
|
|
};
|
|
|
|
int __init zisofs_init(void)
|
|
{
|
|
zisofs_zlib_workspace = vmalloc(zlib_inflate_workspacesize());
|
|
if ( !zisofs_zlib_workspace )
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
void zisofs_cleanup(void)
|
|
{
|
|
vfree(zisofs_zlib_workspace);
|
|
}
|