 08ca0db8aa
			
		
	
	
	08ca0db8aa
	
	
	
		
			
			A read request outside i_size will be handled in do_generic_file_read(). So we just return 0 to avoid getting -EIO as normal reading, let do_generic_file_read do the rest. At the same time we need unlock the page to avoid system stuck. Fixes http://bugzilla.kernel.org/show_bug.cgi?id=10227 Signed-off-by: Dave Young <hidave.darkstar@gmail.com> Acked-by: Jan Kara <jack@suse.cz> Report-by: Christian Perle <chris@linuxinfotag.de> Cc: <stable@kernel.org> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
		
			
				
	
	
		
			343 lines
		
	
	
	
		
			9.8 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			343 lines
		
	
	
	
		
			9.8 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /* -*- linux-c -*- ------------------------------------------------------- *
 | |
|  *   
 | |
|  *   Copyright 2001 H. Peter Anvin - All Rights Reserved
 | |
|  *
 | |
|  *   This program is free software; you can redistribute it and/or modify
 | |
|  *   it under the terms of the GNU General Public License as published by
 | |
|  *   the Free Software Foundation, Inc., 675 Mass Ave, Cambridge MA 02139,
 | |
|  *   USA; either version 2 of the License, or (at your option) any later
 | |
|  *   version; incorporated herein by reference.
 | |
|  *
 | |
|  * ----------------------------------------------------------------------- */
 | |
| 
 | |
| /*
 | |
|  * linux/fs/isofs/compress.c
 | |
|  *
 | |
|  * Transparent decompression of files on an iso9660 filesystem
 | |
|  */
 | |
| 
 | |
| #include <linux/module.h>
 | |
| #include <linux/init.h>
 | |
| 
 | |
| #include <linux/vmalloc.h>
 | |
| #include <linux/zlib.h>
 | |
| 
 | |
| #include "isofs.h"
 | |
| #include "zisofs.h"
 | |
| 
 | |
| /* This should probably be global. */
 | |
| static char zisofs_sink_page[PAGE_CACHE_SIZE];
 | |
| 
 | |
| /*
 | |
|  * This contains the zlib memory allocation and the mutex for the
 | |
|  * allocation; this avoids failures at block-decompression time.
 | |
|  */
 | |
| static void *zisofs_zlib_workspace;
 | |
| static DEFINE_MUTEX(zisofs_zlib_lock);
 | |
| 
 | |
| /*
 | |
|  * When decompressing, we typically obtain more than one page
 | |
|  * per reference.  We inject the additional pages into the page
 | |
|  * cache as a form of readahead.
 | |
|  */
 | |
| static int zisofs_readpage(struct file *file, struct page *page)
 | |
| {
 | |
| 	struct inode *inode = file->f_path.dentry->d_inode;
 | |
| 	struct address_space *mapping = inode->i_mapping;
 | |
| 	unsigned int maxpage, xpage, fpage, blockindex;
 | |
| 	unsigned long offset;
 | |
| 	unsigned long blockptr, blockendptr, cstart, cend, csize;
 | |
| 	struct buffer_head *bh, *ptrbh[2];
 | |
| 	unsigned long bufsize = ISOFS_BUFFER_SIZE(inode);
 | |
| 	unsigned int bufshift = ISOFS_BUFFER_BITS(inode);
 | |
| 	unsigned long bufmask  = bufsize - 1;
 | |
| 	int err = -EIO;
 | |
| 	int i;
 | |
| 	unsigned int header_size = ISOFS_I(inode)->i_format_parm[0];
 | |
| 	unsigned int zisofs_block_shift = ISOFS_I(inode)->i_format_parm[1];
 | |
| 	/* unsigned long zisofs_block_size = 1UL << zisofs_block_shift; */
 | |
| 	unsigned int zisofs_block_page_shift = zisofs_block_shift-PAGE_CACHE_SHIFT;
 | |
| 	unsigned long zisofs_block_pages = 1UL << zisofs_block_page_shift;
 | |
| 	unsigned long zisofs_block_page_mask = zisofs_block_pages-1;
 | |
| 	struct page *pages[zisofs_block_pages];
 | |
| 	unsigned long index = page->index;
 | |
| 	int indexblocks;
 | |
| 
 | |
| 	/* We have already been given one page, this is the one
 | |
| 	   we must do. */
 | |
| 	xpage = index & zisofs_block_page_mask;
 | |
| 	pages[xpage] = page;
 | |
|  
 | |
| 	/* The remaining pages need to be allocated and inserted */
 | |
| 	offset = index & ~zisofs_block_page_mask;
 | |
| 	blockindex = offset >> zisofs_block_page_shift;
 | |
| 	maxpage = (inode->i_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT;
 | |
| 
 | |
| 	/*
 | |
| 	 * If this page is wholly outside i_size we just return zero;
 | |
| 	 * do_generic_file_read() will handle this for us
 | |
| 	 */
 | |
| 	if (page->index >= maxpage) {
 | |
| 		SetPageUptodate(page);
 | |
| 		unlock_page(page);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	maxpage = min(zisofs_block_pages, maxpage-offset);
 | |
| 
 | |
| 	for ( i = 0 ; i < maxpage ; i++, offset++ ) {
 | |
| 		if ( i != xpage ) {
 | |
| 			pages[i] = grab_cache_page_nowait(mapping, offset);
 | |
| 		}
 | |
| 		page = pages[i];
 | |
| 		if ( page ) {
 | |
| 			ClearPageError(page);
 | |
| 			kmap(page);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	/* This is the last page filled, plus one; used in case of abort. */
 | |
| 	fpage = 0;
 | |
| 
 | |
| 	/* Find the pointer to this specific chunk */
 | |
| 	/* Note: we're not using isonum_731() here because the data is known aligned */
 | |
| 	/* Note: header_size is in 32-bit words (4 bytes) */
 | |
| 	blockptr = (header_size + blockindex) << 2;
 | |
| 	blockendptr = blockptr + 4;
 | |
| 
 | |
| 	indexblocks = ((blockptr^blockendptr) >> bufshift) ? 2 : 1;
 | |
| 	ptrbh[0] = ptrbh[1] = NULL;
 | |
| 
 | |
| 	if ( isofs_get_blocks(inode, blockptr >> bufshift, ptrbh, indexblocks) != indexblocks ) {
 | |
| 		if ( ptrbh[0] ) brelse(ptrbh[0]);
 | |
| 		printk(KERN_DEBUG "zisofs: Null buffer on reading block table, inode = %lu, block = %lu\n",
 | |
| 		       inode->i_ino, blockptr >> bufshift);
 | |
| 		goto eio;
 | |
| 	}
 | |
| 	ll_rw_block(READ, indexblocks, ptrbh);
 | |
| 
 | |
| 	bh = ptrbh[0];
 | |
| 	if ( !bh || (wait_on_buffer(bh), !buffer_uptodate(bh)) ) {
 | |
| 		printk(KERN_DEBUG "zisofs: Failed to read block table, inode = %lu, block = %lu\n",
 | |
| 		       inode->i_ino, blockptr >> bufshift);
 | |
| 		if ( ptrbh[1] )
 | |
| 			brelse(ptrbh[1]);
 | |
| 		goto eio;
 | |
| 	}
 | |
| 	cstart = le32_to_cpu(*(__le32 *)(bh->b_data + (blockptr & bufmask)));
 | |
| 
 | |
| 	if ( indexblocks == 2 ) {
 | |
| 		/* We just crossed a block boundary.  Switch to the next block */
 | |
| 		brelse(bh);
 | |
| 		bh = ptrbh[1];
 | |
| 		if ( !bh || (wait_on_buffer(bh), !buffer_uptodate(bh)) ) {
 | |
| 			printk(KERN_DEBUG "zisofs: Failed to read block table, inode = %lu, block = %lu\n",
 | |
| 			       inode->i_ino, blockendptr >> bufshift);
 | |
| 			goto eio;
 | |
| 		}
 | |
| 	}
 | |
| 	cend = le32_to_cpu(*(__le32 *)(bh->b_data + (blockendptr & bufmask)));
 | |
| 	brelse(bh);
 | |
| 
 | |
| 	if (cstart > cend)
 | |
| 		goto eio;
 | |
| 		
 | |
| 	csize = cend-cstart;
 | |
| 
 | |
| 	if (csize > deflateBound(1UL << zisofs_block_shift))
 | |
| 		goto eio;
 | |
| 
 | |
| 	/* Now page[] contains an array of pages, any of which can be NULL,
 | |
| 	   and the locks on which we hold.  We should now read the data and
 | |
| 	   release the pages.  If the pages are NULL the decompressed data
 | |
| 	   for that particular page should be discarded. */
 | |
| 	
 | |
| 	if ( csize == 0 ) {
 | |
| 		/* This data block is empty. */
 | |
| 
 | |
| 		for ( fpage = 0 ; fpage < maxpage ; fpage++ ) {
 | |
| 			if ( (page = pages[fpage]) != NULL ) {
 | |
| 				memset(page_address(page), 0, PAGE_CACHE_SIZE);
 | |
| 				
 | |
| 				flush_dcache_page(page);
 | |
| 				SetPageUptodate(page);
 | |
| 				kunmap(page);
 | |
| 				unlock_page(page);
 | |
| 				if ( fpage == xpage )
 | |
| 					err = 0; /* The critical page */
 | |
| 				else
 | |
| 					page_cache_release(page);
 | |
| 			}
 | |
| 		}
 | |
| 	} else {
 | |
| 		/* This data block is compressed. */
 | |
| 		z_stream stream;
 | |
| 		int bail = 0, left_out = -1;
 | |
| 		int zerr;
 | |
| 		int needblocks = (csize + (cstart & bufmask) + bufmask) >> bufshift;
 | |
| 		int haveblocks;
 | |
| 		struct buffer_head *bhs[needblocks+1];
 | |
| 		struct buffer_head **bhptr;
 | |
| 
 | |
| 		/* Because zlib is not thread-safe, do all the I/O at the top. */
 | |
| 
 | |
| 		blockptr = cstart >> bufshift;
 | |
| 		memset(bhs, 0, (needblocks+1)*sizeof(struct buffer_head *));
 | |
| 		haveblocks = isofs_get_blocks(inode, blockptr, bhs, needblocks);
 | |
| 		ll_rw_block(READ, haveblocks, bhs);
 | |
| 
 | |
| 		bhptr = &bhs[0];
 | |
| 		bh = *bhptr++;
 | |
| 
 | |
| 		/* First block is special since it may be fractional.
 | |
| 		   We also wait for it before grabbing the zlib
 | |
| 		   mutex; odds are that the subsequent blocks are
 | |
| 		   going to come in in short order so we don't hold
 | |
| 		   the zlib mutex longer than necessary. */
 | |
| 
 | |
| 		if ( !bh || (wait_on_buffer(bh), !buffer_uptodate(bh)) ) {
 | |
| 			printk(KERN_DEBUG "zisofs: Hit null buffer, fpage = %d, xpage = %d, csize = %ld\n",
 | |
| 			       fpage, xpage, csize);
 | |
| 			goto b_eio;
 | |
| 		}
 | |
| 		stream.next_in  = bh->b_data + (cstart & bufmask);
 | |
| 		stream.avail_in = min(bufsize-(cstart & bufmask), csize);
 | |
| 		csize -= stream.avail_in;
 | |
| 
 | |
| 		stream.workspace = zisofs_zlib_workspace;
 | |
| 		mutex_lock(&zisofs_zlib_lock);
 | |
| 		
 | |
| 		zerr = zlib_inflateInit(&stream);
 | |
| 		if ( zerr != Z_OK ) {
 | |
| 			if ( err && zerr == Z_MEM_ERROR )
 | |
| 				err = -ENOMEM;
 | |
| 			printk(KERN_DEBUG "zisofs: zisofs_inflateInit returned %d\n",
 | |
| 			       zerr);
 | |
| 			goto z_eio;
 | |
| 		}
 | |
| 
 | |
| 		while ( !bail && fpage < maxpage ) {
 | |
| 			page = pages[fpage];
 | |
| 			if ( page )
 | |
| 				stream.next_out = page_address(page);
 | |
| 			else
 | |
| 				stream.next_out = (void *)&zisofs_sink_page;
 | |
| 			stream.avail_out = PAGE_CACHE_SIZE;
 | |
| 
 | |
| 			while ( stream.avail_out ) {
 | |
| 				int ao, ai;
 | |
| 				if ( stream.avail_in == 0 && left_out ) {
 | |
| 					if ( !csize ) {
 | |
| 						printk(KERN_WARNING "zisofs: ZF read beyond end of input\n");
 | |
| 						bail = 1;
 | |
| 						break;
 | |
| 					} else {
 | |
| 						bh = *bhptr++;
 | |
| 						if ( !bh ||
 | |
| 						     (wait_on_buffer(bh), !buffer_uptodate(bh)) ) {
 | |
| 							/* Reached an EIO */
 | |
|  							printk(KERN_DEBUG "zisofs: Hit null buffer, fpage = %d, xpage = %d, csize = %ld\n",
 | |
| 							       fpage, xpage, csize);
 | |
| 							       
 | |
| 							bail = 1;
 | |
| 							break;
 | |
| 						}
 | |
| 						stream.next_in = bh->b_data;
 | |
| 						stream.avail_in = min(csize,bufsize);
 | |
| 						csize -= stream.avail_in;
 | |
| 					}
 | |
| 				}
 | |
| 				ao = stream.avail_out;  ai = stream.avail_in;
 | |
| 				zerr = zlib_inflate(&stream, Z_SYNC_FLUSH);
 | |
| 				left_out = stream.avail_out;
 | |
| 				if ( zerr == Z_BUF_ERROR && stream.avail_in == 0 )
 | |
| 					continue;
 | |
| 				if ( zerr != Z_OK ) {
 | |
| 					/* EOF, error, or trying to read beyond end of input */
 | |
| 					if ( err && zerr == Z_MEM_ERROR )
 | |
| 						err = -ENOMEM;
 | |
| 					if ( zerr != Z_STREAM_END )
 | |
| 						printk(KERN_DEBUG "zisofs: zisofs_inflate returned %d, inode = %lu, index = %lu, fpage = %d, xpage = %d, avail_in = %d, avail_out = %d, ai = %d, ao = %d\n",
 | |
| 						       zerr, inode->i_ino, index,
 | |
| 						       fpage, xpage,
 | |
| 						       stream.avail_in, stream.avail_out,
 | |
| 						       ai, ao);
 | |
| 					bail = 1;
 | |
| 					break;
 | |
| 				}
 | |
| 			}
 | |
| 
 | |
| 			if ( stream.avail_out && zerr == Z_STREAM_END ) {
 | |
| 				/* Fractional page written before EOF.  This may
 | |
| 				   be the last page in the file. */
 | |
| 				memset(stream.next_out, 0, stream.avail_out);
 | |
| 				stream.avail_out = 0;
 | |
| 			}
 | |
| 
 | |
| 			if ( !stream.avail_out ) {
 | |
| 				/* This page completed */
 | |
| 				if ( page ) {
 | |
| 					flush_dcache_page(page);
 | |
| 					SetPageUptodate(page);
 | |
| 					kunmap(page);
 | |
| 					unlock_page(page);
 | |
| 					if ( fpage == xpage )
 | |
| 						err = 0; /* The critical page */
 | |
| 					else
 | |
| 						page_cache_release(page);
 | |
| 				}
 | |
| 				fpage++;
 | |
| 			}
 | |
| 		}
 | |
| 		zlib_inflateEnd(&stream);
 | |
| 
 | |
| 	z_eio:
 | |
| 		mutex_unlock(&zisofs_zlib_lock);
 | |
| 
 | |
| 	b_eio:
 | |
| 		for ( i = 0 ; i < haveblocks ; i++ ) {
 | |
| 			if ( bhs[i] )
 | |
| 				brelse(bhs[i]);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| eio:
 | |
| 
 | |
| 	/* Release any residual pages, do not SetPageUptodate */
 | |
| 	while ( fpage < maxpage ) {
 | |
| 		page = pages[fpage];
 | |
| 		if ( page ) {
 | |
| 			flush_dcache_page(page);
 | |
| 			if ( fpage == xpage )
 | |
| 				SetPageError(page);
 | |
| 			kunmap(page);
 | |
| 			unlock_page(page);
 | |
| 			if ( fpage != xpage )
 | |
| 				page_cache_release(page);
 | |
| 		}
 | |
| 		fpage++;
 | |
| 	}			
 | |
| 
 | |
| 	/* At this point, err contains 0 or -EIO depending on the "critical" page */
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| const struct address_space_operations zisofs_aops = {
 | |
| 	.readpage = zisofs_readpage,
 | |
| 	/* No sync_page operation supported? */
 | |
| 	/* No bmap operation supported */
 | |
| };
 | |
| 
 | |
| int __init zisofs_init(void)
 | |
| {
 | |
| 	zisofs_zlib_workspace = vmalloc(zlib_inflate_workspacesize());
 | |
| 	if ( !zisofs_zlib_workspace )
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void zisofs_cleanup(void)
 | |
| {
 | |
| 	vfree(zisofs_zlib_workspace);
 | |
| }
 |