 7af1168693
			
		
	
	
	7af1168693
	
	
	
		
			
			Calls into highlevel quota code cannot happen under the write lock. These calls take dqio_mutex which ranks above write lock. So drop write lock before calling back into quota code. CC: stable@vger.kernel.org # >= 3.0 Signed-off-by: Jan Kara <jack@suse.cz>
		
			
				
	
	
		
			2124 lines
		
	
	
	
		
			64 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			2124 lines
		
	
	
	
		
			64 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  *  Copyright 2000 by Hans Reiser, licensing governed by reiserfs/README
 | |
|  */
 | |
| 
 | |
| /*
 | |
|  *  Written by Anatoly P. Pinchuk pap@namesys.botik.ru
 | |
|  *  Programm System Institute
 | |
|  *  Pereslavl-Zalessky Russia
 | |
|  */
 | |
| 
 | |
| /*
 | |
|  *  This file contains functions dealing with S+tree
 | |
|  *
 | |
|  * B_IS_IN_TREE
 | |
|  * copy_item_head
 | |
|  * comp_short_keys
 | |
|  * comp_keys
 | |
|  * comp_short_le_keys
 | |
|  * le_key2cpu_key
 | |
|  * comp_le_keys
 | |
|  * bin_search
 | |
|  * get_lkey
 | |
|  * get_rkey
 | |
|  * key_in_buffer
 | |
|  * decrement_bcount
 | |
|  * reiserfs_check_path
 | |
|  * pathrelse_and_restore
 | |
|  * pathrelse
 | |
|  * search_by_key_reada
 | |
|  * search_by_key
 | |
|  * search_for_position_by_key
 | |
|  * comp_items
 | |
|  * prepare_for_direct_item
 | |
|  * prepare_for_direntry_item
 | |
|  * prepare_for_delete_or_cut
 | |
|  * calc_deleted_bytes_number
 | |
|  * init_tb_struct
 | |
|  * padd_item
 | |
|  * reiserfs_delete_item
 | |
|  * reiserfs_delete_solid_item
 | |
|  * reiserfs_delete_object
 | |
|  * maybe_indirect_to_direct
 | |
|  * indirect_to_direct_roll_back
 | |
|  * reiserfs_cut_from_item
 | |
|  * truncate_directory
 | |
|  * reiserfs_do_truncate
 | |
|  * reiserfs_paste_into_item
 | |
|  * reiserfs_insert_item
 | |
|  */
 | |
| 
 | |
| #include <linux/time.h>
 | |
| #include <linux/string.h>
 | |
| #include <linux/pagemap.h>
 | |
| #include "reiserfs.h"
 | |
| #include <linux/buffer_head.h>
 | |
| #include <linux/quotaops.h>
 | |
| 
 | |
| /* Does the buffer contain a disk block which is in the tree. */
 | |
| inline int B_IS_IN_TREE(const struct buffer_head *bh)
 | |
| {
 | |
| 
 | |
| 	RFALSE(B_LEVEL(bh) > MAX_HEIGHT,
 | |
| 	       "PAP-1010: block (%b) has too big level (%z)", bh, bh);
 | |
| 
 | |
| 	return (B_LEVEL(bh) != FREE_LEVEL);
 | |
| }
 | |
| 
 | |
| //
 | |
| // to gets item head in le form
 | |
| //
 | |
| inline void copy_item_head(struct item_head *to,
 | |
| 			   const struct item_head *from)
 | |
| {
 | |
| 	memcpy(to, from, IH_SIZE);
 | |
| }
 | |
| 
 | |
| /* k1 is pointer to on-disk structure which is stored in little-endian
 | |
|    form. k2 is pointer to cpu variable. For key of items of the same
 | |
|    object this returns 0.
 | |
|    Returns: -1 if key1 < key2
 | |
|    0 if key1 == key2
 | |
|    1 if key1 > key2 */
 | |
| inline int comp_short_keys(const struct reiserfs_key *le_key,
 | |
| 			   const struct cpu_key *cpu_key)
 | |
| {
 | |
| 	__u32 n;
 | |
| 	n = le32_to_cpu(le_key->k_dir_id);
 | |
| 	if (n < cpu_key->on_disk_key.k_dir_id)
 | |
| 		return -1;
 | |
| 	if (n > cpu_key->on_disk_key.k_dir_id)
 | |
| 		return 1;
 | |
| 	n = le32_to_cpu(le_key->k_objectid);
 | |
| 	if (n < cpu_key->on_disk_key.k_objectid)
 | |
| 		return -1;
 | |
| 	if (n > cpu_key->on_disk_key.k_objectid)
 | |
| 		return 1;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* k1 is pointer to on-disk structure which is stored in little-endian
 | |
|    form. k2 is pointer to cpu variable.
 | |
|    Compare keys using all 4 key fields.
 | |
|    Returns: -1 if key1 < key2 0
 | |
|    if key1 = key2 1 if key1 > key2 */
 | |
| static inline int comp_keys(const struct reiserfs_key *le_key,
 | |
| 			    const struct cpu_key *cpu_key)
 | |
| {
 | |
| 	int retval;
 | |
| 
 | |
| 	retval = comp_short_keys(le_key, cpu_key);
 | |
| 	if (retval)
 | |
| 		return retval;
 | |
| 	if (le_key_k_offset(le_key_version(le_key), le_key) <
 | |
| 	    cpu_key_k_offset(cpu_key))
 | |
| 		return -1;
 | |
| 	if (le_key_k_offset(le_key_version(le_key), le_key) >
 | |
| 	    cpu_key_k_offset(cpu_key))
 | |
| 		return 1;
 | |
| 
 | |
| 	if (cpu_key->key_length == 3)
 | |
| 		return 0;
 | |
| 
 | |
| 	/* this part is needed only when tail conversion is in progress */
 | |
| 	if (le_key_k_type(le_key_version(le_key), le_key) <
 | |
| 	    cpu_key_k_type(cpu_key))
 | |
| 		return -1;
 | |
| 
 | |
| 	if (le_key_k_type(le_key_version(le_key), le_key) >
 | |
| 	    cpu_key_k_type(cpu_key))
 | |
| 		return 1;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| inline int comp_short_le_keys(const struct reiserfs_key *key1,
 | |
| 			      const struct reiserfs_key *key2)
 | |
| {
 | |
| 	__u32 *k1_u32, *k2_u32;
 | |
| 	int key_length = REISERFS_SHORT_KEY_LEN;
 | |
| 
 | |
| 	k1_u32 = (__u32 *) key1;
 | |
| 	k2_u32 = (__u32 *) key2;
 | |
| 	for (; key_length--; ++k1_u32, ++k2_u32) {
 | |
| 		if (le32_to_cpu(*k1_u32) < le32_to_cpu(*k2_u32))
 | |
| 			return -1;
 | |
| 		if (le32_to_cpu(*k1_u32) > le32_to_cpu(*k2_u32))
 | |
| 			return 1;
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| inline void le_key2cpu_key(struct cpu_key *to, const struct reiserfs_key *from)
 | |
| {
 | |
| 	int version;
 | |
| 	to->on_disk_key.k_dir_id = le32_to_cpu(from->k_dir_id);
 | |
| 	to->on_disk_key.k_objectid = le32_to_cpu(from->k_objectid);
 | |
| 
 | |
| 	// find out version of the key
 | |
| 	version = le_key_version(from);
 | |
| 	to->version = version;
 | |
| 	to->on_disk_key.k_offset = le_key_k_offset(version, from);
 | |
| 	to->on_disk_key.k_type = le_key_k_type(version, from);
 | |
| }
 | |
| 
 | |
| // this does not say which one is bigger, it only returns 1 if keys
 | |
| // are not equal, 0 otherwise
 | |
| inline int comp_le_keys(const struct reiserfs_key *k1,
 | |
| 			const struct reiserfs_key *k2)
 | |
| {
 | |
| 	return memcmp(k1, k2, sizeof(struct reiserfs_key));
 | |
| }
 | |
| 
 | |
| /**************************************************************************
 | |
|  *  Binary search toolkit function                                        *
 | |
|  *  Search for an item in the array by the item key                       *
 | |
|  *  Returns:    1 if found,  0 if not found;                              *
 | |
|  *        *pos = number of the searched element if found, else the        *
 | |
|  *        number of the first element that is larger than key.            *
 | |
|  **************************************************************************/
 | |
| /* For those not familiar with binary search: lbound is the leftmost item that it
 | |
|  could be, rbound the rightmost item that it could be.  We examine the item
 | |
|  halfway between lbound and rbound, and that tells us either that we can increase
 | |
|  lbound, or decrease rbound, or that we have found it, or if lbound <= rbound that
 | |
|  there are no possible items, and we have not found it. With each examination we
 | |
|  cut the number of possible items it could be by one more than half rounded down,
 | |
|  or we find it. */
 | |
| static inline int bin_search(const void *key,	/* Key to search for. */
 | |
| 			     const void *base,	/* First item in the array. */
 | |
| 			     int num,	/* Number of items in the array. */
 | |
| 			     int width,	/* Item size in the array.
 | |
| 					   searched. Lest the reader be
 | |
| 					   confused, note that this is crafted
 | |
| 					   as a general function, and when it
 | |
| 					   is applied specifically to the array
 | |
| 					   of item headers in a node, width
 | |
| 					   is actually the item header size not
 | |
| 					   the item size. */
 | |
| 			     int *pos /* Number of the searched for element. */
 | |
|     )
 | |
| {
 | |
| 	int rbound, lbound, j;
 | |
| 
 | |
| 	for (j = ((rbound = num - 1) + (lbound = 0)) / 2;
 | |
| 	     lbound <= rbound; j = (rbound + lbound) / 2)
 | |
| 		switch (comp_keys
 | |
| 			((struct reiserfs_key *)((char *)base + j * width),
 | |
| 			 (struct cpu_key *)key)) {
 | |
| 		case -1:
 | |
| 			lbound = j + 1;
 | |
| 			continue;
 | |
| 		case 1:
 | |
| 			rbound = j - 1;
 | |
| 			continue;
 | |
| 		case 0:
 | |
| 			*pos = j;
 | |
| 			return ITEM_FOUND;	/* Key found in the array.  */
 | |
| 		}
 | |
| 
 | |
| 	/* bin_search did not find given key, it returns position of key,
 | |
| 	   that is minimal and greater than the given one. */
 | |
| 	*pos = lbound;
 | |
| 	return ITEM_NOT_FOUND;
 | |
| }
 | |
| 
 | |
| 
 | |
| /* Minimal possible key. It is never in the tree. */
 | |
| const struct reiserfs_key MIN_KEY = { 0, 0, {{0, 0},} };
 | |
| 
 | |
| /* Maximal possible key. It is never in the tree. */
 | |
| static const struct reiserfs_key MAX_KEY = {
 | |
| 	__constant_cpu_to_le32(0xffffffff),
 | |
| 	__constant_cpu_to_le32(0xffffffff),
 | |
| 	{{__constant_cpu_to_le32(0xffffffff),
 | |
| 	  __constant_cpu_to_le32(0xffffffff)},}
 | |
| };
 | |
| 
 | |
| /* Get delimiting key of the buffer by looking for it in the buffers in the path, starting from the bottom
 | |
|    of the path, and going upwards.  We must check the path's validity at each step.  If the key is not in
 | |
|    the path, there is no delimiting key in the tree (buffer is first or last buffer in tree), and in this
 | |
|    case we return a special key, either MIN_KEY or MAX_KEY. */
 | |
| static inline const struct reiserfs_key *get_lkey(const struct treepath *chk_path,
 | |
| 						  const struct super_block *sb)
 | |
| {
 | |
| 	int position, path_offset = chk_path->path_length;
 | |
| 	struct buffer_head *parent;
 | |
| 
 | |
| 	RFALSE(path_offset < FIRST_PATH_ELEMENT_OFFSET,
 | |
| 	       "PAP-5010: invalid offset in the path");
 | |
| 
 | |
| 	/* While not higher in path than first element. */
 | |
| 	while (path_offset-- > FIRST_PATH_ELEMENT_OFFSET) {
 | |
| 
 | |
| 		RFALSE(!buffer_uptodate
 | |
| 		       (PATH_OFFSET_PBUFFER(chk_path, path_offset)),
 | |
| 		       "PAP-5020: parent is not uptodate");
 | |
| 
 | |
| 		/* Parent at the path is not in the tree now. */
 | |
| 		if (!B_IS_IN_TREE
 | |
| 		    (parent =
 | |
| 		     PATH_OFFSET_PBUFFER(chk_path, path_offset)))
 | |
| 			return &MAX_KEY;
 | |
| 		/* Check whether position in the parent is correct. */
 | |
| 		if ((position =
 | |
| 		     PATH_OFFSET_POSITION(chk_path,
 | |
| 					  path_offset)) >
 | |
| 		    B_NR_ITEMS(parent))
 | |
| 			return &MAX_KEY;
 | |
| 		/* Check whether parent at the path really points to the child. */
 | |
| 		if (B_N_CHILD_NUM(parent, position) !=
 | |
| 		    PATH_OFFSET_PBUFFER(chk_path,
 | |
| 					path_offset + 1)->b_blocknr)
 | |
| 			return &MAX_KEY;
 | |
| 		/* Return delimiting key if position in the parent is not equal to zero. */
 | |
| 		if (position)
 | |
| 			return B_N_PDELIM_KEY(parent, position - 1);
 | |
| 	}
 | |
| 	/* Return MIN_KEY if we are in the root of the buffer tree. */
 | |
| 	if (PATH_OFFSET_PBUFFER(chk_path, FIRST_PATH_ELEMENT_OFFSET)->
 | |
| 	    b_blocknr == SB_ROOT_BLOCK(sb))
 | |
| 		return &MIN_KEY;
 | |
| 	return &MAX_KEY;
 | |
| }
 | |
| 
 | |
| /* Get delimiting key of the buffer at the path and its right neighbor. */
 | |
| inline const struct reiserfs_key *get_rkey(const struct treepath *chk_path,
 | |
| 					   const struct super_block *sb)
 | |
| {
 | |
| 	int position, path_offset = chk_path->path_length;
 | |
| 	struct buffer_head *parent;
 | |
| 
 | |
| 	RFALSE(path_offset < FIRST_PATH_ELEMENT_OFFSET,
 | |
| 	       "PAP-5030: invalid offset in the path");
 | |
| 
 | |
| 	while (path_offset-- > FIRST_PATH_ELEMENT_OFFSET) {
 | |
| 
 | |
| 		RFALSE(!buffer_uptodate
 | |
| 		       (PATH_OFFSET_PBUFFER(chk_path, path_offset)),
 | |
| 		       "PAP-5040: parent is not uptodate");
 | |
| 
 | |
| 		/* Parent at the path is not in the tree now. */
 | |
| 		if (!B_IS_IN_TREE
 | |
| 		    (parent =
 | |
| 		     PATH_OFFSET_PBUFFER(chk_path, path_offset)))
 | |
| 			return &MIN_KEY;
 | |
| 		/* Check whether position in the parent is correct. */
 | |
| 		if ((position =
 | |
| 		     PATH_OFFSET_POSITION(chk_path,
 | |
| 					  path_offset)) >
 | |
| 		    B_NR_ITEMS(parent))
 | |
| 			return &MIN_KEY;
 | |
| 		/* Check whether parent at the path really points to the child. */
 | |
| 		if (B_N_CHILD_NUM(parent, position) !=
 | |
| 		    PATH_OFFSET_PBUFFER(chk_path,
 | |
| 					path_offset + 1)->b_blocknr)
 | |
| 			return &MIN_KEY;
 | |
| 		/* Return delimiting key if position in the parent is not the last one. */
 | |
| 		if (position != B_NR_ITEMS(parent))
 | |
| 			return B_N_PDELIM_KEY(parent, position);
 | |
| 	}
 | |
| 	/* Return MAX_KEY if we are in the root of the buffer tree. */
 | |
| 	if (PATH_OFFSET_PBUFFER(chk_path, FIRST_PATH_ELEMENT_OFFSET)->
 | |
| 	    b_blocknr == SB_ROOT_BLOCK(sb))
 | |
| 		return &MAX_KEY;
 | |
| 	return &MIN_KEY;
 | |
| }
 | |
| 
 | |
| /* Check whether a key is contained in the tree rooted from a buffer at a path. */
 | |
| /* This works by looking at the left and right delimiting keys for the buffer in the last path_element in
 | |
|    the path.  These delimiting keys are stored at least one level above that buffer in the tree. If the
 | |
|    buffer is the first or last node in the tree order then one of the delimiting keys may be absent, and in
 | |
|    this case get_lkey and get_rkey return a special key which is MIN_KEY or MAX_KEY. */
 | |
| static inline int key_in_buffer(struct treepath *chk_path,	/* Path which should be checked.  */
 | |
| 				const struct cpu_key *key,	/* Key which should be checked.   */
 | |
| 				struct super_block *sb
 | |
|     )
 | |
| {
 | |
| 
 | |
| 	RFALSE(!key || chk_path->path_length < FIRST_PATH_ELEMENT_OFFSET
 | |
| 	       || chk_path->path_length > MAX_HEIGHT,
 | |
| 	       "PAP-5050: pointer to the key(%p) is NULL or invalid path length(%d)",
 | |
| 	       key, chk_path->path_length);
 | |
| 	RFALSE(!PATH_PLAST_BUFFER(chk_path)->b_bdev,
 | |
| 	       "PAP-5060: device must not be NODEV");
 | |
| 
 | |
| 	if (comp_keys(get_lkey(chk_path, sb), key) == 1)
 | |
| 		/* left delimiting key is bigger, that the key we look for */
 | |
| 		return 0;
 | |
| 	/*  if ( comp_keys(key, get_rkey(chk_path, sb)) != -1 ) */
 | |
| 	if (comp_keys(get_rkey(chk_path, sb), key) != 1)
 | |
| 		/* key must be less than right delimitiing key */
 | |
| 		return 0;
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| int reiserfs_check_path(struct treepath *p)
 | |
| {
 | |
| 	RFALSE(p->path_length != ILLEGAL_PATH_ELEMENT_OFFSET,
 | |
| 	       "path not properly relsed");
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* Drop the reference to each buffer in a path and restore
 | |
|  * dirty bits clean when preparing the buffer for the log.
 | |
|  * This version should only be called from fix_nodes() */
 | |
| void pathrelse_and_restore(struct super_block *sb,
 | |
| 			   struct treepath *search_path)
 | |
| {
 | |
| 	int path_offset = search_path->path_length;
 | |
| 
 | |
| 	RFALSE(path_offset < ILLEGAL_PATH_ELEMENT_OFFSET,
 | |
| 	       "clm-4000: invalid path offset");
 | |
| 
 | |
| 	while (path_offset > ILLEGAL_PATH_ELEMENT_OFFSET) {
 | |
| 		struct buffer_head *bh;
 | |
| 		bh = PATH_OFFSET_PBUFFER(search_path, path_offset--);
 | |
| 		reiserfs_restore_prepared_buffer(sb, bh);
 | |
| 		brelse(bh);
 | |
| 	}
 | |
| 	search_path->path_length = ILLEGAL_PATH_ELEMENT_OFFSET;
 | |
| }
 | |
| 
 | |
| /* Drop the reference to each buffer in a path */
 | |
| void pathrelse(struct treepath *search_path)
 | |
| {
 | |
| 	int path_offset = search_path->path_length;
 | |
| 
 | |
| 	RFALSE(path_offset < ILLEGAL_PATH_ELEMENT_OFFSET,
 | |
| 	       "PAP-5090: invalid path offset");
 | |
| 
 | |
| 	while (path_offset > ILLEGAL_PATH_ELEMENT_OFFSET)
 | |
| 		brelse(PATH_OFFSET_PBUFFER(search_path, path_offset--));
 | |
| 
 | |
| 	search_path->path_length = ILLEGAL_PATH_ELEMENT_OFFSET;
 | |
| }
 | |
| 
 | |
| static int is_leaf(char *buf, int blocksize, struct buffer_head *bh)
 | |
| {
 | |
| 	struct block_head *blkh;
 | |
| 	struct item_head *ih;
 | |
| 	int used_space;
 | |
| 	int prev_location;
 | |
| 	int i;
 | |
| 	int nr;
 | |
| 
 | |
| 	blkh = (struct block_head *)buf;
 | |
| 	if (blkh_level(blkh) != DISK_LEAF_NODE_LEVEL) {
 | |
| 		reiserfs_warning(NULL, "reiserfs-5080",
 | |
| 				 "this should be caught earlier");
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	nr = blkh_nr_item(blkh);
 | |
| 	if (nr < 1 || nr > ((blocksize - BLKH_SIZE) / (IH_SIZE + MIN_ITEM_LEN))) {
 | |
| 		/* item number is too big or too small */
 | |
| 		reiserfs_warning(NULL, "reiserfs-5081",
 | |
| 				 "nr_item seems wrong: %z", bh);
 | |
| 		return 0;
 | |
| 	}
 | |
| 	ih = (struct item_head *)(buf + BLKH_SIZE) + nr - 1;
 | |
| 	used_space = BLKH_SIZE + IH_SIZE * nr + (blocksize - ih_location(ih));
 | |
| 	if (used_space != blocksize - blkh_free_space(blkh)) {
 | |
| 		/* free space does not match to calculated amount of use space */
 | |
| 		reiserfs_warning(NULL, "reiserfs-5082",
 | |
| 				 "free space seems wrong: %z", bh);
 | |
| 		return 0;
 | |
| 	}
 | |
| 	// FIXME: it is_leaf will hit performance too much - we may have
 | |
| 	// return 1 here
 | |
| 
 | |
| 	/* check tables of item heads */
 | |
| 	ih = (struct item_head *)(buf + BLKH_SIZE);
 | |
| 	prev_location = blocksize;
 | |
| 	for (i = 0; i < nr; i++, ih++) {
 | |
| 		if (le_ih_k_type(ih) == TYPE_ANY) {
 | |
| 			reiserfs_warning(NULL, "reiserfs-5083",
 | |
| 					 "wrong item type for item %h",
 | |
| 					 ih);
 | |
| 			return 0;
 | |
| 		}
 | |
| 		if (ih_location(ih) >= blocksize
 | |
| 		    || ih_location(ih) < IH_SIZE * nr) {
 | |
| 			reiserfs_warning(NULL, "reiserfs-5084",
 | |
| 					 "item location seems wrong: %h",
 | |
| 					 ih);
 | |
| 			return 0;
 | |
| 		}
 | |
| 		if (ih_item_len(ih) < 1
 | |
| 		    || ih_item_len(ih) > MAX_ITEM_LEN(blocksize)) {
 | |
| 			reiserfs_warning(NULL, "reiserfs-5085",
 | |
| 					 "item length seems wrong: %h",
 | |
| 					 ih);
 | |
| 			return 0;
 | |
| 		}
 | |
| 		if (prev_location - ih_location(ih) != ih_item_len(ih)) {
 | |
| 			reiserfs_warning(NULL, "reiserfs-5086",
 | |
| 					 "item location seems wrong "
 | |
| 					 "(second one): %h", ih);
 | |
| 			return 0;
 | |
| 		}
 | |
| 		prev_location = ih_location(ih);
 | |
| 	}
 | |
| 
 | |
| 	// one may imagine much more checks
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| /* returns 1 if buf looks like an internal node, 0 otherwise */
 | |
| static int is_internal(char *buf, int blocksize, struct buffer_head *bh)
 | |
| {
 | |
| 	struct block_head *blkh;
 | |
| 	int nr;
 | |
| 	int used_space;
 | |
| 
 | |
| 	blkh = (struct block_head *)buf;
 | |
| 	nr = blkh_level(blkh);
 | |
| 	if (nr <= DISK_LEAF_NODE_LEVEL || nr > MAX_HEIGHT) {
 | |
| 		/* this level is not possible for internal nodes */
 | |
| 		reiserfs_warning(NULL, "reiserfs-5087",
 | |
| 				 "this should be caught earlier");
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	nr = blkh_nr_item(blkh);
 | |
| 	if (nr > (blocksize - BLKH_SIZE - DC_SIZE) / (KEY_SIZE + DC_SIZE)) {
 | |
| 		/* for internal which is not root we might check min number of keys */
 | |
| 		reiserfs_warning(NULL, "reiserfs-5088",
 | |
| 				 "number of key seems wrong: %z", bh);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	used_space = BLKH_SIZE + KEY_SIZE * nr + DC_SIZE * (nr + 1);
 | |
| 	if (used_space != blocksize - blkh_free_space(blkh)) {
 | |
| 		reiserfs_warning(NULL, "reiserfs-5089",
 | |
| 				 "free space seems wrong: %z", bh);
 | |
| 		return 0;
 | |
| 	}
 | |
| 	// one may imagine much more checks
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| // make sure that bh contains formatted node of reiserfs tree of
 | |
| // 'level'-th level
 | |
| static int is_tree_node(struct buffer_head *bh, int level)
 | |
| {
 | |
| 	if (B_LEVEL(bh) != level) {
 | |
| 		reiserfs_warning(NULL, "reiserfs-5090", "node level %d does "
 | |
| 				 "not match to the expected one %d",
 | |
| 				 B_LEVEL(bh), level);
 | |
| 		return 0;
 | |
| 	}
 | |
| 	if (level == DISK_LEAF_NODE_LEVEL)
 | |
| 		return is_leaf(bh->b_data, bh->b_size, bh);
 | |
| 
 | |
| 	return is_internal(bh->b_data, bh->b_size, bh);
 | |
| }
 | |
| 
 | |
| #define SEARCH_BY_KEY_READA 16
 | |
| 
 | |
| /*
 | |
|  * The function is NOT SCHEDULE-SAFE!
 | |
|  * It might unlock the write lock if we needed to wait for a block
 | |
|  * to be read. Note that in this case it won't recover the lock to avoid
 | |
|  * high contention resulting from too much lock requests, especially
 | |
|  * the caller (search_by_key) will perform other schedule-unsafe
 | |
|  * operations just after calling this function.
 | |
|  *
 | |
|  * @return true if we have unlocked
 | |
|  */
 | |
| static bool search_by_key_reada(struct super_block *s,
 | |
| 				struct buffer_head **bh,
 | |
| 				b_blocknr_t *b, int num)
 | |
| {
 | |
| 	int i, j;
 | |
| 	bool unlocked = false;
 | |
| 
 | |
| 	for (i = 0; i < num; i++) {
 | |
| 		bh[i] = sb_getblk(s, b[i]);
 | |
| 	}
 | |
| 	/*
 | |
| 	 * We are going to read some blocks on which we
 | |
| 	 * have a reference. It's safe, though we might be
 | |
| 	 * reading blocks concurrently changed if we release
 | |
| 	 * the lock. But it's still fine because we check later
 | |
| 	 * if the tree changed
 | |
| 	 */
 | |
| 	for (j = 0; j < i; j++) {
 | |
| 		/*
 | |
| 		 * note, this needs attention if we are getting rid of the BKL
 | |
| 		 * you have to make sure the prepared bit isn't set on this buffer
 | |
| 		 */
 | |
| 		if (!buffer_uptodate(bh[j])) {
 | |
| 			if (!unlocked) {
 | |
| 				reiserfs_write_unlock(s);
 | |
| 				unlocked = true;
 | |
| 			}
 | |
| 			ll_rw_block(READA, 1, bh + j);
 | |
| 		}
 | |
| 		brelse(bh[j]);
 | |
| 	}
 | |
| 	return unlocked;
 | |
| }
 | |
| 
 | |
| /**************************************************************************
 | |
|  * Algorithm   SearchByKey                                                *
 | |
|  *             look for item in the Disk S+Tree by its key                *
 | |
|  * Input:  sb   -  super block                                            *
 | |
|  *         key  - pointer to the key to search                            *
 | |
|  * Output: ITEM_FOUND, ITEM_NOT_FOUND or IO_ERROR                         *
 | |
|  *         search_path - path from the root to the needed leaf            *
 | |
|  **************************************************************************/
 | |
| 
 | |
| /* This function fills up the path from the root to the leaf as it
 | |
|    descends the tree looking for the key.  It uses reiserfs_bread to
 | |
|    try to find buffers in the cache given their block number.  If it
 | |
|    does not find them in the cache it reads them from disk.  For each
 | |
|    node search_by_key finds using reiserfs_bread it then uses
 | |
|    bin_search to look through that node.  bin_search will find the
 | |
|    position of the block_number of the next node if it is looking
 | |
|    through an internal node.  If it is looking through a leaf node
 | |
|    bin_search will find the position of the item which has key either
 | |
|    equal to given key, or which is the maximal key less than the given
 | |
|    key.  search_by_key returns a path that must be checked for the
 | |
|    correctness of the top of the path but need not be checked for the
 | |
|    correctness of the bottom of the path */
 | |
| /* The function is NOT SCHEDULE-SAFE! */
 | |
| int search_by_key(struct super_block *sb, const struct cpu_key *key,	/* Key to search. */
 | |
| 		  struct treepath *search_path,/* This structure was
 | |
| 						   allocated and initialized
 | |
| 						   by the calling
 | |
| 						   function. It is filled up
 | |
| 						   by this function.  */
 | |
| 		  int stop_level	/* How far down the tree to search. To
 | |
| 					   stop at leaf level - set to
 | |
| 					   DISK_LEAF_NODE_LEVEL */
 | |
|     )
 | |
| {
 | |
| 	b_blocknr_t block_number;
 | |
| 	int expected_level;
 | |
| 	struct buffer_head *bh;
 | |
| 	struct path_element *last_element;
 | |
| 	int node_level, retval;
 | |
| 	int right_neighbor_of_leaf_node;
 | |
| 	int fs_gen;
 | |
| 	struct buffer_head *reada_bh[SEARCH_BY_KEY_READA];
 | |
| 	b_blocknr_t reada_blocks[SEARCH_BY_KEY_READA];
 | |
| 	int reada_count = 0;
 | |
| 
 | |
| #ifdef CONFIG_REISERFS_CHECK
 | |
| 	int repeat_counter = 0;
 | |
| #endif
 | |
| 
 | |
| 	PROC_INFO_INC(sb, search_by_key);
 | |
| 
 | |
| 	/* As we add each node to a path we increase its count.  This means that
 | |
| 	   we must be careful to release all nodes in a path before we either
 | |
| 	   discard the path struct or re-use the path struct, as we do here. */
 | |
| 
 | |
| 	pathrelse(search_path);
 | |
| 
 | |
| 	right_neighbor_of_leaf_node = 0;
 | |
| 
 | |
| 	/* With each iteration of this loop we search through the items in the
 | |
| 	   current node, and calculate the next current node(next path element)
 | |
| 	   for the next iteration of this loop.. */
 | |
| 	block_number = SB_ROOT_BLOCK(sb);
 | |
| 	expected_level = -1;
 | |
| 	while (1) {
 | |
| 
 | |
| #ifdef CONFIG_REISERFS_CHECK
 | |
| 		if (!(++repeat_counter % 50000))
 | |
| 			reiserfs_warning(sb, "PAP-5100",
 | |
| 					 "%s: there were %d iterations of "
 | |
| 					 "while loop looking for key %K",
 | |
| 					 current->comm, repeat_counter,
 | |
| 					 key);
 | |
| #endif
 | |
| 
 | |
| 		/* prep path to have another element added to it. */
 | |
| 		last_element =
 | |
| 		    PATH_OFFSET_PELEMENT(search_path,
 | |
| 					 ++search_path->path_length);
 | |
| 		fs_gen = get_generation(sb);
 | |
| 
 | |
| 		/* Read the next tree node, and set the last element in the path to
 | |
| 		   have a pointer to it. */
 | |
| 		if ((bh = last_element->pe_buffer =
 | |
| 		     sb_getblk(sb, block_number))) {
 | |
| 			bool unlocked = false;
 | |
| 
 | |
| 			if (!buffer_uptodate(bh) && reada_count > 1)
 | |
| 				/* may unlock the write lock */
 | |
| 				unlocked = search_by_key_reada(sb, reada_bh,
 | |
| 						    reada_blocks, reada_count);
 | |
| 			/*
 | |
| 			 * If we haven't already unlocked the write lock,
 | |
| 			 * then we need to do that here before reading
 | |
| 			 * the current block
 | |
| 			 */
 | |
| 			if (!buffer_uptodate(bh) && !unlocked) {
 | |
| 				reiserfs_write_unlock(sb);
 | |
| 				unlocked = true;
 | |
| 			}
 | |
| 			ll_rw_block(READ, 1, &bh);
 | |
| 			wait_on_buffer(bh);
 | |
| 
 | |
| 			if (unlocked)
 | |
| 				reiserfs_write_lock(sb);
 | |
| 			if (!buffer_uptodate(bh))
 | |
| 				goto io_error;
 | |
| 		} else {
 | |
| 		      io_error:
 | |
| 			search_path->path_length--;
 | |
| 			pathrelse(search_path);
 | |
| 			return IO_ERROR;
 | |
| 		}
 | |
| 		reada_count = 0;
 | |
| 		if (expected_level == -1)
 | |
| 			expected_level = SB_TREE_HEIGHT(sb);
 | |
| 		expected_level--;
 | |
| 
 | |
| 		/* It is possible that schedule occurred. We must check whether the key
 | |
| 		   to search is still in the tree rooted from the current buffer. If
 | |
| 		   not then repeat search from the root. */
 | |
| 		if (fs_changed(fs_gen, sb) &&
 | |
| 		    (!B_IS_IN_TREE(bh) ||
 | |
| 		     B_LEVEL(bh) != expected_level ||
 | |
| 		     !key_in_buffer(search_path, key, sb))) {
 | |
| 			PROC_INFO_INC(sb, search_by_key_fs_changed);
 | |
| 			PROC_INFO_INC(sb, search_by_key_restarted);
 | |
| 			PROC_INFO_INC(sb,
 | |
| 				      sbk_restarted[expected_level - 1]);
 | |
| 			pathrelse(search_path);
 | |
| 
 | |
| 			/* Get the root block number so that we can repeat the search
 | |
| 			   starting from the root. */
 | |
| 			block_number = SB_ROOT_BLOCK(sb);
 | |
| 			expected_level = -1;
 | |
| 			right_neighbor_of_leaf_node = 0;
 | |
| 
 | |
| 			/* repeat search from the root */
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		/* only check that the key is in the buffer if key is not
 | |
| 		   equal to the MAX_KEY. Latter case is only possible in
 | |
| 		   "finish_unfinished()" processing during mount. */
 | |
| 		RFALSE(comp_keys(&MAX_KEY, key) &&
 | |
| 		       !key_in_buffer(search_path, key, sb),
 | |
| 		       "PAP-5130: key is not in the buffer");
 | |
| #ifdef CONFIG_REISERFS_CHECK
 | |
| 		if (REISERFS_SB(sb)->cur_tb) {
 | |
| 			print_cur_tb("5140");
 | |
| 			reiserfs_panic(sb, "PAP-5140",
 | |
| 				       "schedule occurred in do_balance!");
 | |
| 		}
 | |
| #endif
 | |
| 
 | |
| 		// make sure, that the node contents look like a node of
 | |
| 		// certain level
 | |
| 		if (!is_tree_node(bh, expected_level)) {
 | |
| 			reiserfs_error(sb, "vs-5150",
 | |
| 				       "invalid format found in block %ld. "
 | |
| 				       "Fsck?", bh->b_blocknr);
 | |
| 			pathrelse(search_path);
 | |
| 			return IO_ERROR;
 | |
| 		}
 | |
| 
 | |
| 		/* ok, we have acquired next formatted node in the tree */
 | |
| 		node_level = B_LEVEL(bh);
 | |
| 
 | |
| 		PROC_INFO_BH_STAT(sb, bh, node_level - 1);
 | |
| 
 | |
| 		RFALSE(node_level < stop_level,
 | |
| 		       "vs-5152: tree level (%d) is less than stop level (%d)",
 | |
| 		       node_level, stop_level);
 | |
| 
 | |
| 		retval = bin_search(key, B_N_PITEM_HEAD(bh, 0),
 | |
| 				      B_NR_ITEMS(bh),
 | |
| 				      (node_level ==
 | |
| 				       DISK_LEAF_NODE_LEVEL) ? IH_SIZE :
 | |
| 				      KEY_SIZE,
 | |
| 				      &(last_element->pe_position));
 | |
| 		if (node_level == stop_level) {
 | |
| 			return retval;
 | |
| 		}
 | |
| 
 | |
| 		/* we are not in the stop level */
 | |
| 		if (retval == ITEM_FOUND)
 | |
| 			/* item has been found, so we choose the pointer which is to the right of the found one */
 | |
| 			last_element->pe_position++;
 | |
| 
 | |
| 		/* if item was not found we choose the position which is to
 | |
| 		   the left of the found item. This requires no code,
 | |
| 		   bin_search did it already. */
 | |
| 
 | |
| 		/* So we have chosen a position in the current node which is
 | |
| 		   an internal node.  Now we calculate child block number by
 | |
| 		   position in the node. */
 | |
| 		block_number =
 | |
| 		    B_N_CHILD_NUM(bh, last_element->pe_position);
 | |
| 
 | |
| 		/* if we are going to read leaf nodes, try for read ahead as well */
 | |
| 		if ((search_path->reada & PATH_READA) &&
 | |
| 		    node_level == DISK_LEAF_NODE_LEVEL + 1) {
 | |
| 			int pos = last_element->pe_position;
 | |
| 			int limit = B_NR_ITEMS(bh);
 | |
| 			struct reiserfs_key *le_key;
 | |
| 
 | |
| 			if (search_path->reada & PATH_READA_BACK)
 | |
| 				limit = 0;
 | |
| 			while (reada_count < SEARCH_BY_KEY_READA) {
 | |
| 				if (pos == limit)
 | |
| 					break;
 | |
| 				reada_blocks[reada_count++] =
 | |
| 				    B_N_CHILD_NUM(bh, pos);
 | |
| 				if (search_path->reada & PATH_READA_BACK)
 | |
| 					pos--;
 | |
| 				else
 | |
| 					pos++;
 | |
| 
 | |
| 				/*
 | |
| 				 * check to make sure we're in the same object
 | |
| 				 */
 | |
| 				le_key = B_N_PDELIM_KEY(bh, pos);
 | |
| 				if (le32_to_cpu(le_key->k_objectid) !=
 | |
| 				    key->on_disk_key.k_objectid) {
 | |
| 					break;
 | |
| 				}
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /* Form the path to an item and position in this item which contains
 | |
|    file byte defined by key. If there is no such item
 | |
|    corresponding to the key, we point the path to the item with
 | |
|    maximal key less than key, and *pos_in_item is set to one
 | |
|    past the last entry/byte in the item.  If searching for entry in a
 | |
|    directory item, and it is not found, *pos_in_item is set to one
 | |
|    entry more than the entry with maximal key which is less than the
 | |
|    sought key.
 | |
| 
 | |
|    Note that if there is no entry in this same node which is one more,
 | |
|    then we point to an imaginary entry.  for direct items, the
 | |
|    position is in units of bytes, for indirect items the position is
 | |
|    in units of blocknr entries, for directory items the position is in
 | |
|    units of directory entries.  */
 | |
| 
 | |
| /* The function is NOT SCHEDULE-SAFE! */
 | |
| int search_for_position_by_key(struct super_block *sb,	/* Pointer to the super block.          */
 | |
| 			       const struct cpu_key *p_cpu_key,	/* Key to search (cpu variable)         */
 | |
| 			       struct treepath *search_path	/* Filled up by this function.          */
 | |
|     )
 | |
| {
 | |
| 	struct item_head *p_le_ih;	/* pointer to on-disk structure */
 | |
| 	int blk_size;
 | |
| 	loff_t item_offset, offset;
 | |
| 	struct reiserfs_dir_entry de;
 | |
| 	int retval;
 | |
| 
 | |
| 	/* If searching for directory entry. */
 | |
| 	if (is_direntry_cpu_key(p_cpu_key))
 | |
| 		return search_by_entry_key(sb, p_cpu_key, search_path,
 | |
| 					   &de);
 | |
| 
 | |
| 	/* If not searching for directory entry. */
 | |
| 
 | |
| 	/* If item is found. */
 | |
| 	retval = search_item(sb, p_cpu_key, search_path);
 | |
| 	if (retval == IO_ERROR)
 | |
| 		return retval;
 | |
| 	if (retval == ITEM_FOUND) {
 | |
| 
 | |
| 		RFALSE(!ih_item_len
 | |
| 		       (B_N_PITEM_HEAD
 | |
| 			(PATH_PLAST_BUFFER(search_path),
 | |
| 			 PATH_LAST_POSITION(search_path))),
 | |
| 		       "PAP-5165: item length equals zero");
 | |
| 
 | |
| 		pos_in_item(search_path) = 0;
 | |
| 		return POSITION_FOUND;
 | |
| 	}
 | |
| 
 | |
| 	RFALSE(!PATH_LAST_POSITION(search_path),
 | |
| 	       "PAP-5170: position equals zero");
 | |
| 
 | |
| 	/* Item is not found. Set path to the previous item. */
 | |
| 	p_le_ih =
 | |
| 	    B_N_PITEM_HEAD(PATH_PLAST_BUFFER(search_path),
 | |
| 			   --PATH_LAST_POSITION(search_path));
 | |
| 	blk_size = sb->s_blocksize;
 | |
| 
 | |
| 	if (comp_short_keys(&(p_le_ih->ih_key), p_cpu_key)) {
 | |
| 		return FILE_NOT_FOUND;
 | |
| 	}
 | |
| 	// FIXME: quite ugly this far
 | |
| 
 | |
| 	item_offset = le_ih_k_offset(p_le_ih);
 | |
| 	offset = cpu_key_k_offset(p_cpu_key);
 | |
| 
 | |
| 	/* Needed byte is contained in the item pointed to by the path. */
 | |
| 	if (item_offset <= offset &&
 | |
| 	    item_offset + op_bytes_number(p_le_ih, blk_size) > offset) {
 | |
| 		pos_in_item(search_path) = offset - item_offset;
 | |
| 		if (is_indirect_le_ih(p_le_ih)) {
 | |
| 			pos_in_item(search_path) /= blk_size;
 | |
| 		}
 | |
| 		return POSITION_FOUND;
 | |
| 	}
 | |
| 
 | |
| 	/* Needed byte is not contained in the item pointed to by the
 | |
| 	   path. Set pos_in_item out of the item. */
 | |
| 	if (is_indirect_le_ih(p_le_ih))
 | |
| 		pos_in_item(search_path) =
 | |
| 		    ih_item_len(p_le_ih) / UNFM_P_SIZE;
 | |
| 	else
 | |
| 		pos_in_item(search_path) = ih_item_len(p_le_ih);
 | |
| 
 | |
| 	return POSITION_NOT_FOUND;
 | |
| }
 | |
| 
 | |
| /* Compare given item and item pointed to by the path. */
 | |
| int comp_items(const struct item_head *stored_ih, const struct treepath *path)
 | |
| {
 | |
| 	struct buffer_head *bh = PATH_PLAST_BUFFER(path);
 | |
| 	struct item_head *ih;
 | |
| 
 | |
| 	/* Last buffer at the path is not in the tree. */
 | |
| 	if (!B_IS_IN_TREE(bh))
 | |
| 		return 1;
 | |
| 
 | |
| 	/* Last path position is invalid. */
 | |
| 	if (PATH_LAST_POSITION(path) >= B_NR_ITEMS(bh))
 | |
| 		return 1;
 | |
| 
 | |
| 	/* we need only to know, whether it is the same item */
 | |
| 	ih = get_ih(path);
 | |
| 	return memcmp(stored_ih, ih, IH_SIZE);
 | |
| }
 | |
| 
 | |
| /* unformatted nodes are not logged anymore, ever.  This is safe
 | |
| ** now
 | |
| */
 | |
| #define held_by_others(bh) (atomic_read(&(bh)->b_count) > 1)
 | |
| 
 | |
| // block can not be forgotten as it is in I/O or held by someone
 | |
| #define block_in_use(bh) (buffer_locked(bh) || (held_by_others(bh)))
 | |
| 
 | |
| // prepare for delete or cut of direct item
 | |
| static inline int prepare_for_direct_item(struct treepath *path,
 | |
| 					  struct item_head *le_ih,
 | |
| 					  struct inode *inode,
 | |
| 					  loff_t new_file_length, int *cut_size)
 | |
| {
 | |
| 	loff_t round_len;
 | |
| 
 | |
| 	if (new_file_length == max_reiserfs_offset(inode)) {
 | |
| 		/* item has to be deleted */
 | |
| 		*cut_size = -(IH_SIZE + ih_item_len(le_ih));
 | |
| 		return M_DELETE;
 | |
| 	}
 | |
| 	// new file gets truncated
 | |
| 	if (get_inode_item_key_version(inode) == KEY_FORMAT_3_6) {
 | |
| 		//
 | |
| 		round_len = ROUND_UP(new_file_length);
 | |
| 		/* this was new_file_length < le_ih ... */
 | |
| 		if (round_len < le_ih_k_offset(le_ih)) {
 | |
| 			*cut_size = -(IH_SIZE + ih_item_len(le_ih));
 | |
| 			return M_DELETE;	/* Delete this item. */
 | |
| 		}
 | |
| 		/* Calculate first position and size for cutting from item. */
 | |
| 		pos_in_item(path) = round_len - (le_ih_k_offset(le_ih) - 1);
 | |
| 		*cut_size = -(ih_item_len(le_ih) - pos_in_item(path));
 | |
| 
 | |
| 		return M_CUT;	/* Cut from this item. */
 | |
| 	}
 | |
| 
 | |
| 	// old file: items may have any length
 | |
| 
 | |
| 	if (new_file_length < le_ih_k_offset(le_ih)) {
 | |
| 		*cut_size = -(IH_SIZE + ih_item_len(le_ih));
 | |
| 		return M_DELETE;	/* Delete this item. */
 | |
| 	}
 | |
| 	/* Calculate first position and size for cutting from item. */
 | |
| 	*cut_size = -(ih_item_len(le_ih) -
 | |
| 		      (pos_in_item(path) =
 | |
| 		       new_file_length + 1 - le_ih_k_offset(le_ih)));
 | |
| 	return M_CUT;		/* Cut from this item. */
 | |
| }
 | |
| 
 | |
| static inline int prepare_for_direntry_item(struct treepath *path,
 | |
| 					    struct item_head *le_ih,
 | |
| 					    struct inode *inode,
 | |
| 					    loff_t new_file_length,
 | |
| 					    int *cut_size)
 | |
| {
 | |
| 	if (le_ih_k_offset(le_ih) == DOT_OFFSET &&
 | |
| 	    new_file_length == max_reiserfs_offset(inode)) {
 | |
| 		RFALSE(ih_entry_count(le_ih) != 2,
 | |
| 		       "PAP-5220: incorrect empty directory item (%h)", le_ih);
 | |
| 		*cut_size = -(IH_SIZE + ih_item_len(le_ih));
 | |
| 		return M_DELETE;	/* Delete the directory item containing "." and ".." entry. */
 | |
| 	}
 | |
| 
 | |
| 	if (ih_entry_count(le_ih) == 1) {
 | |
| 		/* Delete the directory item such as there is one record only
 | |
| 		   in this item */
 | |
| 		*cut_size = -(IH_SIZE + ih_item_len(le_ih));
 | |
| 		return M_DELETE;
 | |
| 	}
 | |
| 
 | |
| 	/* Cut one record from the directory item. */
 | |
| 	*cut_size =
 | |
| 	    -(DEH_SIZE +
 | |
| 	      entry_length(get_last_bh(path), le_ih, pos_in_item(path)));
 | |
| 	return M_CUT;
 | |
| }
 | |
| 
 | |
| #define JOURNAL_FOR_FREE_BLOCK_AND_UPDATE_SD (2 * JOURNAL_PER_BALANCE_CNT + 1)
 | |
| 
 | |
| /*  If the path points to a directory or direct item, calculate mode and the size cut, for balance.
 | |
|     If the path points to an indirect item, remove some number of its unformatted nodes.
 | |
|     In case of file truncate calculate whether this item must be deleted/truncated or last
 | |
|     unformatted node of this item will be converted to a direct item.
 | |
|     This function returns a determination of what balance mode the calling function should employ. */
 | |
| static char prepare_for_delete_or_cut(struct reiserfs_transaction_handle *th, struct inode *inode, struct treepath *path, const struct cpu_key *item_key, int *removed,	/* Number of unformatted nodes which were removed
 | |
| 																						   from end of the file. */
 | |
| 				      int *cut_size, unsigned long long new_file_length	/* MAX_KEY_OFFSET in case of delete. */
 | |
|     )
 | |
| {
 | |
| 	struct super_block *sb = inode->i_sb;
 | |
| 	struct item_head *p_le_ih = PATH_PITEM_HEAD(path);
 | |
| 	struct buffer_head *bh = PATH_PLAST_BUFFER(path);
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 
 | |
| 	/* Stat_data item. */
 | |
| 	if (is_statdata_le_ih(p_le_ih)) {
 | |
| 
 | |
| 		RFALSE(new_file_length != max_reiserfs_offset(inode),
 | |
| 		       "PAP-5210: mode must be M_DELETE");
 | |
| 
 | |
| 		*cut_size = -(IH_SIZE + ih_item_len(p_le_ih));
 | |
| 		return M_DELETE;
 | |
| 	}
 | |
| 
 | |
| 	/* Directory item. */
 | |
| 	if (is_direntry_le_ih(p_le_ih))
 | |
| 		return prepare_for_direntry_item(path, p_le_ih, inode,
 | |
| 						 new_file_length,
 | |
| 						 cut_size);
 | |
| 
 | |
| 	/* Direct item. */
 | |
| 	if (is_direct_le_ih(p_le_ih))
 | |
| 		return prepare_for_direct_item(path, p_le_ih, inode,
 | |
| 					       new_file_length, cut_size);
 | |
| 
 | |
| 	/* Case of an indirect item. */
 | |
| 	{
 | |
| 	    int blk_size = sb->s_blocksize;
 | |
| 	    struct item_head s_ih;
 | |
| 	    int need_re_search;
 | |
| 	    int delete = 0;
 | |
| 	    int result = M_CUT;
 | |
| 	    int pos = 0;
 | |
| 
 | |
| 	    if ( new_file_length == max_reiserfs_offset (inode) ) {
 | |
| 		/* prepare_for_delete_or_cut() is called by
 | |
| 		 * reiserfs_delete_item() */
 | |
| 		new_file_length = 0;
 | |
| 		delete = 1;
 | |
| 	    }
 | |
| 
 | |
| 	    do {
 | |
| 		need_re_search = 0;
 | |
| 		*cut_size = 0;
 | |
| 		bh = PATH_PLAST_BUFFER(path);
 | |
| 		copy_item_head(&s_ih, PATH_PITEM_HEAD(path));
 | |
| 		pos = I_UNFM_NUM(&s_ih);
 | |
| 
 | |
| 		while (le_ih_k_offset (&s_ih) + (pos - 1) * blk_size > new_file_length) {
 | |
| 		    __le32 *unfm;
 | |
| 		    __u32 block;
 | |
| 
 | |
| 		    /* Each unformatted block deletion may involve one additional
 | |
| 		     * bitmap block into the transaction, thereby the initial
 | |
| 		     * journal space reservation might not be enough. */
 | |
| 		    if (!delete && (*cut_size) != 0 &&
 | |
| 			reiserfs_transaction_free_space(th) < JOURNAL_FOR_FREE_BLOCK_AND_UPDATE_SD)
 | |
| 			break;
 | |
| 
 | |
| 		    unfm = (__le32 *)B_I_PITEM(bh, &s_ih) + pos - 1;
 | |
| 		    block = get_block_num(unfm, 0);
 | |
| 
 | |
| 		    if (block != 0) {
 | |
| 			reiserfs_prepare_for_journal(sb, bh, 1);
 | |
| 			put_block_num(unfm, 0, 0);
 | |
| 			journal_mark_dirty(th, sb, bh);
 | |
| 			reiserfs_free_block(th, inode, block, 1);
 | |
| 		    }
 | |
| 
 | |
| 		    reiserfs_write_unlock(sb);
 | |
| 		    cond_resched();
 | |
| 		    reiserfs_write_lock(sb);
 | |
| 
 | |
| 		    if (item_moved (&s_ih, path))  {
 | |
| 			need_re_search = 1;
 | |
| 			break;
 | |
| 		    }
 | |
| 
 | |
| 		    pos --;
 | |
| 		    (*removed)++;
 | |
| 		    (*cut_size) -= UNFM_P_SIZE;
 | |
| 
 | |
| 		    if (pos == 0) {
 | |
| 			(*cut_size) -= IH_SIZE;
 | |
| 			result = M_DELETE;
 | |
| 			break;
 | |
| 		    }
 | |
| 		}
 | |
| 		/* a trick.  If the buffer has been logged, this will do nothing.  If
 | |
| 		** we've broken the loop without logging it, it will restore the
 | |
| 		** buffer */
 | |
| 		reiserfs_restore_prepared_buffer(sb, bh);
 | |
| 	    } while (need_re_search &&
 | |
| 		     search_for_position_by_key(sb, item_key, path) == POSITION_FOUND);
 | |
| 	    pos_in_item(path) = pos * UNFM_P_SIZE;
 | |
| 
 | |
| 	    if (*cut_size == 0) {
 | |
| 		/* Nothing were cut. maybe convert last unformatted node to the
 | |
| 		 * direct item? */
 | |
| 		result = M_CONVERT;
 | |
| 	    }
 | |
| 	    return result;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /* Calculate number of bytes which will be deleted or cut during balance */
 | |
| static int calc_deleted_bytes_number(struct tree_balance *tb, char mode)
 | |
| {
 | |
| 	int del_size;
 | |
| 	struct item_head *p_le_ih = PATH_PITEM_HEAD(tb->tb_path);
 | |
| 
 | |
| 	if (is_statdata_le_ih(p_le_ih))
 | |
| 		return 0;
 | |
| 
 | |
| 	del_size =
 | |
| 	    (mode ==
 | |
| 	     M_DELETE) ? ih_item_len(p_le_ih) : -tb->insert_size[0];
 | |
| 	if (is_direntry_le_ih(p_le_ih)) {
 | |
| 		/* return EMPTY_DIR_SIZE; We delete emty directoris only.
 | |
| 		 * we can't use EMPTY_DIR_SIZE, as old format dirs have a different
 | |
| 		 * empty size.  ick. FIXME, is this right? */
 | |
| 		return del_size;
 | |
| 	}
 | |
| 
 | |
| 	if (is_indirect_le_ih(p_le_ih))
 | |
| 		del_size = (del_size / UNFM_P_SIZE) *
 | |
| 				(PATH_PLAST_BUFFER(tb->tb_path)->b_size);
 | |
| 	return del_size;
 | |
| }
 | |
| 
 | |
| static void init_tb_struct(struct reiserfs_transaction_handle *th,
 | |
| 			   struct tree_balance *tb,
 | |
| 			   struct super_block *sb,
 | |
| 			   struct treepath *path, int size)
 | |
| {
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 
 | |
| 	memset(tb, '\0', sizeof(struct tree_balance));
 | |
| 	tb->transaction_handle = th;
 | |
| 	tb->tb_sb = sb;
 | |
| 	tb->tb_path = path;
 | |
| 	PATH_OFFSET_PBUFFER(path, ILLEGAL_PATH_ELEMENT_OFFSET) = NULL;
 | |
| 	PATH_OFFSET_POSITION(path, ILLEGAL_PATH_ELEMENT_OFFSET) = 0;
 | |
| 	tb->insert_size[0] = size;
 | |
| }
 | |
| 
 | |
| void padd_item(char *item, int total_length, int length)
 | |
| {
 | |
| 	int i;
 | |
| 
 | |
| 	for (i = total_length; i > length;)
 | |
| 		item[--i] = 0;
 | |
| }
 | |
| 
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| char key2type(struct reiserfs_key *ih)
 | |
| {
 | |
| 	if (is_direntry_le_key(2, ih))
 | |
| 		return 'd';
 | |
| 	if (is_direct_le_key(2, ih))
 | |
| 		return 'D';
 | |
| 	if (is_indirect_le_key(2, ih))
 | |
| 		return 'i';
 | |
| 	if (is_statdata_le_key(2, ih))
 | |
| 		return 's';
 | |
| 	return 'u';
 | |
| }
 | |
| 
 | |
| char head2type(struct item_head *ih)
 | |
| {
 | |
| 	if (is_direntry_le_ih(ih))
 | |
| 		return 'd';
 | |
| 	if (is_direct_le_ih(ih))
 | |
| 		return 'D';
 | |
| 	if (is_indirect_le_ih(ih))
 | |
| 		return 'i';
 | |
| 	if (is_statdata_le_ih(ih))
 | |
| 		return 's';
 | |
| 	return 'u';
 | |
| }
 | |
| #endif
 | |
| 
 | |
| /* Delete object item.
 | |
|  * th       - active transaction handle
 | |
|  * path     - path to the deleted item
 | |
|  * item_key - key to search for the deleted item
 | |
|  * indode   - used for updating i_blocks and quotas
 | |
|  * un_bh    - NULL or unformatted node pointer
 | |
|  */
 | |
| int reiserfs_delete_item(struct reiserfs_transaction_handle *th,
 | |
| 			 struct treepath *path, const struct cpu_key *item_key,
 | |
| 			 struct inode *inode, struct buffer_head *un_bh)
 | |
| {
 | |
| 	struct super_block *sb = inode->i_sb;
 | |
| 	struct tree_balance s_del_balance;
 | |
| 	struct item_head s_ih;
 | |
| 	struct item_head *q_ih;
 | |
| 	int quota_cut_bytes;
 | |
| 	int ret_value, del_size, removed;
 | |
| 
 | |
| #ifdef CONFIG_REISERFS_CHECK
 | |
| 	char mode;
 | |
| 	int iter = 0;
 | |
| #endif
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 
 | |
| 	init_tb_struct(th, &s_del_balance, sb, path,
 | |
| 		       0 /*size is unknown */ );
 | |
| 
 | |
| 	while (1) {
 | |
| 		removed = 0;
 | |
| 
 | |
| #ifdef CONFIG_REISERFS_CHECK
 | |
| 		iter++;
 | |
| 		mode =
 | |
| #endif
 | |
| 		    prepare_for_delete_or_cut(th, inode, path,
 | |
| 					      item_key, &removed,
 | |
| 					      &del_size,
 | |
| 					      max_reiserfs_offset(inode));
 | |
| 
 | |
| 		RFALSE(mode != M_DELETE, "PAP-5320: mode must be M_DELETE");
 | |
| 
 | |
| 		copy_item_head(&s_ih, PATH_PITEM_HEAD(path));
 | |
| 		s_del_balance.insert_size[0] = del_size;
 | |
| 
 | |
| 		ret_value = fix_nodes(M_DELETE, &s_del_balance, NULL, NULL);
 | |
| 		if (ret_value != REPEAT_SEARCH)
 | |
| 			break;
 | |
| 
 | |
| 		PROC_INFO_INC(sb, delete_item_restarted);
 | |
| 
 | |
| 		// file system changed, repeat search
 | |
| 		ret_value =
 | |
| 		    search_for_position_by_key(sb, item_key, path);
 | |
| 		if (ret_value == IO_ERROR)
 | |
| 			break;
 | |
| 		if (ret_value == FILE_NOT_FOUND) {
 | |
| 			reiserfs_warning(sb, "vs-5340",
 | |
| 					 "no items of the file %K found",
 | |
| 					 item_key);
 | |
| 			break;
 | |
| 		}
 | |
| 	}			/* while (1) */
 | |
| 
 | |
| 	if (ret_value != CARRY_ON) {
 | |
| 		unfix_nodes(&s_del_balance);
 | |
| 		return 0;
 | |
| 	}
 | |
| 	// reiserfs_delete_item returns item length when success
 | |
| 	ret_value = calc_deleted_bytes_number(&s_del_balance, M_DELETE);
 | |
| 	q_ih = get_ih(path);
 | |
| 	quota_cut_bytes = ih_item_len(q_ih);
 | |
| 
 | |
| 	/* hack so the quota code doesn't have to guess if the file
 | |
| 	 ** has a tail.  On tail insert, we allocate quota for 1 unformatted node.
 | |
| 	 ** We test the offset because the tail might have been
 | |
| 	 ** split into multiple items, and we only want to decrement for
 | |
| 	 ** the unfm node once
 | |
| 	 */
 | |
| 	if (!S_ISLNK(inode->i_mode) && is_direct_le_ih(q_ih)) {
 | |
| 		if ((le_ih_k_offset(q_ih) & (sb->s_blocksize - 1)) == 1) {
 | |
| 			quota_cut_bytes = sb->s_blocksize + UNFM_P_SIZE;
 | |
| 		} else {
 | |
| 			quota_cut_bytes = 0;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if (un_bh) {
 | |
| 		int off;
 | |
| 		char *data;
 | |
| 
 | |
| 		/* We are in direct2indirect conversion, so move tail contents
 | |
| 		   to the unformatted node */
 | |
| 		/* note, we do the copy before preparing the buffer because we
 | |
| 		 ** don't care about the contents of the unformatted node yet.
 | |
| 		 ** the only thing we really care about is the direct item's data
 | |
| 		 ** is in the unformatted node.
 | |
| 		 **
 | |
| 		 ** Otherwise, we would have to call reiserfs_prepare_for_journal on
 | |
| 		 ** the unformatted node, which might schedule, meaning we'd have to
 | |
| 		 ** loop all the way back up to the start of the while loop.
 | |
| 		 **
 | |
| 		 ** The unformatted node must be dirtied later on.  We can't be
 | |
| 		 ** sure here if the entire tail has been deleted yet.
 | |
| 		 **
 | |
| 		 ** un_bh is from the page cache (all unformatted nodes are
 | |
| 		 ** from the page cache) and might be a highmem page.  So, we
 | |
| 		 ** can't use un_bh->b_data.
 | |
| 		 ** -clm
 | |
| 		 */
 | |
| 
 | |
| 		data = kmap_atomic(un_bh->b_page);
 | |
| 		off = ((le_ih_k_offset(&s_ih) - 1) & (PAGE_CACHE_SIZE - 1));
 | |
| 		memcpy(data + off,
 | |
| 		       B_I_PITEM(PATH_PLAST_BUFFER(path), &s_ih),
 | |
| 		       ret_value);
 | |
| 		kunmap_atomic(data);
 | |
| 	}
 | |
| 	/* Perform balancing after all resources have been collected at once. */
 | |
| 	do_balance(&s_del_balance, NULL, NULL, M_DELETE);
 | |
| 
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 	reiserfs_debug(sb, REISERFS_DEBUG_CODE,
 | |
| 		       "reiserquota delete_item(): freeing %u, id=%u type=%c",
 | |
| 		       quota_cut_bytes, inode->i_uid, head2type(&s_ih));
 | |
| #endif
 | |
| 	dquot_free_space_nodirty(inode, quota_cut_bytes);
 | |
| 
 | |
| 	/* Return deleted body length */
 | |
| 	return ret_value;
 | |
| }
 | |
| 
 | |
| /* Summary Of Mechanisms For Handling Collisions Between Processes:
 | |
| 
 | |
|  deletion of the body of the object is performed by iput(), with the
 | |
|  result that if multiple processes are operating on a file, the
 | |
|  deletion of the body of the file is deferred until the last process
 | |
|  that has an open inode performs its iput().
 | |
| 
 | |
|  writes and truncates are protected from collisions by use of
 | |
|  semaphores.
 | |
| 
 | |
|  creates, linking, and mknod are protected from collisions with other
 | |
|  processes by making the reiserfs_add_entry() the last step in the
 | |
|  creation, and then rolling back all changes if there was a collision.
 | |
|  - Hans
 | |
| */
 | |
| 
 | |
| /* this deletes item which never gets split */
 | |
| void reiserfs_delete_solid_item(struct reiserfs_transaction_handle *th,
 | |
| 				struct inode *inode, struct reiserfs_key *key)
 | |
| {
 | |
| 	struct tree_balance tb;
 | |
| 	INITIALIZE_PATH(path);
 | |
| 	int item_len = 0;
 | |
| 	int tb_init = 0;
 | |
| 	struct cpu_key cpu_key;
 | |
| 	int retval;
 | |
| 	int quota_cut_bytes = 0;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 
 | |
| 	le_key2cpu_key(&cpu_key, key);
 | |
| 
 | |
| 	while (1) {
 | |
| 		retval = search_item(th->t_super, &cpu_key, &path);
 | |
| 		if (retval == IO_ERROR) {
 | |
| 			reiserfs_error(th->t_super, "vs-5350",
 | |
| 				       "i/o failure occurred trying "
 | |
| 				       "to delete %K", &cpu_key);
 | |
| 			break;
 | |
| 		}
 | |
| 		if (retval != ITEM_FOUND) {
 | |
| 			pathrelse(&path);
 | |
| 			// No need for a warning, if there is just no free space to insert '..' item into the newly-created subdir
 | |
| 			if (!
 | |
| 			    ((unsigned long long)
 | |
| 			     GET_HASH_VALUE(le_key_k_offset
 | |
| 					    (le_key_version(key), key)) == 0
 | |
| 			     && (unsigned long long)
 | |
| 			     GET_GENERATION_NUMBER(le_key_k_offset
 | |
| 						   (le_key_version(key),
 | |
| 						    key)) == 1))
 | |
| 				reiserfs_warning(th->t_super, "vs-5355",
 | |
| 						 "%k not found", key);
 | |
| 			break;
 | |
| 		}
 | |
| 		if (!tb_init) {
 | |
| 			tb_init = 1;
 | |
| 			item_len = ih_item_len(PATH_PITEM_HEAD(&path));
 | |
| 			init_tb_struct(th, &tb, th->t_super, &path,
 | |
| 				       -(IH_SIZE + item_len));
 | |
| 		}
 | |
| 		quota_cut_bytes = ih_item_len(PATH_PITEM_HEAD(&path));
 | |
| 
 | |
| 		retval = fix_nodes(M_DELETE, &tb, NULL, NULL);
 | |
| 		if (retval == REPEAT_SEARCH) {
 | |
| 			PROC_INFO_INC(th->t_super, delete_solid_item_restarted);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		if (retval == CARRY_ON) {
 | |
| 			do_balance(&tb, NULL, NULL, M_DELETE);
 | |
| 			if (inode) {	/* Should we count quota for item? (we don't count quotas for save-links) */
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 				reiserfs_debug(th->t_super, REISERFS_DEBUG_CODE,
 | |
| 					       "reiserquota delete_solid_item(): freeing %u id=%u type=%c",
 | |
| 					       quota_cut_bytes, inode->i_uid,
 | |
| 					       key2type(key));
 | |
| #endif
 | |
| 				dquot_free_space_nodirty(inode,
 | |
| 							 quota_cut_bytes);
 | |
| 			}
 | |
| 			break;
 | |
| 		}
 | |
| 		// IO_ERROR, NO_DISK_SPACE, etc
 | |
| 		reiserfs_warning(th->t_super, "vs-5360",
 | |
| 				 "could not delete %K due to fix_nodes failure",
 | |
| 				 &cpu_key);
 | |
| 		unfix_nodes(&tb);
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	reiserfs_check_path(&path);
 | |
| }
 | |
| 
 | |
| int reiserfs_delete_object(struct reiserfs_transaction_handle *th,
 | |
| 			   struct inode *inode)
 | |
| {
 | |
| 	int err;
 | |
| 	inode->i_size = 0;
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 
 | |
| 	/* for directory this deletes item containing "." and ".." */
 | |
| 	err =
 | |
| 	    reiserfs_do_truncate(th, inode, NULL, 0 /*no timestamp updates */ );
 | |
| 	if (err)
 | |
| 		return err;
 | |
| 
 | |
| #if defined( USE_INODE_GENERATION_COUNTER )
 | |
| 	if (!old_format_only(th->t_super)) {
 | |
| 		__le32 *inode_generation;
 | |
| 
 | |
| 		inode_generation =
 | |
| 		    &REISERFS_SB(th->t_super)->s_rs->s_inode_generation;
 | |
| 		le32_add_cpu(inode_generation, 1);
 | |
| 	}
 | |
| /* USE_INODE_GENERATION_COUNTER */
 | |
| #endif
 | |
| 	reiserfs_delete_solid_item(th, inode, INODE_PKEY(inode));
 | |
| 
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| static void unmap_buffers(struct page *page, loff_t pos)
 | |
| {
 | |
| 	struct buffer_head *bh;
 | |
| 	struct buffer_head *head;
 | |
| 	struct buffer_head *next;
 | |
| 	unsigned long tail_index;
 | |
| 	unsigned long cur_index;
 | |
| 
 | |
| 	if (page) {
 | |
| 		if (page_has_buffers(page)) {
 | |
| 			tail_index = pos & (PAGE_CACHE_SIZE - 1);
 | |
| 			cur_index = 0;
 | |
| 			head = page_buffers(page);
 | |
| 			bh = head;
 | |
| 			do {
 | |
| 				next = bh->b_this_page;
 | |
| 
 | |
| 				/* we want to unmap the buffers that contain the tail, and
 | |
| 				 ** all the buffers after it (since the tail must be at the
 | |
| 				 ** end of the file).  We don't want to unmap file data
 | |
| 				 ** before the tail, since it might be dirty and waiting to
 | |
| 				 ** reach disk
 | |
| 				 */
 | |
| 				cur_index += bh->b_size;
 | |
| 				if (cur_index > tail_index) {
 | |
| 					reiserfs_unmap_buffer(bh);
 | |
| 				}
 | |
| 				bh = next;
 | |
| 			} while (bh != head);
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int maybe_indirect_to_direct(struct reiserfs_transaction_handle *th,
 | |
| 				    struct inode *inode,
 | |
| 				    struct page *page,
 | |
| 				    struct treepath *path,
 | |
| 				    const struct cpu_key *item_key,
 | |
| 				    loff_t new_file_size, char *mode)
 | |
| {
 | |
| 	struct super_block *sb = inode->i_sb;
 | |
| 	int block_size = sb->s_blocksize;
 | |
| 	int cut_bytes;
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 	BUG_ON(new_file_size != inode->i_size);
 | |
| 
 | |
| 	/* the page being sent in could be NULL if there was an i/o error
 | |
| 	 ** reading in the last block.  The user will hit problems trying to
 | |
| 	 ** read the file, but for now we just skip the indirect2direct
 | |
| 	 */
 | |
| 	if (atomic_read(&inode->i_count) > 1 ||
 | |
| 	    !tail_has_to_be_packed(inode) ||
 | |
| 	    !page || (REISERFS_I(inode)->i_flags & i_nopack_mask)) {
 | |
| 		/* leave tail in an unformatted node */
 | |
| 		*mode = M_SKIP_BALANCING;
 | |
| 		cut_bytes =
 | |
| 		    block_size - (new_file_size & (block_size - 1));
 | |
| 		pathrelse(path);
 | |
| 		return cut_bytes;
 | |
| 	}
 | |
| 	/* Perform the conversion to a direct_item. */
 | |
| 	/* return indirect_to_direct(inode, path, item_key,
 | |
| 				  new_file_size, mode); */
 | |
| 	return indirect2direct(th, inode, page, path, item_key,
 | |
| 			       new_file_size, mode);
 | |
| }
 | |
| 
 | |
| /* we did indirect_to_direct conversion. And we have inserted direct
 | |
|    item successesfully, but there were no disk space to cut unfm
 | |
|    pointer being converted. Therefore we have to delete inserted
 | |
|    direct item(s) */
 | |
| static void indirect_to_direct_roll_back(struct reiserfs_transaction_handle *th,
 | |
| 					 struct inode *inode, struct treepath *path)
 | |
| {
 | |
| 	struct cpu_key tail_key;
 | |
| 	int tail_len;
 | |
| 	int removed;
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 
 | |
| 	make_cpu_key(&tail_key, inode, inode->i_size + 1, TYPE_DIRECT, 4);	// !!!!
 | |
| 	tail_key.key_length = 4;
 | |
| 
 | |
| 	tail_len =
 | |
| 	    (cpu_key_k_offset(&tail_key) & (inode->i_sb->s_blocksize - 1)) - 1;
 | |
| 	while (tail_len) {
 | |
| 		/* look for the last byte of the tail */
 | |
| 		if (search_for_position_by_key(inode->i_sb, &tail_key, path) ==
 | |
| 		    POSITION_NOT_FOUND)
 | |
| 			reiserfs_panic(inode->i_sb, "vs-5615",
 | |
| 				       "found invalid item");
 | |
| 		RFALSE(path->pos_in_item !=
 | |
| 		       ih_item_len(PATH_PITEM_HEAD(path)) - 1,
 | |
| 		       "vs-5616: appended bytes found");
 | |
| 		PATH_LAST_POSITION(path)--;
 | |
| 
 | |
| 		removed =
 | |
| 		    reiserfs_delete_item(th, path, &tail_key, inode,
 | |
| 					 NULL /*unbh not needed */ );
 | |
| 		RFALSE(removed <= 0
 | |
| 		       || removed > tail_len,
 | |
| 		       "vs-5617: there was tail %d bytes, removed item length %d bytes",
 | |
| 		       tail_len, removed);
 | |
| 		tail_len -= removed;
 | |
| 		set_cpu_key_k_offset(&tail_key,
 | |
| 				     cpu_key_k_offset(&tail_key) - removed);
 | |
| 	}
 | |
| 	reiserfs_warning(inode->i_sb, "reiserfs-5091", "indirect_to_direct "
 | |
| 			 "conversion has been rolled back due to "
 | |
| 			 "lack of disk space");
 | |
| 	//mark_file_without_tail (inode);
 | |
| 	mark_inode_dirty(inode);
 | |
| }
 | |
| 
 | |
| /* (Truncate or cut entry) or delete object item. Returns < 0 on failure */
 | |
| int reiserfs_cut_from_item(struct reiserfs_transaction_handle *th,
 | |
| 			   struct treepath *path,
 | |
| 			   struct cpu_key *item_key,
 | |
| 			   struct inode *inode,
 | |
| 			   struct page *page, loff_t new_file_size)
 | |
| {
 | |
| 	struct super_block *sb = inode->i_sb;
 | |
| 	/* Every function which is going to call do_balance must first
 | |
| 	   create a tree_balance structure.  Then it must fill up this
 | |
| 	   structure by using the init_tb_struct and fix_nodes functions.
 | |
| 	   After that we can make tree balancing. */
 | |
| 	struct tree_balance s_cut_balance;
 | |
| 	struct item_head *p_le_ih;
 | |
| 	int cut_size = 0,	/* Amount to be cut. */
 | |
| 	    ret_value = CARRY_ON, removed = 0,	/* Number of the removed unformatted nodes. */
 | |
| 	    is_inode_locked = 0;
 | |
| 	char mode;		/* Mode of the balance. */
 | |
| 	int retval2 = -1;
 | |
| 	int quota_cut_bytes;
 | |
| 	loff_t tail_pos = 0;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 
 | |
| 	init_tb_struct(th, &s_cut_balance, inode->i_sb, path,
 | |
| 		       cut_size);
 | |
| 
 | |
| 	/* Repeat this loop until we either cut the item without needing
 | |
| 	   to balance, or we fix_nodes without schedule occurring */
 | |
| 	while (1) {
 | |
| 		/* Determine the balance mode, position of the first byte to
 | |
| 		   be cut, and size to be cut.  In case of the indirect item
 | |
| 		   free unformatted nodes which are pointed to by the cut
 | |
| 		   pointers. */
 | |
| 
 | |
| 		mode =
 | |
| 		    prepare_for_delete_or_cut(th, inode, path,
 | |
| 					      item_key, &removed,
 | |
| 					      &cut_size, new_file_size);
 | |
| 		if (mode == M_CONVERT) {
 | |
| 			/* convert last unformatted node to direct item or leave
 | |
| 			   tail in the unformatted node */
 | |
| 			RFALSE(ret_value != CARRY_ON,
 | |
| 			       "PAP-5570: can not convert twice");
 | |
| 
 | |
| 			ret_value =
 | |
| 			    maybe_indirect_to_direct(th, inode, page,
 | |
| 						     path, item_key,
 | |
| 						     new_file_size, &mode);
 | |
| 			if (mode == M_SKIP_BALANCING)
 | |
| 				/* tail has been left in the unformatted node */
 | |
| 				return ret_value;
 | |
| 
 | |
| 			is_inode_locked = 1;
 | |
| 
 | |
| 			/* removing of last unformatted node will change value we
 | |
| 			   have to return to truncate. Save it */
 | |
| 			retval2 = ret_value;
 | |
| 			/*retval2 = sb->s_blocksize - (new_file_size & (sb->s_blocksize - 1)); */
 | |
| 
 | |
| 			/* So, we have performed the first part of the conversion:
 | |
| 			   inserting the new direct item.  Now we are removing the
 | |
| 			   last unformatted node pointer. Set key to search for
 | |
| 			   it. */
 | |
| 			set_cpu_key_k_type(item_key, TYPE_INDIRECT);
 | |
| 			item_key->key_length = 4;
 | |
| 			new_file_size -=
 | |
| 			    (new_file_size & (sb->s_blocksize - 1));
 | |
| 			tail_pos = new_file_size;
 | |
| 			set_cpu_key_k_offset(item_key, new_file_size + 1);
 | |
| 			if (search_for_position_by_key
 | |
| 			    (sb, item_key,
 | |
| 			     path) == POSITION_NOT_FOUND) {
 | |
| 				print_block(PATH_PLAST_BUFFER(path), 3,
 | |
| 					    PATH_LAST_POSITION(path) - 1,
 | |
| 					    PATH_LAST_POSITION(path) + 1);
 | |
| 				reiserfs_panic(sb, "PAP-5580", "item to "
 | |
| 					       "convert does not exist (%K)",
 | |
| 					       item_key);
 | |
| 			}
 | |
| 			continue;
 | |
| 		}
 | |
| 		if (cut_size == 0) {
 | |
| 			pathrelse(path);
 | |
| 			return 0;
 | |
| 		}
 | |
| 
 | |
| 		s_cut_balance.insert_size[0] = cut_size;
 | |
| 
 | |
| 		ret_value = fix_nodes(mode, &s_cut_balance, NULL, NULL);
 | |
| 		if (ret_value != REPEAT_SEARCH)
 | |
| 			break;
 | |
| 
 | |
| 		PROC_INFO_INC(sb, cut_from_item_restarted);
 | |
| 
 | |
| 		ret_value =
 | |
| 		    search_for_position_by_key(sb, item_key, path);
 | |
| 		if (ret_value == POSITION_FOUND)
 | |
| 			continue;
 | |
| 
 | |
| 		reiserfs_warning(sb, "PAP-5610", "item %K not found",
 | |
| 				 item_key);
 | |
| 		unfix_nodes(&s_cut_balance);
 | |
| 		return (ret_value == IO_ERROR) ? -EIO : -ENOENT;
 | |
| 	}			/* while */
 | |
| 
 | |
| 	// check fix_nodes results (IO_ERROR or NO_DISK_SPACE)
 | |
| 	if (ret_value != CARRY_ON) {
 | |
| 		if (is_inode_locked) {
 | |
| 			// FIXME: this seems to be not needed: we are always able
 | |
| 			// to cut item
 | |
| 			indirect_to_direct_roll_back(th, inode, path);
 | |
| 		}
 | |
| 		if (ret_value == NO_DISK_SPACE)
 | |
| 			reiserfs_warning(sb, "reiserfs-5092",
 | |
| 					 "NO_DISK_SPACE");
 | |
| 		unfix_nodes(&s_cut_balance);
 | |
| 		return -EIO;
 | |
| 	}
 | |
| 
 | |
| 	/* go ahead and perform balancing */
 | |
| 
 | |
| 	RFALSE(mode == M_PASTE || mode == M_INSERT, "invalid mode");
 | |
| 
 | |
| 	/* Calculate number of bytes that need to be cut from the item. */
 | |
| 	quota_cut_bytes =
 | |
| 	    (mode ==
 | |
| 	     M_DELETE) ? ih_item_len(get_ih(path)) : -s_cut_balance.
 | |
| 	    insert_size[0];
 | |
| 	if (retval2 == -1)
 | |
| 		ret_value = calc_deleted_bytes_number(&s_cut_balance, mode);
 | |
| 	else
 | |
| 		ret_value = retval2;
 | |
| 
 | |
| 	/* For direct items, we only change the quota when deleting the last
 | |
| 	 ** item.
 | |
| 	 */
 | |
| 	p_le_ih = PATH_PITEM_HEAD(s_cut_balance.tb_path);
 | |
| 	if (!S_ISLNK(inode->i_mode) && is_direct_le_ih(p_le_ih)) {
 | |
| 		if (mode == M_DELETE &&
 | |
| 		    (le_ih_k_offset(p_le_ih) & (sb->s_blocksize - 1)) ==
 | |
| 		    1) {
 | |
| 			// FIXME: this is to keep 3.5 happy
 | |
| 			REISERFS_I(inode)->i_first_direct_byte = U32_MAX;
 | |
| 			quota_cut_bytes = sb->s_blocksize + UNFM_P_SIZE;
 | |
| 		} else {
 | |
| 			quota_cut_bytes = 0;
 | |
| 		}
 | |
| 	}
 | |
| #ifdef CONFIG_REISERFS_CHECK
 | |
| 	if (is_inode_locked) {
 | |
| 		struct item_head *le_ih =
 | |
| 		    PATH_PITEM_HEAD(s_cut_balance.tb_path);
 | |
| 		/* we are going to complete indirect2direct conversion. Make
 | |
| 		   sure, that we exactly remove last unformatted node pointer
 | |
| 		   of the item */
 | |
| 		if (!is_indirect_le_ih(le_ih))
 | |
| 			reiserfs_panic(sb, "vs-5652",
 | |
| 				       "item must be indirect %h", le_ih);
 | |
| 
 | |
| 		if (mode == M_DELETE && ih_item_len(le_ih) != UNFM_P_SIZE)
 | |
| 			reiserfs_panic(sb, "vs-5653", "completing "
 | |
| 				       "indirect2direct conversion indirect "
 | |
| 				       "item %h being deleted must be of "
 | |
| 				       "4 byte long", le_ih);
 | |
| 
 | |
| 		if (mode == M_CUT
 | |
| 		    && s_cut_balance.insert_size[0] != -UNFM_P_SIZE) {
 | |
| 			reiserfs_panic(sb, "vs-5654", "can not complete "
 | |
| 				       "indirect2direct conversion of %h "
 | |
| 				       "(CUT, insert_size==%d)",
 | |
| 				       le_ih, s_cut_balance.insert_size[0]);
 | |
| 		}
 | |
| 		/* it would be useful to make sure, that right neighboring
 | |
| 		   item is direct item of this file */
 | |
| 	}
 | |
| #endif
 | |
| 
 | |
| 	do_balance(&s_cut_balance, NULL, NULL, mode);
 | |
| 	if (is_inode_locked) {
 | |
| 		/* we've done an indirect->direct conversion.  when the data block
 | |
| 		 ** was freed, it was removed from the list of blocks that must
 | |
| 		 ** be flushed before the transaction commits, make sure to
 | |
| 		 ** unmap and invalidate it
 | |
| 		 */
 | |
| 		unmap_buffers(page, tail_pos);
 | |
| 		REISERFS_I(inode)->i_flags &= ~i_pack_on_close_mask;
 | |
| 	}
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 	reiserfs_debug(inode->i_sb, REISERFS_DEBUG_CODE,
 | |
| 		       "reiserquota cut_from_item(): freeing %u id=%u type=%c",
 | |
| 		       quota_cut_bytes, inode->i_uid, '?');
 | |
| #endif
 | |
| 	dquot_free_space_nodirty(inode, quota_cut_bytes);
 | |
| 	return ret_value;
 | |
| }
 | |
| 
 | |
| static void truncate_directory(struct reiserfs_transaction_handle *th,
 | |
| 			       struct inode *inode)
 | |
| {
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 	if (inode->i_nlink)
 | |
| 		reiserfs_error(inode->i_sb, "vs-5655", "link count != 0");
 | |
| 
 | |
| 	set_le_key_k_offset(KEY_FORMAT_3_5, INODE_PKEY(inode), DOT_OFFSET);
 | |
| 	set_le_key_k_type(KEY_FORMAT_3_5, INODE_PKEY(inode), TYPE_DIRENTRY);
 | |
| 	reiserfs_delete_solid_item(th, inode, INODE_PKEY(inode));
 | |
| 	reiserfs_update_sd(th, inode);
 | |
| 	set_le_key_k_offset(KEY_FORMAT_3_5, INODE_PKEY(inode), SD_OFFSET);
 | |
| 	set_le_key_k_type(KEY_FORMAT_3_5, INODE_PKEY(inode), TYPE_STAT_DATA);
 | |
| }
 | |
| 
 | |
| /* Truncate file to the new size. Note, this must be called with a transaction
 | |
|    already started */
 | |
| int reiserfs_do_truncate(struct reiserfs_transaction_handle *th,
 | |
| 			  struct inode *inode,	/* ->i_size contains new size */
 | |
| 			 struct page *page,	/* up to date for last block */
 | |
| 			 int update_timestamps	/* when it is called by
 | |
| 						   file_release to convert
 | |
| 						   the tail - no timestamps
 | |
| 						   should be updated */
 | |
|     )
 | |
| {
 | |
| 	INITIALIZE_PATH(s_search_path);	/* Path to the current object item. */
 | |
| 	struct item_head *p_le_ih;	/* Pointer to an item header. */
 | |
| 	struct cpu_key s_item_key;	/* Key to search for a previous file item. */
 | |
| 	loff_t file_size,	/* Old file size. */
 | |
| 	 new_file_size;	/* New file size. */
 | |
| 	int deleted;		/* Number of deleted or truncated bytes. */
 | |
| 	int retval;
 | |
| 	int err = 0;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 	if (!
 | |
| 	    (S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode)
 | |
| 	     || S_ISLNK(inode->i_mode)))
 | |
| 		return 0;
 | |
| 
 | |
| 	if (S_ISDIR(inode->i_mode)) {
 | |
| 		// deletion of directory - no need to update timestamps
 | |
| 		truncate_directory(th, inode);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	/* Get new file size. */
 | |
| 	new_file_size = inode->i_size;
 | |
| 
 | |
| 	// FIXME: note, that key type is unimportant here
 | |
| 	make_cpu_key(&s_item_key, inode, max_reiserfs_offset(inode),
 | |
| 		     TYPE_DIRECT, 3);
 | |
| 
 | |
| 	retval =
 | |
| 	    search_for_position_by_key(inode->i_sb, &s_item_key,
 | |
| 				       &s_search_path);
 | |
| 	if (retval == IO_ERROR) {
 | |
| 		reiserfs_error(inode->i_sb, "vs-5657",
 | |
| 			       "i/o failure occurred trying to truncate %K",
 | |
| 			       &s_item_key);
 | |
| 		err = -EIO;
 | |
| 		goto out;
 | |
| 	}
 | |
| 	if (retval == POSITION_FOUND || retval == FILE_NOT_FOUND) {
 | |
| 		reiserfs_error(inode->i_sb, "PAP-5660",
 | |
| 			       "wrong result %d of search for %K", retval,
 | |
| 			       &s_item_key);
 | |
| 
 | |
| 		err = -EIO;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	s_search_path.pos_in_item--;
 | |
| 
 | |
| 	/* Get real file size (total length of all file items) */
 | |
| 	p_le_ih = PATH_PITEM_HEAD(&s_search_path);
 | |
| 	if (is_statdata_le_ih(p_le_ih))
 | |
| 		file_size = 0;
 | |
| 	else {
 | |
| 		loff_t offset = le_ih_k_offset(p_le_ih);
 | |
| 		int bytes =
 | |
| 		    op_bytes_number(p_le_ih, inode->i_sb->s_blocksize);
 | |
| 
 | |
| 		/* this may mismatch with real file size: if last direct item
 | |
| 		   had no padding zeros and last unformatted node had no free
 | |
| 		   space, this file would have this file size */
 | |
| 		file_size = offset + bytes - 1;
 | |
| 	}
 | |
| 	/*
 | |
| 	 * are we doing a full truncate or delete, if so
 | |
| 	 * kick in the reada code
 | |
| 	 */
 | |
| 	if (new_file_size == 0)
 | |
| 		s_search_path.reada = PATH_READA | PATH_READA_BACK;
 | |
| 
 | |
| 	if (file_size == 0 || file_size < new_file_size) {
 | |
| 		goto update_and_out;
 | |
| 	}
 | |
| 
 | |
| 	/* Update key to search for the last file item. */
 | |
| 	set_cpu_key_k_offset(&s_item_key, file_size);
 | |
| 
 | |
| 	do {
 | |
| 		/* Cut or delete file item. */
 | |
| 		deleted =
 | |
| 		    reiserfs_cut_from_item(th, &s_search_path, &s_item_key,
 | |
| 					   inode, page, new_file_size);
 | |
| 		if (deleted < 0) {
 | |
| 			reiserfs_warning(inode->i_sb, "vs-5665",
 | |
| 					 "reiserfs_cut_from_item failed");
 | |
| 			reiserfs_check_path(&s_search_path);
 | |
| 			return 0;
 | |
| 		}
 | |
| 
 | |
| 		RFALSE(deleted > file_size,
 | |
| 		       "PAP-5670: reiserfs_cut_from_item: too many bytes deleted: deleted %d, file_size %lu, item_key %K",
 | |
| 		       deleted, file_size, &s_item_key);
 | |
| 
 | |
| 		/* Change key to search the last file item. */
 | |
| 		file_size -= deleted;
 | |
| 
 | |
| 		set_cpu_key_k_offset(&s_item_key, file_size);
 | |
| 
 | |
| 		/* While there are bytes to truncate and previous file item is presented in the tree. */
 | |
| 
 | |
| 		/*
 | |
| 		 ** This loop could take a really long time, and could log
 | |
| 		 ** many more blocks than a transaction can hold.  So, we do a polite
 | |
| 		 ** journal end here, and if the transaction needs ending, we make
 | |
| 		 ** sure the file is consistent before ending the current trans
 | |
| 		 ** and starting a new one
 | |
| 		 */
 | |
| 		if (journal_transaction_should_end(th, 0) ||
 | |
| 		    reiserfs_transaction_free_space(th) <= JOURNAL_FOR_FREE_BLOCK_AND_UPDATE_SD) {
 | |
| 			int orig_len_alloc = th->t_blocks_allocated;
 | |
| 			pathrelse(&s_search_path);
 | |
| 
 | |
| 			if (update_timestamps) {
 | |
| 				inode->i_mtime = CURRENT_TIME_SEC;
 | |
| 				inode->i_ctime = CURRENT_TIME_SEC;
 | |
| 			}
 | |
| 			reiserfs_update_sd(th, inode);
 | |
| 
 | |
| 			err = journal_end(th, inode->i_sb, orig_len_alloc);
 | |
| 			if (err)
 | |
| 				goto out;
 | |
| 			err = journal_begin(th, inode->i_sb,
 | |
| 					    JOURNAL_FOR_FREE_BLOCK_AND_UPDATE_SD + JOURNAL_PER_BALANCE_CNT * 4) ;
 | |
| 			if (err)
 | |
| 				goto out;
 | |
| 			reiserfs_update_inode_transaction(inode);
 | |
| 		}
 | |
| 	} while (file_size > ROUND_UP(new_file_size) &&
 | |
| 		 search_for_position_by_key(inode->i_sb, &s_item_key,
 | |
| 					    &s_search_path) == POSITION_FOUND);
 | |
| 
 | |
| 	RFALSE(file_size > ROUND_UP(new_file_size),
 | |
| 	       "PAP-5680: truncate did not finish: new_file_size %Ld, current %Ld, oid %d",
 | |
| 	       new_file_size, file_size, s_item_key.on_disk_key.k_objectid);
 | |
| 
 | |
|       update_and_out:
 | |
| 	if (update_timestamps) {
 | |
| 		// this is truncate, not file closing
 | |
| 		inode->i_mtime = CURRENT_TIME_SEC;
 | |
| 		inode->i_ctime = CURRENT_TIME_SEC;
 | |
| 	}
 | |
| 	reiserfs_update_sd(th, inode);
 | |
| 
 | |
|       out:
 | |
| 	pathrelse(&s_search_path);
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| #ifdef CONFIG_REISERFS_CHECK
 | |
| // this makes sure, that we __append__, not overwrite or add holes
 | |
| static void check_research_for_paste(struct treepath *path,
 | |
| 				     const struct cpu_key *key)
 | |
| {
 | |
| 	struct item_head *found_ih = get_ih(path);
 | |
| 
 | |
| 	if (is_direct_le_ih(found_ih)) {
 | |
| 		if (le_ih_k_offset(found_ih) +
 | |
| 		    op_bytes_number(found_ih,
 | |
| 				    get_last_bh(path)->b_size) !=
 | |
| 		    cpu_key_k_offset(key)
 | |
| 		    || op_bytes_number(found_ih,
 | |
| 				       get_last_bh(path)->b_size) !=
 | |
| 		    pos_in_item(path))
 | |
| 			reiserfs_panic(NULL, "PAP-5720", "found direct item "
 | |
| 				       "%h or position (%d) does not match "
 | |
| 				       "to key %K", found_ih,
 | |
| 				       pos_in_item(path), key);
 | |
| 	}
 | |
| 	if (is_indirect_le_ih(found_ih)) {
 | |
| 		if (le_ih_k_offset(found_ih) +
 | |
| 		    op_bytes_number(found_ih,
 | |
| 				    get_last_bh(path)->b_size) !=
 | |
| 		    cpu_key_k_offset(key)
 | |
| 		    || I_UNFM_NUM(found_ih) != pos_in_item(path)
 | |
| 		    || get_ih_free_space(found_ih) != 0)
 | |
| 			reiserfs_panic(NULL, "PAP-5730", "found indirect "
 | |
| 				       "item (%h) or position (%d) does not "
 | |
| 				       "match to key (%K)",
 | |
| 				       found_ih, pos_in_item(path), key);
 | |
| 	}
 | |
| }
 | |
| #endif				/* config reiserfs check */
 | |
| 
 | |
| /* Paste bytes to the existing item. Returns bytes number pasted into the item. */
 | |
| int reiserfs_paste_into_item(struct reiserfs_transaction_handle *th, struct treepath *search_path,	/* Path to the pasted item.	  */
 | |
| 			     const struct cpu_key *key,	/* Key to search for the needed item. */
 | |
| 			     struct inode *inode,	/* Inode item belongs to */
 | |
| 			     const char *body,	/* Pointer to the bytes to paste.    */
 | |
| 			     int pasted_size)
 | |
| {				/* Size of pasted bytes.             */
 | |
| 	struct tree_balance s_paste_balance;
 | |
| 	int retval;
 | |
| 	int fs_gen;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 
 | |
| 	fs_gen = get_generation(inode->i_sb);
 | |
| 
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 	reiserfs_debug(inode->i_sb, REISERFS_DEBUG_CODE,
 | |
| 		       "reiserquota paste_into_item(): allocating %u id=%u type=%c",
 | |
| 		       pasted_size, inode->i_uid,
 | |
| 		       key2type(&(key->on_disk_key)));
 | |
| #endif
 | |
| 
 | |
| 	reiserfs_write_unlock(inode->i_sb);
 | |
| 	retval = dquot_alloc_space_nodirty(inode, pasted_size);
 | |
| 	reiserfs_write_lock(inode->i_sb);
 | |
| 	if (retval) {
 | |
| 		pathrelse(search_path);
 | |
| 		return retval;
 | |
| 	}
 | |
| 	init_tb_struct(th, &s_paste_balance, th->t_super, search_path,
 | |
| 		       pasted_size);
 | |
| #ifdef DISPLACE_NEW_PACKING_LOCALITIES
 | |
| 	s_paste_balance.key = key->on_disk_key;
 | |
| #endif
 | |
| 
 | |
| 	/* DQUOT_* can schedule, must check before the fix_nodes */
 | |
| 	if (fs_changed(fs_gen, inode->i_sb)) {
 | |
| 		goto search_again;
 | |
| 	}
 | |
| 
 | |
| 	while ((retval =
 | |
| 		fix_nodes(M_PASTE, &s_paste_balance, NULL,
 | |
| 			  body)) == REPEAT_SEARCH) {
 | |
| 	      search_again:
 | |
| 		/* file system changed while we were in the fix_nodes */
 | |
| 		PROC_INFO_INC(th->t_super, paste_into_item_restarted);
 | |
| 		retval =
 | |
| 		    search_for_position_by_key(th->t_super, key,
 | |
| 					       search_path);
 | |
| 		if (retval == IO_ERROR) {
 | |
| 			retval = -EIO;
 | |
| 			goto error_out;
 | |
| 		}
 | |
| 		if (retval == POSITION_FOUND) {
 | |
| 			reiserfs_warning(inode->i_sb, "PAP-5710",
 | |
| 					 "entry or pasted byte (%K) exists",
 | |
| 					 key);
 | |
| 			retval = -EEXIST;
 | |
| 			goto error_out;
 | |
| 		}
 | |
| #ifdef CONFIG_REISERFS_CHECK
 | |
| 		check_research_for_paste(search_path, key);
 | |
| #endif
 | |
| 	}
 | |
| 
 | |
| 	/* Perform balancing after all resources are collected by fix_nodes, and
 | |
| 	   accessing them will not risk triggering schedule. */
 | |
| 	if (retval == CARRY_ON) {
 | |
| 		do_balance(&s_paste_balance, NULL /*ih */ , body, M_PASTE);
 | |
| 		return 0;
 | |
| 	}
 | |
| 	retval = (retval == NO_DISK_SPACE) ? -ENOSPC : -EIO;
 | |
|       error_out:
 | |
| 	/* this also releases the path */
 | |
| 	unfix_nodes(&s_paste_balance);
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 	reiserfs_debug(inode->i_sb, REISERFS_DEBUG_CODE,
 | |
| 		       "reiserquota paste_into_item(): freeing %u id=%u type=%c",
 | |
| 		       pasted_size, inode->i_uid,
 | |
| 		       key2type(&(key->on_disk_key)));
 | |
| #endif
 | |
| 	dquot_free_space_nodirty(inode, pasted_size);
 | |
| 	return retval;
 | |
| }
 | |
| 
 | |
| /* Insert new item into the buffer at the path.
 | |
|  * th   - active transaction handle
 | |
|  * path - path to the inserted item
 | |
|  * ih   - pointer to the item header to insert
 | |
|  * body - pointer to the bytes to insert
 | |
|  */
 | |
| int reiserfs_insert_item(struct reiserfs_transaction_handle *th,
 | |
| 			 struct treepath *path, const struct cpu_key *key,
 | |
| 			 struct item_head *ih, struct inode *inode,
 | |
| 			 const char *body)
 | |
| {
 | |
| 	struct tree_balance s_ins_balance;
 | |
| 	int retval;
 | |
| 	int fs_gen = 0;
 | |
| 	int quota_bytes = 0;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 
 | |
| 	if (inode) {		/* Do we count quotas for item? */
 | |
| 		fs_gen = get_generation(inode->i_sb);
 | |
| 		quota_bytes = ih_item_len(ih);
 | |
| 
 | |
| 		/* hack so the quota code doesn't have to guess if the file has
 | |
| 		 ** a tail, links are always tails, so there's no guessing needed
 | |
| 		 */
 | |
| 		if (!S_ISLNK(inode->i_mode) && is_direct_le_ih(ih))
 | |
| 			quota_bytes = inode->i_sb->s_blocksize + UNFM_P_SIZE;
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 		reiserfs_debug(inode->i_sb, REISERFS_DEBUG_CODE,
 | |
| 			       "reiserquota insert_item(): allocating %u id=%u type=%c",
 | |
| 			       quota_bytes, inode->i_uid, head2type(ih));
 | |
| #endif
 | |
| 		reiserfs_write_unlock(inode->i_sb);
 | |
| 		/* We can't dirty inode here. It would be immediately written but
 | |
| 		 * appropriate stat item isn't inserted yet... */
 | |
| 		retval = dquot_alloc_space_nodirty(inode, quota_bytes);
 | |
| 		reiserfs_write_lock(inode->i_sb);
 | |
| 		if (retval) {
 | |
| 			pathrelse(path);
 | |
| 			return retval;
 | |
| 		}
 | |
| 	}
 | |
| 	init_tb_struct(th, &s_ins_balance, th->t_super, path,
 | |
| 		       IH_SIZE + ih_item_len(ih));
 | |
| #ifdef DISPLACE_NEW_PACKING_LOCALITIES
 | |
| 	s_ins_balance.key = key->on_disk_key;
 | |
| #endif
 | |
| 	/* DQUOT_* can schedule, must check to be sure calling fix_nodes is safe */
 | |
| 	if (inode && fs_changed(fs_gen, inode->i_sb)) {
 | |
| 		goto search_again;
 | |
| 	}
 | |
| 
 | |
| 	while ((retval =
 | |
| 		fix_nodes(M_INSERT, &s_ins_balance, ih,
 | |
| 			  body)) == REPEAT_SEARCH) {
 | |
| 	      search_again:
 | |
| 		/* file system changed while we were in the fix_nodes */
 | |
| 		PROC_INFO_INC(th->t_super, insert_item_restarted);
 | |
| 		retval = search_item(th->t_super, key, path);
 | |
| 		if (retval == IO_ERROR) {
 | |
| 			retval = -EIO;
 | |
| 			goto error_out;
 | |
| 		}
 | |
| 		if (retval == ITEM_FOUND) {
 | |
| 			reiserfs_warning(th->t_super, "PAP-5760",
 | |
| 					 "key %K already exists in the tree",
 | |
| 					 key);
 | |
| 			retval = -EEXIST;
 | |
| 			goto error_out;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	/* make balancing after all resources will be collected at a time */
 | |
| 	if (retval == CARRY_ON) {
 | |
| 		do_balance(&s_ins_balance, ih, body, M_INSERT);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	retval = (retval == NO_DISK_SPACE) ? -ENOSPC : -EIO;
 | |
|       error_out:
 | |
| 	/* also releases the path */
 | |
| 	unfix_nodes(&s_ins_balance);
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 	reiserfs_debug(th->t_super, REISERFS_DEBUG_CODE,
 | |
| 		       "reiserquota insert_item(): freeing %u id=%u type=%c",
 | |
| 		       quota_bytes, inode->i_uid, head2type(ih));
 | |
| #endif
 | |
| 	if (inode)
 | |
| 		dquot_free_space_nodirty(inode, quota_bytes);
 | |
| 	return retval;
 | |
| }
 |