mirror of
				git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
				synced 2025-09-04 20:19:47 +08:00 
			
		
		
		
	 42bc47b353
			
		
	
	
		42bc47b353
		
	
	
	
	
		
			
			The vmalloc() function has no 2-factor argument form, so multiplication
factors need to be wrapped in array_size(). This patch replaces cases of:
        vmalloc(a * b)
with:
        vmalloc(array_size(a, b))
as well as handling cases of:
        vmalloc(a * b * c)
with:
        vmalloc(array3_size(a, b, c))
This does, however, attempt to ignore constant size factors like:
        vmalloc(4 * 1024)
though any constants defined via macros get caught up in the conversion.
Any factors with a sizeof() of "unsigned char", "char", and "u8" were
dropped, since they're redundant.
The Coccinelle script used for this was:
// Fix redundant parens around sizeof().
@@
type TYPE;
expression THING, E;
@@
(
  vmalloc(
-	(sizeof(TYPE)) * E
+	sizeof(TYPE) * E
  , ...)
|
  vmalloc(
-	(sizeof(THING)) * E
+	sizeof(THING) * E
  , ...)
)
// Drop single-byte sizes and redundant parens.
@@
expression COUNT;
typedef u8;
typedef __u8;
@@
(
  vmalloc(
-	sizeof(u8) * (COUNT)
+	COUNT
  , ...)
|
  vmalloc(
-	sizeof(__u8) * (COUNT)
+	COUNT
  , ...)
|
  vmalloc(
-	sizeof(char) * (COUNT)
+	COUNT
  , ...)
|
  vmalloc(
-	sizeof(unsigned char) * (COUNT)
+	COUNT
  , ...)
|
  vmalloc(
-	sizeof(u8) * COUNT
+	COUNT
  , ...)
|
  vmalloc(
-	sizeof(__u8) * COUNT
+	COUNT
  , ...)
|
  vmalloc(
-	sizeof(char) * COUNT
+	COUNT
  , ...)
|
  vmalloc(
-	sizeof(unsigned char) * COUNT
+	COUNT
  , ...)
)
// 2-factor product with sizeof(type/expression) and identifier or constant.
@@
type TYPE;
expression THING;
identifier COUNT_ID;
constant COUNT_CONST;
@@
(
  vmalloc(
-	sizeof(TYPE) * (COUNT_ID)
+	array_size(COUNT_ID, sizeof(TYPE))
  , ...)
|
  vmalloc(
-	sizeof(TYPE) * COUNT_ID
+	array_size(COUNT_ID, sizeof(TYPE))
  , ...)
|
  vmalloc(
-	sizeof(TYPE) * (COUNT_CONST)
+	array_size(COUNT_CONST, sizeof(TYPE))
  , ...)
|
  vmalloc(
-	sizeof(TYPE) * COUNT_CONST
+	array_size(COUNT_CONST, sizeof(TYPE))
  , ...)
|
  vmalloc(
-	sizeof(THING) * (COUNT_ID)
+	array_size(COUNT_ID, sizeof(THING))
  , ...)
|
  vmalloc(
-	sizeof(THING) * COUNT_ID
+	array_size(COUNT_ID, sizeof(THING))
  , ...)
|
  vmalloc(
-	sizeof(THING) * (COUNT_CONST)
+	array_size(COUNT_CONST, sizeof(THING))
  , ...)
|
  vmalloc(
-	sizeof(THING) * COUNT_CONST
+	array_size(COUNT_CONST, sizeof(THING))
  , ...)
)
// 2-factor product, only identifiers.
@@
identifier SIZE, COUNT;
@@
  vmalloc(
-	SIZE * COUNT
+	array_size(COUNT, SIZE)
  , ...)
// 3-factor product with 1 sizeof(type) or sizeof(expression), with
// redundant parens removed.
@@
expression THING;
identifier STRIDE, COUNT;
type TYPE;
@@
(
  vmalloc(
-	sizeof(TYPE) * (COUNT) * (STRIDE)
+	array3_size(COUNT, STRIDE, sizeof(TYPE))
  , ...)
|
  vmalloc(
-	sizeof(TYPE) * (COUNT) * STRIDE
+	array3_size(COUNT, STRIDE, sizeof(TYPE))
  , ...)
|
  vmalloc(
-	sizeof(TYPE) * COUNT * (STRIDE)
+	array3_size(COUNT, STRIDE, sizeof(TYPE))
  , ...)
|
  vmalloc(
-	sizeof(TYPE) * COUNT * STRIDE
+	array3_size(COUNT, STRIDE, sizeof(TYPE))
  , ...)
|
  vmalloc(
-	sizeof(THING) * (COUNT) * (STRIDE)
+	array3_size(COUNT, STRIDE, sizeof(THING))
  , ...)
|
  vmalloc(
-	sizeof(THING) * (COUNT) * STRIDE
+	array3_size(COUNT, STRIDE, sizeof(THING))
  , ...)
|
  vmalloc(
-	sizeof(THING) * COUNT * (STRIDE)
+	array3_size(COUNT, STRIDE, sizeof(THING))
  , ...)
|
  vmalloc(
-	sizeof(THING) * COUNT * STRIDE
+	array3_size(COUNT, STRIDE, sizeof(THING))
  , ...)
)
// 3-factor product with 2 sizeof(variable), with redundant parens removed.
@@
expression THING1, THING2;
identifier COUNT;
type TYPE1, TYPE2;
@@
(
  vmalloc(
-	sizeof(TYPE1) * sizeof(TYPE2) * COUNT
+	array3_size(COUNT, sizeof(TYPE1), sizeof(TYPE2))
  , ...)
|
  vmalloc(
-	sizeof(TYPE1) * sizeof(THING2) * (COUNT)
+	array3_size(COUNT, sizeof(TYPE1), sizeof(TYPE2))
  , ...)
|
  vmalloc(
-	sizeof(THING1) * sizeof(THING2) * COUNT
+	array3_size(COUNT, sizeof(THING1), sizeof(THING2))
  , ...)
|
  vmalloc(
-	sizeof(THING1) * sizeof(THING2) * (COUNT)
+	array3_size(COUNT, sizeof(THING1), sizeof(THING2))
  , ...)
|
  vmalloc(
-	sizeof(TYPE1) * sizeof(THING2) * COUNT
+	array3_size(COUNT, sizeof(TYPE1), sizeof(THING2))
  , ...)
|
  vmalloc(
-	sizeof(TYPE1) * sizeof(THING2) * (COUNT)
+	array3_size(COUNT, sizeof(TYPE1), sizeof(THING2))
  , ...)
)
// 3-factor product, only identifiers, with redundant parens removed.
@@
identifier STRIDE, SIZE, COUNT;
@@
(
  vmalloc(
-	(COUNT) * STRIDE * SIZE
+	array3_size(COUNT, STRIDE, SIZE)
  , ...)
|
  vmalloc(
-	COUNT * (STRIDE) * SIZE
+	array3_size(COUNT, STRIDE, SIZE)
  , ...)
|
  vmalloc(
-	COUNT * STRIDE * (SIZE)
+	array3_size(COUNT, STRIDE, SIZE)
  , ...)
|
  vmalloc(
-	(COUNT) * (STRIDE) * SIZE
+	array3_size(COUNT, STRIDE, SIZE)
  , ...)
|
  vmalloc(
-	COUNT * (STRIDE) * (SIZE)
+	array3_size(COUNT, STRIDE, SIZE)
  , ...)
|
  vmalloc(
-	(COUNT) * STRIDE * (SIZE)
+	array3_size(COUNT, STRIDE, SIZE)
  , ...)
|
  vmalloc(
-	(COUNT) * (STRIDE) * (SIZE)
+	array3_size(COUNT, STRIDE, SIZE)
  , ...)
|
  vmalloc(
-	COUNT * STRIDE * SIZE
+	array3_size(COUNT, STRIDE, SIZE)
  , ...)
)
// Any remaining multi-factor products, first at least 3-factor products
// when they're not all constants...
@@
expression E1, E2, E3;
constant C1, C2, C3;
@@
(
  vmalloc(C1 * C2 * C3, ...)
|
  vmalloc(
-	E1 * E2 * E3
+	array3_size(E1, E2, E3)
  , ...)
)
// And then all remaining 2 factors products when they're not all constants.
@@
expression E1, E2;
constant C1, C2;
@@
(
  vmalloc(C1 * C2, ...)
|
  vmalloc(
-	E1 * E2
+	array_size(E1, E2)
  , ...)
)
Signed-off-by: Kees Cook <keescook@chromium.org>
		
	
			
		
			
				
	
	
		
			1477 lines
		
	
	
		
			40 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1477 lines
		
	
	
		
			40 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright 2000 by Hans Reiser, licensing governed by reiserfs/README
 | |
|  */
 | |
| /* Reiserfs block (de)allocator, bitmap-based. */
 | |
| 
 | |
| #include <linux/time.h>
 | |
| #include "reiserfs.h"
 | |
| #include <linux/errno.h>
 | |
| #include <linux/buffer_head.h>
 | |
| #include <linux/kernel.h>
 | |
| #include <linux/pagemap.h>
 | |
| #include <linux/vmalloc.h>
 | |
| #include <linux/quotaops.h>
 | |
| #include <linux/seq_file.h>
 | |
| 
 | |
| #define PREALLOCATION_SIZE 9
 | |
| 
 | |
| /* different reiserfs block allocator options */
 | |
| 
 | |
| #define SB_ALLOC_OPTS(s) (REISERFS_SB(s)->s_alloc_options.bits)
 | |
| 
 | |
| #define  _ALLOC_concentrating_formatted_nodes 0
 | |
| #define  _ALLOC_displacing_large_files 1
 | |
| #define  _ALLOC_displacing_new_packing_localities 2
 | |
| #define  _ALLOC_old_hashed_relocation 3
 | |
| #define  _ALLOC_new_hashed_relocation 4
 | |
| #define  _ALLOC_skip_busy 5
 | |
| #define  _ALLOC_displace_based_on_dirid 6
 | |
| #define  _ALLOC_hashed_formatted_nodes 7
 | |
| #define  _ALLOC_old_way 8
 | |
| #define  _ALLOC_hundredth_slices 9
 | |
| #define  _ALLOC_dirid_groups 10
 | |
| #define  _ALLOC_oid_groups 11
 | |
| #define  _ALLOC_packing_groups 12
 | |
| 
 | |
| #define  concentrating_formatted_nodes(s)	test_bit(_ALLOC_concentrating_formatted_nodes, &SB_ALLOC_OPTS(s))
 | |
| #define  displacing_large_files(s)		test_bit(_ALLOC_displacing_large_files, &SB_ALLOC_OPTS(s))
 | |
| #define  displacing_new_packing_localities(s)	test_bit(_ALLOC_displacing_new_packing_localities, &SB_ALLOC_OPTS(s))
 | |
| 
 | |
| #define SET_OPTION(optname) \
 | |
|    do { \
 | |
| 	reiserfs_info(s, "block allocator option \"%s\" is set", #optname); \
 | |
| 	set_bit(_ALLOC_ ## optname , &SB_ALLOC_OPTS(s)); \
 | |
|     } while(0)
 | |
| #define TEST_OPTION(optname, s) \
 | |
|     test_bit(_ALLOC_ ## optname , &SB_ALLOC_OPTS(s))
 | |
| 
 | |
| static inline void get_bit_address(struct super_block *s,
 | |
| 				   b_blocknr_t block,
 | |
| 				   unsigned int *bmap_nr,
 | |
| 				   unsigned int *offset)
 | |
| {
 | |
| 	/*
 | |
| 	 * It is in the bitmap block number equal to the block
 | |
| 	 * number divided by the number of bits in a block.
 | |
| 	 */
 | |
| 	*bmap_nr = block >> (s->s_blocksize_bits + 3);
 | |
| 	/* Within that bitmap block it is located at bit offset *offset. */
 | |
| 	*offset = block & ((s->s_blocksize << 3) - 1);
 | |
| }
 | |
| 
 | |
| int is_reusable(struct super_block *s, b_blocknr_t block, int bit_value)
 | |
| {
 | |
| 	unsigned int bmap, offset;
 | |
| 	unsigned int bmap_count = reiserfs_bmap_count(s);
 | |
| 
 | |
| 	if (block == 0 || block >= SB_BLOCK_COUNT(s)) {
 | |
| 		reiserfs_error(s, "vs-4010",
 | |
| 			       "block number is out of range %lu (%u)",
 | |
| 			       block, SB_BLOCK_COUNT(s));
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	get_bit_address(s, block, &bmap, &offset);
 | |
| 
 | |
| 	/*
 | |
| 	 * Old format filesystem? Unlikely, but the bitmaps are all
 | |
| 	 * up front so we need to account for it.
 | |
| 	 */
 | |
| 	if (unlikely(test_bit(REISERFS_OLD_FORMAT,
 | |
| 			      &REISERFS_SB(s)->s_properties))) {
 | |
| 		b_blocknr_t bmap1 = REISERFS_SB(s)->s_sbh->b_blocknr + 1;
 | |
| 		if (block >= bmap1 &&
 | |
| 		    block <= bmap1 + bmap_count) {
 | |
| 			reiserfs_error(s, "vs-4019", "bitmap block %lu(%u) "
 | |
| 				       "can't be freed or reused",
 | |
| 				       block, bmap_count);
 | |
| 			return 0;
 | |
| 		}
 | |
| 	} else {
 | |
| 		if (offset == 0) {
 | |
| 			reiserfs_error(s, "vs-4020", "bitmap block %lu(%u) "
 | |
| 				       "can't be freed or reused",
 | |
| 				       block, bmap_count);
 | |
| 			return 0;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	if (bmap >= bmap_count) {
 | |
| 		reiserfs_error(s, "vs-4030", "bitmap for requested block "
 | |
| 			       "is out of range: block=%lu, bitmap_nr=%u",
 | |
| 			       block, bmap);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	if (bit_value == 0 && block == SB_ROOT_BLOCK(s)) {
 | |
| 		reiserfs_error(s, "vs-4050", "this is root block (%u), "
 | |
| 			       "it must be busy", SB_ROOT_BLOCK(s));
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Searches in journal structures for a given block number (bmap, off).
 | |
|  * If block is found in reiserfs journal it suggests next free block
 | |
|  * candidate to test.
 | |
|  */
 | |
| static inline int is_block_in_journal(struct super_block *s, unsigned int bmap,
 | |
| 				      int off, int *next)
 | |
| {
 | |
| 	b_blocknr_t tmp;
 | |
| 
 | |
| 	if (reiserfs_in_journal(s, bmap, off, 1, &tmp)) {
 | |
| 		if (tmp) {	/* hint supplied */
 | |
| 			*next = tmp;
 | |
| 			PROC_INFO_INC(s, scan_bitmap.in_journal_hint);
 | |
| 		} else {
 | |
| 			(*next) = off + 1;  /* inc offset to avoid looping. */
 | |
| 			PROC_INFO_INC(s, scan_bitmap.in_journal_nohint);
 | |
| 		}
 | |
| 		PROC_INFO_INC(s, scan_bitmap.retry);
 | |
| 		return 1;
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Searches for a window of zero bits with given minimum and maximum
 | |
|  * lengths in one bitmap block
 | |
|  */
 | |
| static int scan_bitmap_block(struct reiserfs_transaction_handle *th,
 | |
| 			     unsigned int bmap_n, int *beg, int boundary,
 | |
| 			     int min, int max, int unfm)
 | |
| {
 | |
| 	struct super_block *s = th->t_super;
 | |
| 	struct reiserfs_bitmap_info *bi = &SB_AP_BITMAP(s)[bmap_n];
 | |
| 	struct buffer_head *bh;
 | |
| 	int end, next;
 | |
| 	int org = *beg;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 	RFALSE(bmap_n >= reiserfs_bmap_count(s), "Bitmap %u is out of "
 | |
| 	       "range (0..%u)", bmap_n, reiserfs_bmap_count(s) - 1);
 | |
| 	PROC_INFO_INC(s, scan_bitmap.bmap);
 | |
| 
 | |
| 	if (!bi) {
 | |
| 		reiserfs_error(s, "jdm-4055", "NULL bitmap info pointer "
 | |
| 			       "for bitmap %d", bmap_n);
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	bh = reiserfs_read_bitmap_block(s, bmap_n);
 | |
| 	if (bh == NULL)
 | |
| 		return 0;
 | |
| 
 | |
| 	while (1) {
 | |
| cont:
 | |
| 		if (bi->free_count < min) {
 | |
| 			brelse(bh);
 | |
| 			return 0;	/* No free blocks in this bitmap */
 | |
| 		}
 | |
| 
 | |
| 		/* search for a first zero bit -- beginning of a window */
 | |
| 		*beg = reiserfs_find_next_zero_le_bit
 | |
| 		    ((unsigned long *)(bh->b_data), boundary, *beg);
 | |
| 
 | |
| 		/*
 | |
| 		 * search for a zero bit fails or the rest of bitmap block
 | |
| 		 * cannot contain a zero window of minimum size
 | |
| 		 */
 | |
| 		if (*beg + min > boundary) {
 | |
| 			brelse(bh);
 | |
| 			return 0;
 | |
| 		}
 | |
| 
 | |
| 		if (unfm && is_block_in_journal(s, bmap_n, *beg, beg))
 | |
| 			continue;
 | |
| 		/* first zero bit found; we check next bits */
 | |
| 		for (end = *beg + 1;; end++) {
 | |
| 			if (end >= *beg + max || end >= boundary
 | |
| 			    || reiserfs_test_le_bit(end, bh->b_data)) {
 | |
| 				next = end;
 | |
| 				break;
 | |
| 			}
 | |
| 
 | |
| 			/*
 | |
| 			 * finding the other end of zero bit window requires
 | |
| 			 * looking into journal structures (in case of
 | |
| 			 * searching for free blocks for unformatted nodes)
 | |
| 			 */
 | |
| 			if (unfm && is_block_in_journal(s, bmap_n, end, &next))
 | |
| 				break;
 | |
| 		}
 | |
| 
 | |
| 		/*
 | |
| 		 * now (*beg) points to beginning of zero bits window,
 | |
| 		 * (end) points to one bit after the window end
 | |
| 		 */
 | |
| 
 | |
| 		/* found window of proper size */
 | |
| 		if (end - *beg >= min) {
 | |
| 			int i;
 | |
| 			reiserfs_prepare_for_journal(s, bh, 1);
 | |
| 			/*
 | |
| 			 * try to set all blocks used checking are
 | |
| 			 * they still free
 | |
| 			 */
 | |
| 			for (i = *beg; i < end; i++) {
 | |
| 				/* Don't check in journal again. */
 | |
| 				if (reiserfs_test_and_set_le_bit
 | |
| 				    (i, bh->b_data)) {
 | |
| 					/*
 | |
| 					 * bit was set by another process while
 | |
| 					 * we slept in prepare_for_journal()
 | |
| 					 */
 | |
| 					PROC_INFO_INC(s, scan_bitmap.stolen);
 | |
| 
 | |
| 					/*
 | |
| 					 * we can continue with smaller set
 | |
| 					 * of allocated blocks, if length of
 | |
| 					 * this set is more or equal to `min'
 | |
| 					 */
 | |
| 					if (i >= *beg + min) {
 | |
| 						end = i;
 | |
| 						break;
 | |
| 					}
 | |
| 
 | |
| 					/*
 | |
| 					 * otherwise we clear all bit
 | |
| 					 * were set ...
 | |
| 					 */
 | |
| 					while (--i >= *beg)
 | |
| 						reiserfs_clear_le_bit
 | |
| 						    (i, bh->b_data);
 | |
| 					reiserfs_restore_prepared_buffer(s, bh);
 | |
| 					*beg = org;
 | |
| 
 | |
| 					/*
 | |
| 					 * Search again in current block
 | |
| 					 * from beginning
 | |
| 					 */
 | |
| 					goto cont;
 | |
| 				}
 | |
| 			}
 | |
| 			bi->free_count -= (end - *beg);
 | |
| 			journal_mark_dirty(th, bh);
 | |
| 			brelse(bh);
 | |
| 
 | |
| 			/* free block count calculation */
 | |
| 			reiserfs_prepare_for_journal(s, SB_BUFFER_WITH_SB(s),
 | |
| 						     1);
 | |
| 			PUT_SB_FREE_BLOCKS(s, SB_FREE_BLOCKS(s) - (end - *beg));
 | |
| 			journal_mark_dirty(th, SB_BUFFER_WITH_SB(s));
 | |
| 
 | |
| 			return end - (*beg);
 | |
| 		} else {
 | |
| 			*beg = next;
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static int bmap_hash_id(struct super_block *s, u32 id)
 | |
| {
 | |
| 	char *hash_in = NULL;
 | |
| 	unsigned long hash;
 | |
| 	unsigned bm;
 | |
| 
 | |
| 	if (id <= 2) {
 | |
| 		bm = 1;
 | |
| 	} else {
 | |
| 		hash_in = (char *)(&id);
 | |
| 		hash = keyed_hash(hash_in, 4);
 | |
| 		bm = hash % reiserfs_bmap_count(s);
 | |
| 		if (!bm)
 | |
| 			bm = 1;
 | |
| 	}
 | |
| 	/* this can only be true when SB_BMAP_NR = 1 */
 | |
| 	if (bm >= reiserfs_bmap_count(s))
 | |
| 		bm = 0;
 | |
| 	return bm;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * hashes the id and then returns > 0 if the block group for the
 | |
|  * corresponding hash is full
 | |
|  */
 | |
| static inline int block_group_used(struct super_block *s, u32 id)
 | |
| {
 | |
| 	int bm = bmap_hash_id(s, id);
 | |
| 	struct reiserfs_bitmap_info *info = &SB_AP_BITMAP(s)[bm];
 | |
| 
 | |
| 	/*
 | |
| 	 * If we don't have cached information on this bitmap block, we're
 | |
| 	 * going to have to load it later anyway. Loading it here allows us
 | |
| 	 * to make a better decision. This favors long-term performance gain
 | |
| 	 * with a better on-disk layout vs. a short term gain of skipping the
 | |
| 	 * read and potentially having a bad placement.
 | |
| 	 */
 | |
| 	if (info->free_count == UINT_MAX) {
 | |
| 		struct buffer_head *bh = reiserfs_read_bitmap_block(s, bm);
 | |
| 		brelse(bh);
 | |
| 	}
 | |
| 
 | |
| 	if (info->free_count > ((s->s_blocksize << 3) * 60 / 100)) {
 | |
| 		return 0;
 | |
| 	}
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * the packing is returned in disk byte order
 | |
|  */
 | |
| __le32 reiserfs_choose_packing(struct inode * dir)
 | |
| {
 | |
| 	__le32 packing;
 | |
| 	if (TEST_OPTION(packing_groups, dir->i_sb)) {
 | |
| 		u32 parent_dir = le32_to_cpu(INODE_PKEY(dir)->k_dir_id);
 | |
| 		/*
 | |
| 		 * some versions of reiserfsck expect packing locality 1 to be
 | |
| 		 * special
 | |
| 		 */
 | |
| 		if (parent_dir == 1 || block_group_used(dir->i_sb, parent_dir))
 | |
| 			packing = INODE_PKEY(dir)->k_objectid;
 | |
| 		else
 | |
| 			packing = INODE_PKEY(dir)->k_dir_id;
 | |
| 	} else
 | |
| 		packing = INODE_PKEY(dir)->k_objectid;
 | |
| 	return packing;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Tries to find contiguous zero bit window (given size) in given region of
 | |
|  * bitmap and place new blocks there. Returns number of allocated blocks.
 | |
|  */
 | |
| static int scan_bitmap(struct reiserfs_transaction_handle *th,
 | |
| 		       b_blocknr_t * start, b_blocknr_t finish,
 | |
| 		       int min, int max, int unfm, sector_t file_block)
 | |
| {
 | |
| 	int nr_allocated = 0;
 | |
| 	struct super_block *s = th->t_super;
 | |
| 	unsigned int bm, off;
 | |
| 	unsigned int end_bm, end_off;
 | |
| 	unsigned int off_max = s->s_blocksize << 3;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 	PROC_INFO_INC(s, scan_bitmap.call);
 | |
| 
 | |
| 	/* No point in looking for more free blocks */
 | |
| 	if (SB_FREE_BLOCKS(s) <= 0)
 | |
| 		return 0;
 | |
| 
 | |
| 	get_bit_address(s, *start, &bm, &off);
 | |
| 	get_bit_address(s, finish, &end_bm, &end_off);
 | |
| 	if (bm > reiserfs_bmap_count(s))
 | |
| 		return 0;
 | |
| 	if (end_bm > reiserfs_bmap_count(s))
 | |
| 		end_bm = reiserfs_bmap_count(s);
 | |
| 
 | |
| 	/*
 | |
| 	 * When the bitmap is more than 10% free, anyone can allocate.
 | |
| 	 * When it's less than 10% free, only files that already use the
 | |
| 	 * bitmap are allowed. Once we pass 80% full, this restriction
 | |
| 	 * is lifted.
 | |
| 	 *
 | |
| 	 * We do this so that files that grow later still have space close to
 | |
| 	 * their original allocation. This improves locality, and presumably
 | |
| 	 * performance as a result.
 | |
| 	 *
 | |
| 	 * This is only an allocation policy and does not make up for getting a
 | |
| 	 * bad hint. Decent hinting must be implemented for this to work well.
 | |
| 	 */
 | |
| 	if (TEST_OPTION(skip_busy, s)
 | |
| 	    && SB_FREE_BLOCKS(s) > SB_BLOCK_COUNT(s) / 20) {
 | |
| 		for (; bm < end_bm; bm++, off = 0) {
 | |
| 			if ((off && (!unfm || (file_block != 0)))
 | |
| 			    || SB_AP_BITMAP(s)[bm].free_count >
 | |
| 			    (s->s_blocksize << 3) / 10)
 | |
| 				nr_allocated =
 | |
| 				    scan_bitmap_block(th, bm, &off, off_max,
 | |
| 						      min, max, unfm);
 | |
| 			if (nr_allocated)
 | |
| 				goto ret;
 | |
| 		}
 | |
| 		/* we know from above that start is a reasonable number */
 | |
| 		get_bit_address(s, *start, &bm, &off);
 | |
| 	}
 | |
| 
 | |
| 	for (; bm < end_bm; bm++, off = 0) {
 | |
| 		nr_allocated =
 | |
| 		    scan_bitmap_block(th, bm, &off, off_max, min, max, unfm);
 | |
| 		if (nr_allocated)
 | |
| 			goto ret;
 | |
| 	}
 | |
| 
 | |
| 	nr_allocated =
 | |
| 	    scan_bitmap_block(th, bm, &off, end_off + 1, min, max, unfm);
 | |
| 
 | |
| ret:
 | |
| 	*start = bm * off_max + off;
 | |
| 	return nr_allocated;
 | |
| 
 | |
| }
 | |
| 
 | |
| static void _reiserfs_free_block(struct reiserfs_transaction_handle *th,
 | |
| 				 struct inode *inode, b_blocknr_t block,
 | |
| 				 int for_unformatted)
 | |
| {
 | |
| 	struct super_block *s = th->t_super;
 | |
| 	struct reiserfs_super_block *rs;
 | |
| 	struct buffer_head *sbh, *bmbh;
 | |
| 	struct reiserfs_bitmap_info *apbi;
 | |
| 	unsigned int nr, offset;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 	PROC_INFO_INC(s, free_block);
 | |
| 	rs = SB_DISK_SUPER_BLOCK(s);
 | |
| 	sbh = SB_BUFFER_WITH_SB(s);
 | |
| 	apbi = SB_AP_BITMAP(s);
 | |
| 
 | |
| 	get_bit_address(s, block, &nr, &offset);
 | |
| 
 | |
| 	if (nr >= reiserfs_bmap_count(s)) {
 | |
| 		reiserfs_error(s, "vs-4075", "block %lu is out of range",
 | |
| 			       block);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	bmbh = reiserfs_read_bitmap_block(s, nr);
 | |
| 	if (!bmbh)
 | |
| 		return;
 | |
| 
 | |
| 	reiserfs_prepare_for_journal(s, bmbh, 1);
 | |
| 
 | |
| 	/* clear bit for the given block in bit map */
 | |
| 	if (!reiserfs_test_and_clear_le_bit(offset, bmbh->b_data)) {
 | |
| 		reiserfs_error(s, "vs-4080",
 | |
| 			       "block %lu: bit already cleared", block);
 | |
| 	}
 | |
| 	apbi[nr].free_count++;
 | |
| 	journal_mark_dirty(th, bmbh);
 | |
| 	brelse(bmbh);
 | |
| 
 | |
| 	reiserfs_prepare_for_journal(s, sbh, 1);
 | |
| 	/* update super block */
 | |
| 	set_sb_free_blocks(rs, sb_free_blocks(rs) + 1);
 | |
| 
 | |
| 	journal_mark_dirty(th, sbh);
 | |
| 	if (for_unformatted) {
 | |
| 		int depth = reiserfs_write_unlock_nested(s);
 | |
| 		dquot_free_block_nodirty(inode, 1);
 | |
| 		reiserfs_write_lock_nested(s, depth);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| void reiserfs_free_block(struct reiserfs_transaction_handle *th,
 | |
| 			 struct inode *inode, b_blocknr_t block,
 | |
| 			 int for_unformatted)
 | |
| {
 | |
| 	struct super_block *s = th->t_super;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 	RFALSE(!s, "vs-4061: trying to free block on nonexistent device");
 | |
| 	if (!is_reusable(s, block, 1))
 | |
| 		return;
 | |
| 
 | |
| 	if (block > sb_block_count(REISERFS_SB(s)->s_rs)) {
 | |
| 		reiserfs_error(th->t_super, "bitmap-4072",
 | |
| 			       "Trying to free block outside file system "
 | |
| 			       "boundaries (%lu > %lu)",
 | |
| 			       block, sb_block_count(REISERFS_SB(s)->s_rs));
 | |
| 		return;
 | |
| 	}
 | |
| 	/* mark it before we clear it, just in case */
 | |
| 	journal_mark_freed(th, s, block);
 | |
| 	_reiserfs_free_block(th, inode, block, for_unformatted);
 | |
| }
 | |
| 
 | |
| /* preallocated blocks don't need to be run through journal_mark_freed */
 | |
| static void reiserfs_free_prealloc_block(struct reiserfs_transaction_handle *th,
 | |
| 					 struct inode *inode, b_blocknr_t block)
 | |
| {
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 	RFALSE(!th->t_super,
 | |
| 	       "vs-4060: trying to free block on nonexistent device");
 | |
| 	if (!is_reusable(th->t_super, block, 1))
 | |
| 		return;
 | |
| 	_reiserfs_free_block(th, inode, block, 1);
 | |
| }
 | |
| 
 | |
| static void __discard_prealloc(struct reiserfs_transaction_handle *th,
 | |
| 			       struct reiserfs_inode_info *ei)
 | |
| {
 | |
| 	unsigned long save = ei->i_prealloc_block;
 | |
| 	int dirty = 0;
 | |
| 	struct inode *inode = &ei->vfs_inode;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| #ifdef CONFIG_REISERFS_CHECK
 | |
| 	if (ei->i_prealloc_count < 0)
 | |
| 		reiserfs_error(th->t_super, "zam-4001",
 | |
| 			       "inode has negative prealloc blocks count.");
 | |
| #endif
 | |
| 	while (ei->i_prealloc_count > 0) {
 | |
| 		b_blocknr_t block_to_free;
 | |
| 
 | |
| 		/*
 | |
| 		 * reiserfs_free_prealloc_block can drop the write lock,
 | |
| 		 * which could allow another caller to free the same block.
 | |
| 		 * We can protect against it by modifying the prealloc
 | |
| 		 * state before calling it.
 | |
| 		 */
 | |
| 		block_to_free = ei->i_prealloc_block++;
 | |
| 		ei->i_prealloc_count--;
 | |
| 		reiserfs_free_prealloc_block(th, inode, block_to_free);
 | |
| 		dirty = 1;
 | |
| 	}
 | |
| 	if (dirty)
 | |
| 		reiserfs_update_sd(th, inode);
 | |
| 	ei->i_prealloc_block = save;
 | |
| 	list_del_init(&ei->i_prealloc_list);
 | |
| }
 | |
| 
 | |
| /* FIXME: It should be inline function */
 | |
| void reiserfs_discard_prealloc(struct reiserfs_transaction_handle *th,
 | |
| 			       struct inode *inode)
 | |
| {
 | |
| 	struct reiserfs_inode_info *ei = REISERFS_I(inode);
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 	if (ei->i_prealloc_count)
 | |
| 		__discard_prealloc(th, ei);
 | |
| }
 | |
| 
 | |
| void reiserfs_discard_all_prealloc(struct reiserfs_transaction_handle *th)
 | |
| {
 | |
| 	struct list_head *plist = &SB_JOURNAL(th->t_super)->j_prealloc_list;
 | |
| 
 | |
| 	BUG_ON(!th->t_trans_id);
 | |
| 	while (!list_empty(plist)) {
 | |
| 		struct reiserfs_inode_info *ei;
 | |
| 		ei = list_entry(plist->next, struct reiserfs_inode_info,
 | |
| 				i_prealloc_list);
 | |
| #ifdef CONFIG_REISERFS_CHECK
 | |
| 		if (!ei->i_prealloc_count) {
 | |
| 			reiserfs_error(th->t_super, "zam-4001",
 | |
| 				       "inode is in prealloc list but has "
 | |
| 				       "no preallocated blocks.");
 | |
| 		}
 | |
| #endif
 | |
| 		__discard_prealloc(th, ei);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| void reiserfs_init_alloc_options(struct super_block *s)
 | |
| {
 | |
| 	set_bit(_ALLOC_skip_busy, &SB_ALLOC_OPTS(s));
 | |
| 	set_bit(_ALLOC_dirid_groups, &SB_ALLOC_OPTS(s));
 | |
| 	set_bit(_ALLOC_packing_groups, &SB_ALLOC_OPTS(s));
 | |
| }
 | |
| 
 | |
| /* block allocator related options are parsed here */
 | |
| int reiserfs_parse_alloc_options(struct super_block *s, char *options)
 | |
| {
 | |
| 	char *this_char, *value;
 | |
| 
 | |
| 	/* clear default settings */
 | |
| 	REISERFS_SB(s)->s_alloc_options.bits = 0;
 | |
| 
 | |
| 	while ((this_char = strsep(&options, ":")) != NULL) {
 | |
| 		if ((value = strchr(this_char, '=')) != NULL)
 | |
| 			*value++ = 0;
 | |
| 
 | |
| 		if (!strcmp(this_char, "concentrating_formatted_nodes")) {
 | |
| 			int temp;
 | |
| 			SET_OPTION(concentrating_formatted_nodes);
 | |
| 			temp = (value
 | |
| 				&& *value) ? simple_strtoul(value, &value,
 | |
| 							    0) : 10;
 | |
| 			if (temp <= 0 || temp > 100) {
 | |
| 				REISERFS_SB(s)->s_alloc_options.border = 10;
 | |
| 			} else {
 | |
| 				REISERFS_SB(s)->s_alloc_options.border =
 | |
| 				    100 / temp;
 | |
| 			}
 | |
| 			continue;
 | |
| 		}
 | |
| 		if (!strcmp(this_char, "displacing_large_files")) {
 | |
| 			SET_OPTION(displacing_large_files);
 | |
| 			REISERFS_SB(s)->s_alloc_options.large_file_size =
 | |
| 			    (value
 | |
| 			     && *value) ? simple_strtoul(value, &value, 0) : 16;
 | |
| 			continue;
 | |
| 		}
 | |
| 		if (!strcmp(this_char, "displacing_new_packing_localities")) {
 | |
| 			SET_OPTION(displacing_new_packing_localities);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		if (!strcmp(this_char, "old_hashed_relocation")) {
 | |
| 			SET_OPTION(old_hashed_relocation);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		if (!strcmp(this_char, "new_hashed_relocation")) {
 | |
| 			SET_OPTION(new_hashed_relocation);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		if (!strcmp(this_char, "dirid_groups")) {
 | |
| 			SET_OPTION(dirid_groups);
 | |
| 			continue;
 | |
| 		}
 | |
| 		if (!strcmp(this_char, "oid_groups")) {
 | |
| 			SET_OPTION(oid_groups);
 | |
| 			continue;
 | |
| 		}
 | |
| 		if (!strcmp(this_char, "packing_groups")) {
 | |
| 			SET_OPTION(packing_groups);
 | |
| 			continue;
 | |
| 		}
 | |
| 		if (!strcmp(this_char, "hashed_formatted_nodes")) {
 | |
| 			SET_OPTION(hashed_formatted_nodes);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		if (!strcmp(this_char, "skip_busy")) {
 | |
| 			SET_OPTION(skip_busy);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		if (!strcmp(this_char, "hundredth_slices")) {
 | |
| 			SET_OPTION(hundredth_slices);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		if (!strcmp(this_char, "old_way")) {
 | |
| 			SET_OPTION(old_way);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		if (!strcmp(this_char, "displace_based_on_dirid")) {
 | |
| 			SET_OPTION(displace_based_on_dirid);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		if (!strcmp(this_char, "preallocmin")) {
 | |
| 			REISERFS_SB(s)->s_alloc_options.preallocmin =
 | |
| 			    (value
 | |
| 			     && *value) ? simple_strtoul(value, &value, 0) : 4;
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		if (!strcmp(this_char, "preallocsize")) {
 | |
| 			REISERFS_SB(s)->s_alloc_options.preallocsize =
 | |
| 			    (value
 | |
| 			     && *value) ? simple_strtoul(value, &value,
 | |
| 							 0) :
 | |
| 			    PREALLOCATION_SIZE;
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		reiserfs_warning(s, "zam-4001", "unknown option - %s",
 | |
| 				 this_char);
 | |
| 		return 1;
 | |
| 	}
 | |
| 
 | |
| 	reiserfs_info(s, "allocator options = [%08x]\n", SB_ALLOC_OPTS(s));
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void print_sep(struct seq_file *seq, int *first)
 | |
| {
 | |
| 	if (!*first)
 | |
| 		seq_puts(seq, ":");
 | |
| 	else
 | |
| 		*first = 0;
 | |
| }
 | |
| 
 | |
| void show_alloc_options(struct seq_file *seq, struct super_block *s)
 | |
| {
 | |
| 	int first = 1;
 | |
| 
 | |
| 	if (SB_ALLOC_OPTS(s) == ((1 << _ALLOC_skip_busy) |
 | |
| 		(1 << _ALLOC_dirid_groups) | (1 << _ALLOC_packing_groups)))
 | |
| 		return;
 | |
| 
 | |
| 	seq_puts(seq, ",alloc=");
 | |
| 
 | |
| 	if (TEST_OPTION(concentrating_formatted_nodes, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		if (REISERFS_SB(s)->s_alloc_options.border != 10) {
 | |
| 			seq_printf(seq, "concentrating_formatted_nodes=%d",
 | |
| 				100 / REISERFS_SB(s)->s_alloc_options.border);
 | |
| 		} else
 | |
| 			seq_puts(seq, "concentrating_formatted_nodes");
 | |
| 	}
 | |
| 	if (TEST_OPTION(displacing_large_files, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		if (REISERFS_SB(s)->s_alloc_options.large_file_size != 16) {
 | |
| 			seq_printf(seq, "displacing_large_files=%lu",
 | |
| 			    REISERFS_SB(s)->s_alloc_options.large_file_size);
 | |
| 		} else
 | |
| 			seq_puts(seq, "displacing_large_files");
 | |
| 	}
 | |
| 	if (TEST_OPTION(displacing_new_packing_localities, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "displacing_new_packing_localities");
 | |
| 	}
 | |
| 	if (TEST_OPTION(old_hashed_relocation, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "old_hashed_relocation");
 | |
| 	}
 | |
| 	if (TEST_OPTION(new_hashed_relocation, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "new_hashed_relocation");
 | |
| 	}
 | |
| 	if (TEST_OPTION(dirid_groups, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "dirid_groups");
 | |
| 	}
 | |
| 	if (TEST_OPTION(oid_groups, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "oid_groups");
 | |
| 	}
 | |
| 	if (TEST_OPTION(packing_groups, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "packing_groups");
 | |
| 	}
 | |
| 	if (TEST_OPTION(hashed_formatted_nodes, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "hashed_formatted_nodes");
 | |
| 	}
 | |
| 	if (TEST_OPTION(skip_busy, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "skip_busy");
 | |
| 	}
 | |
| 	if (TEST_OPTION(hundredth_slices, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "hundredth_slices");
 | |
| 	}
 | |
| 	if (TEST_OPTION(old_way, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "old_way");
 | |
| 	}
 | |
| 	if (TEST_OPTION(displace_based_on_dirid, s)) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_puts(seq, "displace_based_on_dirid");
 | |
| 	}
 | |
| 	if (REISERFS_SB(s)->s_alloc_options.preallocmin != 0) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_printf(seq, "preallocmin=%d",
 | |
| 				REISERFS_SB(s)->s_alloc_options.preallocmin);
 | |
| 	}
 | |
| 	if (REISERFS_SB(s)->s_alloc_options.preallocsize != 17) {
 | |
| 		print_sep(seq, &first);
 | |
| 		seq_printf(seq, "preallocsize=%d",
 | |
| 				REISERFS_SB(s)->s_alloc_options.preallocsize);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static inline void new_hashed_relocation(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	char *hash_in;
 | |
| 
 | |
| 	if (hint->formatted_node) {
 | |
| 		hash_in = (char *)&hint->key.k_dir_id;
 | |
| 	} else {
 | |
| 		if (!hint->inode) {
 | |
| 			/*hint->search_start = hint->beg;*/
 | |
| 			hash_in = (char *)&hint->key.k_dir_id;
 | |
| 		} else
 | |
| 		    if (TEST_OPTION(displace_based_on_dirid, hint->th->t_super))
 | |
| 			hash_in = (char *)(&INODE_PKEY(hint->inode)->k_dir_id);
 | |
| 		else
 | |
| 			hash_in =
 | |
| 			    (char *)(&INODE_PKEY(hint->inode)->k_objectid);
 | |
| 	}
 | |
| 
 | |
| 	hint->search_start =
 | |
| 	    hint->beg + keyed_hash(hash_in, 4) % (hint->end - hint->beg);
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Relocation based on dirid, hashing them into a given bitmap block
 | |
|  * files. Formatted nodes are unaffected, a separate policy covers them
 | |
|  */
 | |
| static void dirid_groups(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	unsigned long hash;
 | |
| 	__u32 dirid = 0;
 | |
| 	int bm = 0;
 | |
| 	struct super_block *sb = hint->th->t_super;
 | |
| 
 | |
| 	if (hint->inode)
 | |
| 		dirid = le32_to_cpu(INODE_PKEY(hint->inode)->k_dir_id);
 | |
| 	else if (hint->formatted_node)
 | |
| 		dirid = hint->key.k_dir_id;
 | |
| 
 | |
| 	if (dirid) {
 | |
| 		bm = bmap_hash_id(sb, dirid);
 | |
| 		hash = bm * (sb->s_blocksize << 3);
 | |
| 		/* give a portion of the block group to metadata */
 | |
| 		if (hint->inode)
 | |
| 			hash += sb->s_blocksize / 2;
 | |
| 		hint->search_start = hash;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * Relocation based on oid, hashing them into a given bitmap block
 | |
|  * files. Formatted nodes are unaffected, a separate policy covers them
 | |
|  */
 | |
| static void oid_groups(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	if (hint->inode) {
 | |
| 		unsigned long hash;
 | |
| 		__u32 oid;
 | |
| 		__u32 dirid;
 | |
| 		int bm;
 | |
| 
 | |
| 		dirid = le32_to_cpu(INODE_PKEY(hint->inode)->k_dir_id);
 | |
| 
 | |
| 		/*
 | |
| 		 * keep the root dir and it's first set of subdirs close to
 | |
| 		 * the start of the disk
 | |
| 		 */
 | |
| 		if (dirid <= 2)
 | |
| 			hash = (hint->inode->i_sb->s_blocksize << 3);
 | |
| 		else {
 | |
| 			oid = le32_to_cpu(INODE_PKEY(hint->inode)->k_objectid);
 | |
| 			bm = bmap_hash_id(hint->inode->i_sb, oid);
 | |
| 			hash = bm * (hint->inode->i_sb->s_blocksize << 3);
 | |
| 		}
 | |
| 		hint->search_start = hash;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * returns 1 if it finds an indirect item and gets valid hint info
 | |
|  * from it, otherwise 0
 | |
|  */
 | |
| static int get_left_neighbor(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	struct treepath *path;
 | |
| 	struct buffer_head *bh;
 | |
| 	struct item_head *ih;
 | |
| 	int pos_in_item;
 | |
| 	__le32 *item;
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	/*
 | |
| 	 * reiserfs code can call this function w/o pointer to path
 | |
| 	 * structure supplied; then we rely on supplied search_start
 | |
| 	 */
 | |
| 	if (!hint->path)
 | |
| 		return 0;
 | |
| 
 | |
| 	path = hint->path;
 | |
| 	bh = get_last_bh(path);
 | |
| 	RFALSE(!bh, "green-4002: Illegal path specified to get_left_neighbor");
 | |
| 	ih = tp_item_head(path);
 | |
| 	pos_in_item = path->pos_in_item;
 | |
| 	item = tp_item_body(path);
 | |
| 
 | |
| 	hint->search_start = bh->b_blocknr;
 | |
| 
 | |
| 	/*
 | |
| 	 * for indirect item: go to left and look for the first non-hole entry
 | |
| 	 * in the indirect item
 | |
| 	 */
 | |
| 	if (!hint->formatted_node && is_indirect_le_ih(ih)) {
 | |
| 		if (pos_in_item == I_UNFM_NUM(ih))
 | |
| 			pos_in_item--;
 | |
| 		while (pos_in_item >= 0) {
 | |
| 			int t = get_block_num(item, pos_in_item);
 | |
| 			if (t) {
 | |
| 				hint->search_start = t;
 | |
| 				ret = 1;
 | |
| 				break;
 | |
| 			}
 | |
| 			pos_in_item--;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	/* does result value fit into specified region? */
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| /*
 | |
|  * should be, if formatted node, then try to put on first part of the device
 | |
|  * specified as number of percent with mount option device, else try to put
 | |
|  * on last of device.  This is not to say it is good code to do so,
 | |
|  * but the effect should be measured.
 | |
|  */
 | |
| static inline void set_border_in_hint(struct super_block *s,
 | |
| 				      reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	b_blocknr_t border =
 | |
| 	    SB_BLOCK_COUNT(s) / REISERFS_SB(s)->s_alloc_options.border;
 | |
| 
 | |
| 	if (hint->formatted_node)
 | |
| 		hint->end = border - 1;
 | |
| 	else
 | |
| 		hint->beg = border;
 | |
| }
 | |
| 
 | |
| static inline void displace_large_file(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	if (TEST_OPTION(displace_based_on_dirid, hint->th->t_super))
 | |
| 		hint->search_start =
 | |
| 		    hint->beg +
 | |
| 		    keyed_hash((char *)(&INODE_PKEY(hint->inode)->k_dir_id),
 | |
| 			       4) % (hint->end - hint->beg);
 | |
| 	else
 | |
| 		hint->search_start =
 | |
| 		    hint->beg +
 | |
| 		    keyed_hash((char *)(&INODE_PKEY(hint->inode)->k_objectid),
 | |
| 			       4) % (hint->end - hint->beg);
 | |
| }
 | |
| 
 | |
| static inline void hash_formatted_node(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	char *hash_in;
 | |
| 
 | |
| 	if (!hint->inode)
 | |
| 		hash_in = (char *)&hint->key.k_dir_id;
 | |
| 	else if (TEST_OPTION(displace_based_on_dirid, hint->th->t_super))
 | |
| 		hash_in = (char *)(&INODE_PKEY(hint->inode)->k_dir_id);
 | |
| 	else
 | |
| 		hash_in = (char *)(&INODE_PKEY(hint->inode)->k_objectid);
 | |
| 
 | |
| 	hint->search_start =
 | |
| 	    hint->beg + keyed_hash(hash_in, 4) % (hint->end - hint->beg);
 | |
| }
 | |
| 
 | |
| static inline int
 | |
| this_blocknr_allocation_would_make_it_a_large_file(reiserfs_blocknr_hint_t *
 | |
| 						   hint)
 | |
| {
 | |
| 	return hint->block ==
 | |
| 	    REISERFS_SB(hint->th->t_super)->s_alloc_options.large_file_size;
 | |
| }
 | |
| 
 | |
| #ifdef DISPLACE_NEW_PACKING_LOCALITIES
 | |
| static inline void displace_new_packing_locality(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	struct in_core_key *key = &hint->key;
 | |
| 
 | |
| 	hint->th->displace_new_blocks = 0;
 | |
| 	hint->search_start =
 | |
| 	    hint->beg + keyed_hash((char *)(&key->k_objectid),
 | |
| 				   4) % (hint->end - hint->beg);
 | |
| }
 | |
| #endif
 | |
| 
 | |
| static inline int old_hashed_relocation(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	b_blocknr_t border;
 | |
| 	u32 hash_in;
 | |
| 
 | |
| 	if (hint->formatted_node || hint->inode == NULL) {
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	hash_in = le32_to_cpu((INODE_PKEY(hint->inode))->k_dir_id);
 | |
| 	border =
 | |
| 	    hint->beg + (u32) keyed_hash(((char *)(&hash_in)),
 | |
| 					 4) % (hint->end - hint->beg - 1);
 | |
| 	if (border > hint->search_start)
 | |
| 		hint->search_start = border;
 | |
| 
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| static inline int old_way(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	b_blocknr_t border;
 | |
| 
 | |
| 	if (hint->formatted_node || hint->inode == NULL) {
 | |
| 		return 0;
 | |
| 	}
 | |
| 
 | |
| 	border =
 | |
| 	    hint->beg +
 | |
| 	    le32_to_cpu(INODE_PKEY(hint->inode)->k_dir_id) % (hint->end -
 | |
| 							      hint->beg);
 | |
| 	if (border > hint->search_start)
 | |
| 		hint->search_start = border;
 | |
| 
 | |
| 	return 1;
 | |
| }
 | |
| 
 | |
| static inline void hundredth_slices(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	struct in_core_key *key = &hint->key;
 | |
| 	b_blocknr_t slice_start;
 | |
| 
 | |
| 	slice_start =
 | |
| 	    (keyed_hash((char *)(&key->k_dir_id), 4) % 100) * (hint->end / 100);
 | |
| 	if (slice_start > hint->search_start
 | |
| 	    || slice_start + (hint->end / 100) <= hint->search_start) {
 | |
| 		hint->search_start = slice_start;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static void determine_search_start(reiserfs_blocknr_hint_t * hint,
 | |
| 				   int amount_needed)
 | |
| {
 | |
| 	struct super_block *s = hint->th->t_super;
 | |
| 	int unfm_hint;
 | |
| 
 | |
| 	hint->beg = 0;
 | |
| 	hint->end = SB_BLOCK_COUNT(s) - 1;
 | |
| 
 | |
| 	/* This is former border algorithm. Now with tunable border offset */
 | |
| 	if (concentrating_formatted_nodes(s))
 | |
| 		set_border_in_hint(s, hint);
 | |
| 
 | |
| #ifdef DISPLACE_NEW_PACKING_LOCALITIES
 | |
| 	/*
 | |
| 	 * whenever we create a new directory, we displace it.  At first
 | |
| 	 * we will hash for location, later we might look for a moderately
 | |
| 	 * empty place for it
 | |
| 	 */
 | |
| 	if (displacing_new_packing_localities(s)
 | |
| 	    && hint->th->displace_new_blocks) {
 | |
| 		displace_new_packing_locality(hint);
 | |
| 
 | |
| 		/*
 | |
| 		 * we do not continue determine_search_start,
 | |
| 		 * if new packing locality is being displaced
 | |
| 		 */
 | |
| 		return;
 | |
| 	}
 | |
| #endif
 | |
| 
 | |
| 	/*
 | |
| 	 * all persons should feel encouraged to add more special cases
 | |
| 	 * here and test them
 | |
| 	 */
 | |
| 
 | |
| 	if (displacing_large_files(s) && !hint->formatted_node
 | |
| 	    && this_blocknr_allocation_would_make_it_a_large_file(hint)) {
 | |
| 		displace_large_file(hint);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	/*
 | |
| 	 * if none of our special cases is relevant, use the left
 | |
| 	 * neighbor in the tree order of the new node we are allocating for
 | |
| 	 */
 | |
| 	if (hint->formatted_node && TEST_OPTION(hashed_formatted_nodes, s)) {
 | |
| 		hash_formatted_node(hint);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	unfm_hint = get_left_neighbor(hint);
 | |
| 
 | |
| 	/*
 | |
| 	 * Mimic old block allocator behaviour, that is if VFS allowed for
 | |
| 	 * preallocation, new blocks are displaced based on directory ID.
 | |
| 	 * Also, if suggested search_start is less than last preallocated
 | |
| 	 * block, we start searching from it, assuming that HDD dataflow
 | |
| 	 * is faster in forward direction
 | |
| 	 */
 | |
| 	if (TEST_OPTION(old_way, s)) {
 | |
| 		if (!hint->formatted_node) {
 | |
| 			if (!reiserfs_hashed_relocation(s))
 | |
| 				old_way(hint);
 | |
| 			else if (!reiserfs_no_unhashed_relocation(s))
 | |
| 				old_hashed_relocation(hint);
 | |
| 
 | |
| 			if (hint->inode
 | |
| 			    && hint->search_start <
 | |
| 			    REISERFS_I(hint->inode)->i_prealloc_block)
 | |
| 				hint->search_start =
 | |
| 				    REISERFS_I(hint->inode)->i_prealloc_block;
 | |
| 		}
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	/* This is an approach proposed by Hans */
 | |
| 	if (TEST_OPTION(hundredth_slices, s)
 | |
| 	    && !(displacing_large_files(s) && !hint->formatted_node)) {
 | |
| 		hundredth_slices(hint);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	/* old_hashed_relocation only works on unformatted */
 | |
| 	if (!unfm_hint && !hint->formatted_node &&
 | |
| 	    TEST_OPTION(old_hashed_relocation, s)) {
 | |
| 		old_hashed_relocation(hint);
 | |
| 	}
 | |
| 
 | |
| 	/* new_hashed_relocation works with both formatted/unformatted nodes */
 | |
| 	if ((!unfm_hint || hint->formatted_node) &&
 | |
| 	    TEST_OPTION(new_hashed_relocation, s)) {
 | |
| 		new_hashed_relocation(hint);
 | |
| 	}
 | |
| 
 | |
| 	/* dirid grouping works only on unformatted nodes */
 | |
| 	if (!unfm_hint && !hint->formatted_node && TEST_OPTION(dirid_groups, s)) {
 | |
| 		dirid_groups(hint);
 | |
| 	}
 | |
| #ifdef DISPLACE_NEW_PACKING_LOCALITIES
 | |
| 	if (hint->formatted_node && TEST_OPTION(dirid_groups, s)) {
 | |
| 		dirid_groups(hint);
 | |
| 	}
 | |
| #endif
 | |
| 
 | |
| 	/* oid grouping works only on unformatted nodes */
 | |
| 	if (!unfm_hint && !hint->formatted_node && TEST_OPTION(oid_groups, s)) {
 | |
| 		oid_groups(hint);
 | |
| 	}
 | |
| 	return;
 | |
| }
 | |
| 
 | |
| static int determine_prealloc_size(reiserfs_blocknr_hint_t * hint)
 | |
| {
 | |
| 	/* make minimum size a mount option and benchmark both ways */
 | |
| 	/* we preallocate blocks only for regular files, specific size */
 | |
| 	/* benchmark preallocating always and see what happens */
 | |
| 
 | |
| 	hint->prealloc_size = 0;
 | |
| 
 | |
| 	if (!hint->formatted_node && hint->preallocate) {
 | |
| 		if (S_ISREG(hint->inode->i_mode) && !IS_PRIVATE(hint->inode)
 | |
| 		    && hint->inode->i_size >=
 | |
| 		    REISERFS_SB(hint->th->t_super)->s_alloc_options.
 | |
| 		    preallocmin * hint->inode->i_sb->s_blocksize)
 | |
| 			hint->prealloc_size =
 | |
| 			    REISERFS_SB(hint->th->t_super)->s_alloc_options.
 | |
| 			    preallocsize - 1;
 | |
| 	}
 | |
| 	return CARRY_ON;
 | |
| }
 | |
| 
 | |
| static inline int allocate_without_wrapping_disk(reiserfs_blocknr_hint_t * hint,
 | |
| 						 b_blocknr_t * new_blocknrs,
 | |
| 						 b_blocknr_t start,
 | |
| 						 b_blocknr_t finish, int min,
 | |
| 						 int amount_needed,
 | |
| 						 int prealloc_size)
 | |
| {
 | |
| 	int rest = amount_needed;
 | |
| 	int nr_allocated;
 | |
| 
 | |
| 	while (rest > 0 && start <= finish) {
 | |
| 		nr_allocated = scan_bitmap(hint->th, &start, finish, min,
 | |
| 					   rest + prealloc_size,
 | |
| 					   !hint->formatted_node, hint->block);
 | |
| 
 | |
| 		if (nr_allocated == 0)	/* no new blocks allocated, return */
 | |
| 			break;
 | |
| 
 | |
| 		/* fill free_blocknrs array first */
 | |
| 		while (rest > 0 && nr_allocated > 0) {
 | |
| 			*new_blocknrs++ = start++;
 | |
| 			rest--;
 | |
| 			nr_allocated--;
 | |
| 		}
 | |
| 
 | |
| 		/* do we have something to fill prealloc. array also ? */
 | |
| 		if (nr_allocated > 0) {
 | |
| 			/*
 | |
| 			 * it means prealloc_size was greater that 0 and
 | |
| 			 * we do preallocation
 | |
| 			 */
 | |
| 			list_add(&REISERFS_I(hint->inode)->i_prealloc_list,
 | |
| 				 &SB_JOURNAL(hint->th->t_super)->
 | |
| 				 j_prealloc_list);
 | |
| 			REISERFS_I(hint->inode)->i_prealloc_block = start;
 | |
| 			REISERFS_I(hint->inode)->i_prealloc_count =
 | |
| 			    nr_allocated;
 | |
| 			break;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return (amount_needed - rest);
 | |
| }
 | |
| 
 | |
| static inline int blocknrs_and_prealloc_arrays_from_search_start
 | |
|     (reiserfs_blocknr_hint_t * hint, b_blocknr_t * new_blocknrs,
 | |
|      int amount_needed) {
 | |
| 	struct super_block *s = hint->th->t_super;
 | |
| 	b_blocknr_t start = hint->search_start;
 | |
| 	b_blocknr_t finish = SB_BLOCK_COUNT(s) - 1;
 | |
| 	int passno = 0;
 | |
| 	int nr_allocated = 0;
 | |
| 	int depth;
 | |
| 
 | |
| 	determine_prealloc_size(hint);
 | |
| 	if (!hint->formatted_node) {
 | |
| 		int quota_ret;
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 		reiserfs_debug(s, REISERFS_DEBUG_CODE,
 | |
| 			       "reiserquota: allocating %d blocks id=%u",
 | |
| 			       amount_needed, hint->inode->i_uid);
 | |
| #endif
 | |
| 		depth = reiserfs_write_unlock_nested(s);
 | |
| 		quota_ret =
 | |
| 		    dquot_alloc_block_nodirty(hint->inode, amount_needed);
 | |
| 		if (quota_ret) {	/* Quota exceeded? */
 | |
| 			reiserfs_write_lock_nested(s, depth);
 | |
| 			return QUOTA_EXCEEDED;
 | |
| 		}
 | |
| 		if (hint->preallocate && hint->prealloc_size) {
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 			reiserfs_debug(s, REISERFS_DEBUG_CODE,
 | |
| 				       "reiserquota: allocating (prealloc) %d blocks id=%u",
 | |
| 				       hint->prealloc_size, hint->inode->i_uid);
 | |
| #endif
 | |
| 			quota_ret = dquot_prealloc_block_nodirty(hint->inode,
 | |
| 							 hint->prealloc_size);
 | |
| 			if (quota_ret)
 | |
| 				hint->preallocate = hint->prealloc_size = 0;
 | |
| 		}
 | |
| 		/* for unformatted nodes, force large allocations */
 | |
| 		reiserfs_write_lock_nested(s, depth);
 | |
| 	}
 | |
| 
 | |
| 	do {
 | |
| 		switch (passno++) {
 | |
| 		case 0:	/* Search from hint->search_start to end of disk */
 | |
| 			start = hint->search_start;
 | |
| 			finish = SB_BLOCK_COUNT(s) - 1;
 | |
| 			break;
 | |
| 		case 1:	/* Search from hint->beg to hint->search_start */
 | |
| 			start = hint->beg;
 | |
| 			finish = hint->search_start;
 | |
| 			break;
 | |
| 		case 2:	/* Last chance: Search from 0 to hint->beg */
 | |
| 			start = 0;
 | |
| 			finish = hint->beg;
 | |
| 			break;
 | |
| 		default:
 | |
| 			/* We've tried searching everywhere, not enough space */
 | |
| 			/* Free the blocks */
 | |
| 			if (!hint->formatted_node) {
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 				reiserfs_debug(s, REISERFS_DEBUG_CODE,
 | |
| 					       "reiserquota: freeing (nospace) %d blocks id=%u",
 | |
| 					       amount_needed +
 | |
| 					       hint->prealloc_size -
 | |
| 					       nr_allocated,
 | |
| 					       hint->inode->i_uid);
 | |
| #endif
 | |
| 				/* Free not allocated blocks */
 | |
| 				depth = reiserfs_write_unlock_nested(s);
 | |
| 				dquot_free_block_nodirty(hint->inode,
 | |
| 					amount_needed + hint->prealloc_size -
 | |
| 					nr_allocated);
 | |
| 				reiserfs_write_lock_nested(s, depth);
 | |
| 			}
 | |
| 			while (nr_allocated--)
 | |
| 				reiserfs_free_block(hint->th, hint->inode,
 | |
| 						    new_blocknrs[nr_allocated],
 | |
| 						    !hint->formatted_node);
 | |
| 
 | |
| 			return NO_DISK_SPACE;
 | |
| 		}
 | |
| 	} while ((nr_allocated += allocate_without_wrapping_disk(hint,
 | |
| 								 new_blocknrs +
 | |
| 								 nr_allocated,
 | |
| 								 start, finish,
 | |
| 								 1,
 | |
| 								 amount_needed -
 | |
| 								 nr_allocated,
 | |
| 								 hint->
 | |
| 								 prealloc_size))
 | |
| 		 < amount_needed);
 | |
| 	if (!hint->formatted_node &&
 | |
| 	    amount_needed + hint->prealloc_size >
 | |
| 	    nr_allocated + REISERFS_I(hint->inode)->i_prealloc_count) {
 | |
| 		/* Some of preallocation blocks were not allocated */
 | |
| #ifdef REISERQUOTA_DEBUG
 | |
| 		reiserfs_debug(s, REISERFS_DEBUG_CODE,
 | |
| 			       "reiserquota: freeing (failed prealloc) %d blocks id=%u",
 | |
| 			       amount_needed + hint->prealloc_size -
 | |
| 			       nr_allocated -
 | |
| 			       REISERFS_I(hint->inode)->i_prealloc_count,
 | |
| 			       hint->inode->i_uid);
 | |
| #endif
 | |
| 
 | |
| 		depth = reiserfs_write_unlock_nested(s);
 | |
| 		dquot_free_block_nodirty(hint->inode, amount_needed +
 | |
| 					 hint->prealloc_size - nr_allocated -
 | |
| 					 REISERFS_I(hint->inode)->
 | |
| 					 i_prealloc_count);
 | |
| 		reiserfs_write_lock_nested(s, depth);
 | |
| 	}
 | |
| 
 | |
| 	return CARRY_ON;
 | |
| }
 | |
| 
 | |
| /* grab new blocknrs from preallocated list */
 | |
| /* return amount still needed after using them */
 | |
| static int use_preallocated_list_if_available(reiserfs_blocknr_hint_t * hint,
 | |
| 					      b_blocknr_t * new_blocknrs,
 | |
| 					      int amount_needed)
 | |
| {
 | |
| 	struct inode *inode = hint->inode;
 | |
| 
 | |
| 	if (REISERFS_I(inode)->i_prealloc_count > 0) {
 | |
| 		while (amount_needed) {
 | |
| 
 | |
| 			*new_blocknrs++ = REISERFS_I(inode)->i_prealloc_block++;
 | |
| 			REISERFS_I(inode)->i_prealloc_count--;
 | |
| 
 | |
| 			amount_needed--;
 | |
| 
 | |
| 			if (REISERFS_I(inode)->i_prealloc_count <= 0) {
 | |
| 				list_del(&REISERFS_I(inode)->i_prealloc_list);
 | |
| 				break;
 | |
| 			}
 | |
| 		}
 | |
| 	}
 | |
| 	/* return amount still needed after using preallocated blocks */
 | |
| 	return amount_needed;
 | |
| }
 | |
| 
 | |
| int reiserfs_allocate_blocknrs(reiserfs_blocknr_hint_t *hint,
 | |
| 			       b_blocknr_t *new_blocknrs,
 | |
| 			       int amount_needed,
 | |
| 			       /* Amount of blocks we have already reserved */
 | |
| 			       int reserved_by_us)
 | |
| {
 | |
| 	int initial_amount_needed = amount_needed;
 | |
| 	int ret;
 | |
| 	struct super_block *s = hint->th->t_super;
 | |
| 
 | |
| 	/* Check if there is enough space, taking into account reserved space */
 | |
| 	if (SB_FREE_BLOCKS(s) - REISERFS_SB(s)->reserved_blocks <
 | |
| 	    amount_needed - reserved_by_us)
 | |
| 		return NO_DISK_SPACE;
 | |
| 	/* should this be if !hint->inode &&  hint->preallocate? */
 | |
| 	/* do you mean hint->formatted_node can be removed ? - Zam */
 | |
| 	/*
 | |
| 	 * hint->formatted_node cannot be removed because we try to access
 | |
| 	 * inode information here, and there is often no inode associated with
 | |
| 	 * metadata allocations - green
 | |
| 	 */
 | |
| 
 | |
| 	if (!hint->formatted_node && hint->preallocate) {
 | |
| 		amount_needed = use_preallocated_list_if_available
 | |
| 		    (hint, new_blocknrs, amount_needed);
 | |
| 
 | |
| 		/*
 | |
| 		 * We have all the block numbers we need from the
 | |
| 		 * prealloc list
 | |
| 		 */
 | |
| 		if (amount_needed == 0)
 | |
| 			return CARRY_ON;
 | |
| 		new_blocknrs += (initial_amount_needed - amount_needed);
 | |
| 	}
 | |
| 
 | |
| 	/* find search start and save it in hint structure */
 | |
| 	determine_search_start(hint, amount_needed);
 | |
| 	if (hint->search_start >= SB_BLOCK_COUNT(s))
 | |
| 		hint->search_start = SB_BLOCK_COUNT(s) - 1;
 | |
| 
 | |
| 	/* allocation itself; fill new_blocknrs and preallocation arrays */
 | |
| 	ret = blocknrs_and_prealloc_arrays_from_search_start
 | |
| 	    (hint, new_blocknrs, amount_needed);
 | |
| 
 | |
| 	/*
 | |
| 	 * We used prealloc. list to fill (partially) new_blocknrs array.
 | |
| 	 * If final allocation fails we need to return blocks back to
 | |
| 	 * prealloc. list or just free them. -- Zam (I chose second
 | |
| 	 * variant)
 | |
| 	 */
 | |
| 	if (ret != CARRY_ON) {
 | |
| 		while (amount_needed++ < initial_amount_needed) {
 | |
| 			reiserfs_free_block(hint->th, hint->inode,
 | |
| 					    *(--new_blocknrs), 1);
 | |
| 		}
 | |
| 	}
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void reiserfs_cache_bitmap_metadata(struct super_block *sb,
 | |
|                                     struct buffer_head *bh,
 | |
|                                     struct reiserfs_bitmap_info *info)
 | |
| {
 | |
| 	unsigned long *cur = (unsigned long *)(bh->b_data + bh->b_size);
 | |
| 
 | |
| 	/* The first bit must ALWAYS be 1 */
 | |
| 	if (!reiserfs_test_le_bit(0, (unsigned long *)bh->b_data))
 | |
| 		reiserfs_error(sb, "reiserfs-2025", "bitmap block %lu is "
 | |
| 			       "corrupted: first bit must be 1", bh->b_blocknr);
 | |
| 
 | |
| 	info->free_count = 0;
 | |
| 
 | |
| 	while (--cur >= (unsigned long *)bh->b_data) {
 | |
| 		/* 0 and ~0 are special, we can optimize for them */
 | |
| 		if (*cur == 0)
 | |
| 			info->free_count += BITS_PER_LONG;
 | |
| 		else if (*cur != ~0L)	/* A mix, investigate */
 | |
| 			info->free_count += BITS_PER_LONG - hweight_long(*cur);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| struct buffer_head *reiserfs_read_bitmap_block(struct super_block *sb,
 | |
|                                                unsigned int bitmap)
 | |
| {
 | |
| 	b_blocknr_t block = (sb->s_blocksize << 3) * bitmap;
 | |
| 	struct reiserfs_bitmap_info *info = SB_AP_BITMAP(sb) + bitmap;
 | |
| 	struct buffer_head *bh;
 | |
| 
 | |
| 	/*
 | |
| 	 * Way old format filesystems had the bitmaps packed up front.
 | |
| 	 * I doubt there are any of these left, but just in case...
 | |
| 	 */
 | |
| 	if (unlikely(test_bit(REISERFS_OLD_FORMAT,
 | |
| 			      &REISERFS_SB(sb)->s_properties)))
 | |
| 		block = REISERFS_SB(sb)->s_sbh->b_blocknr + 1 + bitmap;
 | |
| 	else if (bitmap == 0)
 | |
| 		block = (REISERFS_DISK_OFFSET_IN_BYTES >> sb->s_blocksize_bits) + 1;
 | |
| 
 | |
| 	bh = sb_bread(sb, block);
 | |
| 	if (bh == NULL)
 | |
| 		reiserfs_warning(sb, "sh-2029: %s: bitmap block (#%u) "
 | |
| 		                 "reading failed", __func__, block);
 | |
| 	else {
 | |
| 		if (buffer_locked(bh)) {
 | |
| 			int depth;
 | |
| 			PROC_INFO_INC(sb, scan_bitmap.wait);
 | |
| 			depth = reiserfs_write_unlock_nested(sb);
 | |
| 			__wait_on_buffer(bh);
 | |
| 			reiserfs_write_lock_nested(sb, depth);
 | |
| 		}
 | |
| 		BUG_ON(!buffer_uptodate(bh));
 | |
| 		BUG_ON(atomic_read(&bh->b_count) == 0);
 | |
| 
 | |
| 		if (info->free_count == UINT_MAX)
 | |
| 			reiserfs_cache_bitmap_metadata(sb, bh, info);
 | |
| 	}
 | |
| 
 | |
| 	return bh;
 | |
| }
 | |
| 
 | |
| int reiserfs_init_bitmap_cache(struct super_block *sb)
 | |
| {
 | |
| 	struct reiserfs_bitmap_info *bitmap;
 | |
| 	unsigned int bmap_nr = reiserfs_bmap_count(sb);
 | |
| 
 | |
| 	bitmap = vmalloc(array_size(bmap_nr, sizeof(*bitmap)));
 | |
| 	if (bitmap == NULL)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	memset(bitmap, 0xff, sizeof(*bitmap) * bmap_nr);
 | |
| 
 | |
| 	SB_AP_BITMAP(sb) = bitmap;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| void reiserfs_free_bitmap_cache(struct super_block *sb)
 | |
| {
 | |
| 	if (SB_AP_BITMAP(sb)) {
 | |
| 		vfree(SB_AP_BITMAP(sb));
 | |
| 		SB_AP_BITMAP(sb) = NULL;
 | |
| 	}
 | |
| }
 |