diff --git a/ldiskfs/kernel_patches/patches/ext3-extents-bug11324.patch b/ldiskfs/kernel_patches/patches/ext3-extents-bug11324.patch deleted file mode 100644 index c7ed475da265de7eb48b8e003bc0631f08d7b85c..0000000000000000000000000000000000000000 --- a/ldiskfs/kernel_patches/patches/ext3-extents-bug11324.patch +++ /dev/null @@ -1,252 +0,0 @@ -Index: linux-stage/fs/ext3/extents.c -=================================================================== ---- linux-stage.orig/fs/ext3/extents.c 2007-04-17 22:09:19.000000000 -0700 -+++ linux-stage/fs/ext3/extents.c 2007-04-17 22:12:05.000000000 -0700 -@@ -44,26 +44,49 @@ - #include <asm/uaccess.h> - - --static inline int ext3_ext_check_header(struct ext3_extent_header *eh) --{ -- if (eh->eh_magic != EXT3_EXT_MAGIC) { -- printk(KERN_ERR "EXT3-fs: invalid magic = 0x%x\n", -- (unsigned)eh->eh_magic); -- return -EIO; -- } -- if (eh->eh_max == 0) { -- printk(KERN_ERR "EXT3-fs: invalid eh_max = %u\n", -- (unsigned)eh->eh_max); -- return -EIO; -- } -- if (eh->eh_entries > eh->eh_max) { -- printk(KERN_ERR "EXT3-fs: invalid eh_entries = %u\n", -- (unsigned)eh->eh_entries); -- return -EIO; -+static int __ext3_ext_check_header(const char *function, struct inode *inode, -+ struct ext3_extent_header *eh, int depth, -+ int max) -+{ -+ const char *error_msg = NULL; -+ -+ if (unlikely(eh->eh_magic != EXT3_EXT_MAGIC)) { -+ error_msg = "invalid magic"; -+ goto corrupted; -+ } -+ if (unlikely(eh->eh_depth != depth)) { -+ error_msg = "unexpected eh_depth"; -+ goto corrupted; -+ } -+ if (unlikely(eh->eh_max == 0)) { -+ error_msg = "too small eh_max"; -+ goto corrupted; -+ } -+ if (unlikely(eh->eh_max > max)) { -+ error_msg = "too large eh_max"; -+ goto corrupted; -+ } -+ if (unlikely(eh->eh_entries > eh->eh_max)) { -+ error_msg = "invalid eh_entries"; -+ goto corrupted; - } - return 0; -+ -+corrupted: -+ ext3_error(inode->i_sb, function, -+ "bad header in inode #%lu: %s - magic %x, " -+ "entries %u, max %u(%u), depth %u(%u)", -+ inode->i_ino, error_msg, eh->eh_magic, -+ eh->eh_entries, eh->eh_max, max, -+ eh->eh_depth, depth); -+ -+ return -EIO; - } - -+#define ext3_ext_check_header(inode,eh,depth,max) \ -+ __ext3_ext_check_header(__FUNCTION__,inode,eh,depth,max) -+ -+ - static handle_t *ext3_ext_journal_restart(handle_t *handle, int needed) - { - int err; -@@ -226,6 +249,26 @@ - return size; - } - -+static inline int -+ext3_ext_max_entries(struct ext3_extents_tree *tree, int root, int depth) -+{ -+ int max; -+ -+ if (root) { -+ if (depth == 0) -+ max = ext3_ext_space_root(tree); -+ else -+ max = ext3_ext_space_root_idx(tree); -+ } else { -+ if (depth == 0) -+ max = ext3_ext_space_block(tree); -+ else -+ max = ext3_ext_space_block_idx(tree); -+ } -+ -+ return max; -+} -+ - static void ext3_ext_show_path(struct ext3_extents_tree *tree, - struct ext3_ext_path *path) - { -@@ -296,10 +339,6 @@ - struct ext3_extent_idx *ix; - int l = 0, k, r; - -- EXT_ASSERT(eh->eh_magic == EXT3_EXT_MAGIC); -- EXT_ASSERT(eh->eh_entries <= eh->eh_max); -- EXT_ASSERT(eh->eh_entries > 0); -- - ext_debug(tree, "binsearch for %d(idx): ", block); - - path->p_idx = ix = EXT_FIRST_INDEX(eh); -@@ -359,9 +398,6 @@ - struct ext3_extent *ex; - int l = 0, k, r; - -- EXT_ASSERT(eh->eh_magic == EXT3_EXT_MAGIC); -- EXT_ASSERT(eh->eh_entries <= eh->eh_max); -- - if (eh->eh_entries == 0) { - /* - * this leaf is empty yet: -@@ -436,6 +472,7 @@ - struct ext3_extent_header *eh; - struct buffer_head *bh; - int depth, i, ppos = 0; -+ int max; - - EXT_ASSERT(tree); - EXT_ASSERT(tree->inode); -@@ -443,17 +480,15 @@ - - eh = EXT_ROOT_HDR(tree); - EXT_ASSERT(eh); -- if (ext3_ext_check_header(eh)) { -+ i = depth = EXT_DEPTH(tree); -+ max = ext3_ext_max_entries(tree, 1, i); -+ if (ext3_ext_check_header(tree->inode, eh, i, max)) { - /* don't free previously allocated path - * -- caller should take care */ - path = NULL; - goto err; - } - -- i = depth = EXT_DEPTH(tree); -- EXT_ASSERT(eh->eh_max); -- EXT_ASSERT(eh->eh_magic == EXT3_EXT_MAGIC); -- - /* account possible depth increase */ - if (!path) { - path = kmalloc(sizeof(struct ext3_ext_path) * (depth + 2), -@@ -484,8 +519,10 @@ - path[ppos].p_hdr = eh; - i--; - -- if (ext3_ext_check_header(eh)) -+ max = ext3_ext_max_entries(tree, 0, i); -+ if (ext3_ext_check_header(tree->inode, eh, i, max)) - goto err; -+ - } - - path[ppos].p_depth = i; -@@ -493,9 +530,6 @@ - path[ppos].p_ext = NULL; - path[ppos].p_idx = NULL; - -- if (ext3_ext_check_header(eh)) -- goto err; -- - /* find extent */ - ext3_ext_binsearch(tree, path + ppos, block); - -@@ -1545,6 +1579,8 @@ - ext_debug(tree, "remove [%lu:%lu] in leaf\n", start, end); - if (!path[depth].p_hdr) - path[depth].p_hdr = EXT_BLOCK_HDR(path[depth].p_bh); -+ -+ /* the header must be checked already in ext3_ext_remove_space() */ - eh = path[depth].p_hdr; - EXT_ASSERT(eh); - EXT_ASSERT(eh->eh_entries <= eh->eh_max); -@@ -1707,7 +1743,7 @@ - int depth = EXT_DEPTH(tree); - struct ext3_ext_path *path; - handle_t *handle; -- int i = 0, err = 0; -+ int i = 0, err = 0, max; - - ext_debug(tree, "space to be removed: %lu:%lu\n", start, end); - -@@ -1730,7 +1766,13 @@ - } - memset(path, 0, sizeof(struct ext3_ext_path) * (depth + 1)); - path[i].p_hdr = EXT_ROOT_HDR(tree); -- -+ -+ max = ext3_ext_max_entries(tree, 1, depth); -+ if (ext3_ext_check_header(inode, path[i].p_hdr, depth, max)) { -+ err = -EIO; -+ goto out; -+ } -+ - while (i >= 0 && err == 0) { - if (i == depth) { - /* this is leaf block */ -@@ -1740,16 +1782,13 @@ - i--; - continue; - } -- -+ - /* this is index block */ - if (!path[i].p_hdr) { - ext_debug(tree, "initialize header\n"); - path[i].p_hdr = EXT_BLOCK_HDR(path[i].p_bh); - } - -- EXT_ASSERT(path[i].p_hdr->eh_entries <= path[i].p_hdr->eh_max); -- EXT_ASSERT(path[i].p_hdr->eh_magic == EXT3_EXT_MAGIC); -- - if (!path[i].p_idx) { - /* this level hasn't touched yet */ - path[i].p_idx = -@@ -1776,6 +1815,14 @@ - err = -EIO; - break; - } -+ BUG_ON(i + 1 > depth); -+ max = ext3_ext_max_entries(tree, 0, depth - i - 1); -+ if (ext3_ext_check_header(inode, -+ EXT_BLOCK_HDR(path[i+1].p_bh), -+ depth - i - 1, max)) { -+ err = -EIO; -+ break; -+ } - /* put actual number of indexes to know is this - * number got changed at the next iteration */ - path[i].p_block = path[i].p_hdr->eh_entries; -@@ -1796,7 +1843,7 @@ - } - - /* TODO: flexible tree reduction should be here */ -- if (path->p_hdr->eh_entries == 0) { -+ if (err == 0 && path->p_hdr->eh_entries == 0) { - /* - * truncate to zero freed all the tree - * so, we need to correct eh_depth -@@ -1810,6 +1857,7 @@ - } - ext3_ext_tree_changed(tree); - -+out: - kfree(path); - ext3_journal_stop(handle); - diff --git a/ldiskfs/kernel_patches/patches/ext3-extents-fixes-2.6.9-rhel4.patch b/ldiskfs/kernel_patches/patches/ext3-extents-fixes-2.6.9-rhel4.patch new file mode 100644 index 0000000000000000000000000000000000000000..ffb9700950659afdb954b99bc2918418770b0764 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-extents-fixes-2.6.9-rhel4.patch @@ -0,0 +1,86 @@ + - minor fixes + - multiblock get_block() for direct I/O + +Index: linux-2.6.9-full/include/linux/ext3_extents.h +=================================================================== +--- linux-2.6.9-full.orig/include/linux/ext3_extents.h 2007-03-23 15:57:00.000000000 +0300 ++++ linux-2.6.9-full/include/linux/ext3_extents.h 2007-03-28 00:59:32.000000000 +0400 +@@ -205,7 +205,7 @@ typedef int (*ext_prepare_callback)(stru + #define EXT_DEPTH(__tree__) (EXT_ROOT_HDR(__tree__)->eh_depth) + #define EXT_GENERATION(__tree__) EXT_HDR_GEN(EXT_ROOT_HDR(__tree__)) + +-#define EXT_ASSERT(__x__) if (!(__x__)) BUG(); ++#define EXT_ASSERT(__x__) if (unlikely(!(__x__))) BUG(); + + #define EXT_CHECK_PATH(tree,path) \ + { \ +Index: linux-2.6.9-full/fs/ext3/extents.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/extents.c 2007-03-23 15:57:00.000000000 +0300 ++++ linux-2.6.9-full/fs/ext3/extents.c 2007-03-28 00:59:41.000000000 +0400 +@@ -895,6 +895,8 @@ repeat: + /* if we found index with free entry, then use that + * entry: create all needed subtree and add new leaf */ + err = ext3_ext_split(handle, tree, path, newext, i); ++ if (err) ++ goto out; + + /* refill path */ + ext3_ext_drop_refs(path); +@@ -904,12 +906,16 @@ repeat: + } else { + /* tree is full, time to grow in depth */ + err = ext3_ext_grow_indepth(handle, tree, path, newext); ++ if (err) ++ goto out; + + /* refill path */ + ext3_ext_drop_refs(path); + path = ext3_ext_find_extent(tree, newext->ee_block, path); +- if (IS_ERR(path)) ++ if (IS_ERR(path)) { + err = PTR_ERR(path); ++ goto out; ++ } + + /* + * only first (depth 0 -> 1) produces free space +@@ -922,10 +928,8 @@ repeat: + } + } + +- if (err) +- return err; +- +- return 0; ++out: ++ return err; + } + + /* +@@ -1992,21 +1996,10 @@ static int ext3_new_block_cb(handle_t *h + EXT_ASSERT(ex->ee_start); + EXT_ASSERT(ex->ee_len); + +- /* reuse block from the extent to order data/metadata */ +- newblock = ex->ee_start++; +- ex->ee_len--; +- if (ex->ee_len == 0) { +- ex->ee_len = 1; +- /* allocate new block for the extent */ +- goal = ext3_ext_find_goal(inode, path, ex->ee_block); +- ex->ee_start = ext3_new_block(handle, inode, goal, err); +- ex->ee_start_hi = 0; +- if (ex->ee_start == 0) { +- /* error occured: restore old extent */ +- ex->ee_start = newblock; +- return 0; +- } +- } ++ /* allocate new block for the extent */ ++ goal = ext3_ext_find_goal(inode, path, ex->ee_block); ++ newblock = ext3_new_block(handle, inode, goal, err); ++ + return newblock; + } + diff --git a/ldiskfs/kernel_patches/patches/ext3-extents-multiblock-directio-2.6.5-suse.patch b/ldiskfs/kernel_patches/patches/ext3-extents-multiblock-directio-2.6.5-suse.patch new file mode 100644 index 0000000000000000000000000000000000000000..744cc4589220b60fae517a451a6c0f3d2ce00857 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-extents-multiblock-directio-2.6.5-suse.patch @@ -0,0 +1,157 @@ +Index: linux-2.6.5-7.283-full/include/linux/ext3_fs.h +=================================================================== +--- linux-2.6.5-7.283-full.orig/include/linux/ext3_fs.h 2007-03-28 02:13:37.000000000 +0400 ++++ linux-2.6.5-7.283-full/include/linux/ext3_fs.h 2007-03-28 02:21:37.000000000 +0400 +@@ -815,7 +815,7 @@ extern struct inode_operations ext3_fast + + /* extents.c */ + extern int ext3_ext_writepage_trans_blocks(struct inode *, int); +-extern int ext3_ext_get_block(handle_t *, struct inode *, long, ++extern int ext3_ext_get_block(handle_t *, struct inode *, long, int, + struct buffer_head *, int, int); + extern void ext3_ext_truncate(struct inode *, struct page *); + extern void ext3_ext_init(struct super_block *); +Index: linux-2.6.5-7.283-full/fs/ext3/extents.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/extents.c 2007-03-28 02:14:25.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/extents.c 2007-03-28 02:21:37.000000000 +0400 +@@ -2024,7 +2024,8 @@ void ext3_init_tree_desc(struct ext3_ext + } + + int ext3_ext_get_block(handle_t *handle, struct inode *inode, +- long iblock, struct buffer_head *bh_result, ++ long iblock, int max_blocks, ++ struct buffer_head *bh_result, + int create, int extend_disksize) + { + struct ext3_ext_path *path = NULL; +@@ -2032,6 +2033,11 @@ int ext3_ext_get_block(handle_t *handle, + struct ext3_extent *ex; + int goal, newblock, err = 0, depth; + struct ext3_extents_tree tree; ++ unsigned long next; ++ int allocated = 0; ++ ++ /* until we have multiblock allocation */ ++ max_blocks = 1; + + __clear_bit(BH_New, &bh_result->b_state); + ext3_init_tree_desc(&tree, inode); +@@ -2051,6 +2057,9 @@ int ext3_ext_get_block(handle_t *handle, + } else if (goal == EXT3_EXT_CACHE_EXTENT) { + /* block is already allocated */ + newblock = iblock - newex.ee_block + newex.ee_start; ++ /* number of remaining blocks in the extent */ ++ EXT_ASSERT(iblock >= newex.ee_block); ++ allocated = newex.ee_len - (iblock - newex.ee_block); + goto out; + } else { + EXT_ASSERT(0); +@@ -2078,6 +2087,8 @@ int ext3_ext_get_block(handle_t *handle, + /* if found exent covers block, simple return it */ + if (iblock >= ex->ee_block && iblock < ex->ee_block + ex->ee_len) { + newblock = iblock - ex->ee_block + ex->ee_start; ++ /* number of remaining blocks in the extent */ ++ allocated = ex->ee_len - (iblock - ex->ee_block); + ext_debug(&tree, "%d fit into %d:%d -> %d\n", + (int) iblock, ex->ee_block, ex->ee_len, + newblock); +@@ -2098,6 +2109,15 @@ int ext3_ext_get_block(handle_t *handle, + goto out2; + } + ++ /* find next allocated block so that we know how many ++ * blocks we can allocate without ovelapping next extent */ ++ EXT_ASSERT(iblock >= ex->ee_block + ex->ee_len); ++ next = ext3_ext_next_allocated_block(path); ++ EXT_ASSERT(next > iblock); ++ allocated = next - iblock; ++ if (allocated > max_blocks) ++ allocated = max_blocks; ++ + /* allocate new block */ + goal = ext3_ext_find_goal(inode, path, iblock); + newblock = ext3_new_block(handle, inode, goal, &err); +@@ -2112,8 +2132,11 @@ int ext3_ext_get_block(handle_t *handle, + newex.ee_start_hi = 0; + newex.ee_len = 1; + err = ext3_ext_insert_extent(handle, &tree, path, &newex); +- if (err) ++ if (err) { ++ /* free data blocks we just allocated */ ++ ext3_free_blocks(handle, inode, newex.ee_start, newex.ee_len); + goto out2; ++ } + + if (extend_disksize && inode->i_size > EXT3_I(inode)->i_disksize) + EXT3_I(inode)->i_disksize = inode->i_size; +@@ -2125,10 +2148,13 @@ int ext3_ext_get_block(handle_t *handle, + ext3_ext_put_in_cache(&tree, newex.ee_block, newex.ee_len, + newex.ee_start, EXT3_EXT_CACHE_EXTENT); + out: ++ if (allocated > max_blocks) ++ allocated = max_blocks; + ext3_ext_show_leaf(&tree, path); + __set_bit(BH_Mapped, &bh_result->b_state); + bh_result->b_bdev = inode->i_sb->s_bdev; + bh_result->b_blocknr = newblock; ++ bh_result->b_size = (allocated << inode->i_blkbits); + out2: + if (path) { + ext3_ext_drop_refs(path); +Index: linux-2.6.5-7.283-full/fs/ext3/inode.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/inode.c 2007-03-28 02:13:37.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/inode.c 2007-03-28 02:50:19.000000000 +0400 +@@ -800,13 +800,17 @@ changed: + + static inline int + ext3_get_block_wrap(handle_t *handle, struct inode *inode, long block, +- struct buffer_head *bh, int create, int extend_disksize) ++ int max_blocks, struct buffer_head *bh, int create, ++ int extend_disksize) + { ++ int ret; + if (EXT3_I(inode)->i_flags & EXT3_EXTENTS_FL) +- return ext3_ext_get_block(handle, inode, block, bh, create, +- extend_disksize); +- return ext3_get_block_handle(handle, inode, block, bh, create, ++ return ext3_ext_get_block(handle, inode, block, max_blocks, ++ bh, create, extend_disksize); ++ ret = ext3_get_block_handle(handle, inode, block, bh, create, + extend_disksize); ++ bh->b_size = (1 << inode->i_blkbits); ++ return ret; + } + + static int ext3_get_block(struct inode *inode, sector_t iblock, +@@ -819,7 +823,7 @@ static int ext3_get_block(struct inode * + handle = ext3_journal_current_handle(); + J_ASSERT(handle != 0); + } +- ret = ext3_get_block_wrap(handle, inode, iblock, ++ ret = ext3_get_block_wrap(handle, inode, iblock, 1, + bh_result, create, 1); + return ret; + } +@@ -847,10 +851,8 @@ ext3_direct_io_get_blocks(struct inode * + } + } + if (ret == 0) +- ret = ext3_get_block_wrap(handle, inode, iblock, ++ ret = ext3_get_block_wrap(handle, inode, iblock, max_blocks, + bh_result, create, 0); +- if (ret == 0) +- bh_result->b_size = (1 << inode->i_blkbits); + return ret; + } + +@@ -869,7 +871,7 @@ struct buffer_head *ext3_getblk(handle_t + dummy.b_state = 0; + dummy.b_blocknr = -1000; + buffer_trace_init(&dummy.b_history); +- *errp = ext3_get_block_wrap(handle, inode, block, &dummy, create, 1); ++ *errp = ext3_get_block_wrap(handle, inode, block, 1, &dummy, create, 1); + if (!*errp && buffer_mapped(&dummy)) { + struct buffer_head *bh; + bh = sb_getblk(inode->i_sb, dummy.b_blocknr); diff --git a/ldiskfs/kernel_patches/patches/ext3-extents-multiblock-directio-2.6.9-rhel4.patch b/ldiskfs/kernel_patches/patches/ext3-extents-multiblock-directio-2.6.9-rhel4.patch new file mode 100644 index 0000000000000000000000000000000000000000..726a47373829be221d61ed9fec366bbc9636351c --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-extents-multiblock-directio-2.6.9-rhel4.patch @@ -0,0 +1,149 @@ +--- linux-2.6.9-full/include/linux/ext3_fs.h 2007-03-23 15:57:00.000000000 +0300 ++++ linux-2.6.9-full/include/linux/ext3_fs.h 2007-02-16 17:16:23.000000000 +0300 +@@ -850,7 +850,7 @@ extern struct inode_operations ext3_fast + + /* extents.c */ + extern int ext3_ext_writepage_trans_blocks(struct inode *, int); +-extern int ext3_ext_get_block(handle_t *, struct inode *, long, ++extern int ext3_ext_get_block(handle_t *, struct inode *, long, int, + struct buffer_head *, int, int); + extern void ext3_ext_truncate(struct inode *, struct page *); + extern void ext3_ext_init(struct super_block *); + +--- linux-2.6.9-full/fs/ext3/extents.c 2007-03-23 15:57:00.000000000 +0300 ++++ linux-2.6.9-full/fs/ext3/extents.c 2007-02-22 17:45:05.000000000 +0300 +@@ -2031,7 +2168,8 @@ void ext3_init_tree_desc(struct ext3_ext + } + + int ext3_ext_get_block(handle_t *handle, struct inode *inode, +- long iblock, struct buffer_head *bh_result, ++ long iblock, int max_blocks, ++ struct buffer_head *bh_result, + int create, int extend_disksize) + { + struct ext3_ext_path *path = NULL; +@@ -2039,6 +2177,11 @@ int ext3_ext_get_block(handle_t *handle, + struct ext3_extent *ex; + int goal, newblock, err = 0, depth; + struct ext3_extents_tree tree; ++ unsigned long next; ++ int allocated = 0; ++ ++ /* until we have multiblock allocation */ ++ max_blocks = 1; + + clear_buffer_new(bh_result); + ext3_init_tree_desc(&tree, inode); +@@ -2058,6 +2201,9 @@ int ext3_ext_get_block(handle_t *handle, + } else if (goal == EXT3_EXT_CACHE_EXTENT) { + /* block is already allocated */ + newblock = iblock - newex.ee_block + newex.ee_start; ++ /* number of remaining blocks in the extent */ ++ EXT_ASSERT(iblock >= newex.ee_block); ++ allocated = newex.ee_len - (iblock - newex.ee_block); + goto out; + } else { + EXT_ASSERT(0); +@@ -2085,6 +2231,8 @@ int ext3_ext_get_block(handle_t *handle, + /* if found exent covers block, simple return it */ + if (iblock >= ex->ee_block && iblock < ex->ee_block + ex->ee_len) { + newblock = iblock - ex->ee_block + ex->ee_start; ++ /* number of remaining blocks in the extent */ ++ allocated = ex->ee_len - (iblock - ex->ee_block); + ext_debug(&tree, "%d fit into %d:%d -> %d\n", + (int) iblock, ex->ee_block, ex->ee_len, + newblock); +@@ -2105,6 +2253,15 @@ int ext3_ext_get_block(handle_t *handle, + goto out2; + } + ++ /* find next allocated block so that we know how many ++ * blocks we can allocate without ovelapping next extent */ ++ EXT_ASSERT(iblock >= ex->ee_block + ex->ee_len); ++ next = ext3_ext_next_allocated_block(path); ++ EXT_ASSERT(next > iblock); ++ allocated = next - iblock; ++ if (allocated > max_blocks) ++ allocated = max_blocks; ++ + /* allocate new block */ + goal = ext3_ext_find_goal(inode, path, iblock); + newblock = ext3_new_block(handle, inode, goal, &err); +@@ -2119,8 +2276,11 @@ int ext3_ext_get_block(handle_t *handle, + newex.ee_start_hi = 0; + newex.ee_len = 1; + err = ext3_ext_insert_extent(handle, &tree, path, &newex); +- if (err) ++ if (err) { ++ /* free data blocks we just allocated */ ++ ext3_free_blocks(handle, inode, newex.ee_start, newex.ee_len); + goto out2; ++ } + + if (extend_disksize && inode->i_size > EXT3_I(inode)->i_disksize) + EXT3_I(inode)->i_disksize = inode->i_size; +@@ -2132,8 +2292,11 @@ int ext3_ext_get_block(handle_t *handle, + ext3_ext_put_in_cache(&tree, newex.ee_block, newex.ee_len, + newex.ee_start, EXT3_EXT_CACHE_EXTENT); + out: ++ if (allocated > max_blocks) ++ allocated = max_blocks; + ext3_ext_show_leaf(&tree, path); + map_bh(bh_result, inode->i_sb, newblock); ++ bh_result->b_size = (allocated << inode->i_blkbits); + out2: + if (path) { + ext3_ext_drop_refs(path); +--- linux-2.6.9-full/fs/ext3/inode.c 2007-03-23 15:57:00.000000000 +0300 ++++ linux-2.6.9-full/fs/ext3/inode.c 2007-02-16 17:17:03.000000000 +0300 +@@ -798,13 +798,17 @@ changed: + + static inline int + ext3_get_block_wrap(handle_t *handle, struct inode *inode, long block, +- struct buffer_head *bh, int create, int extend_disksize) ++ int max_blocks, struct buffer_head *bh, int create, ++ int extend_disksize) + { ++ int ret; + if (EXT3_I(inode)->i_flags & EXT3_EXTENTS_FL) +- return ext3_ext_get_block(handle, inode, block, bh, create, +- extend_disksize); +- return ext3_get_block_handle(handle, inode, block, bh, create, ++ return ext3_ext_get_block(handle, inode, block, max_blocks, ++ bh, create, extend_disksize); ++ ret = ext3_get_block_handle(handle, inode, block, bh, create, + extend_disksize); ++ bh->b_size = (1 << inode->i_blkbits); ++ return ret; + } + + static int ext3_get_block(struct inode *inode, sector_t iblock, +@@ -817,7 +821,7 @@ static int ext3_get_block(struct inode * + handle = ext3_journal_current_handle(); + J_ASSERT(handle != 0); + } +- ret = ext3_get_block_wrap(handle, inode, iblock, ++ ret = ext3_get_block_wrap(handle, inode, iblock, 1, + bh_result, create, 1); + return ret; + } +@@ -862,9 +866,8 @@ ext3_direct_io_get_blocks(struct inode * + + get_block: + if (ret == 0) +- ret = ext3_get_block_wrap(handle, inode, iblock, ++ ret = ext3_get_block_wrap(handle, inode, iblock, max_blocks, + bh_result, create, 0); +- bh_result->b_size = (1 << inode->i_blkbits); + return ret; + } + +@@ -882,7 +885,7 @@ struct buffer_head *ext3_getblk(handle_t + dummy.b_state = 0; + dummy.b_blocknr = -1000; + buffer_trace_init(&dummy.b_history); +- *errp = ext3_get_block_wrap(handle, inode, block, &dummy, create, 1); ++ *errp = ext3_get_block_wrap(handle, inode, block, 1, &dummy, create, 1); + if (!*errp && buffer_mapped(&dummy)) { + struct buffer_head *bh; + bh = sb_getblk(inode->i_sb, dummy.b_blocknr); diff --git a/ldiskfs/kernel_patches/patches/ext3-extents-search-2.6.9-rhel4.patch b/ldiskfs/kernel_patches/patches/ext3-extents-search-2.6.9-rhel4.patch new file mode 100644 index 0000000000000000000000000000000000000000..2ad69c8c4c3f509847cf9acdf62cbae2532dd6c5 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-extents-search-2.6.9-rhel4.patch @@ -0,0 +1,168 @@ +Index: linux-2.6.9-full/include/linux/ext3_extents.h +=================================================================== +--- linux-2.6.9-full.orig/include/linux/ext3_extents.h 2007-03-23 15:57:00.000000000 +0300 ++++ linux-2.6.9-full/include/linux/ext3_extents.h 2007-03-26 22:08:16.000000000 +0400 +@@ -242,6 +242,8 @@ struct ext3_extent_tree_stats { + int leaf_num; + }; + ++extern int ext3_ext_search_left(struct ext3_extents_tree *, struct ext3_ext_path *, unsigned long *, unsigned long *); ++extern int ext3_ext_search_right(struct ext3_extents_tree *, struct ext3_ext_path *, unsigned long *, unsigned long *); + extern void ext3_init_tree_desc(struct ext3_extents_tree *, struct inode *); + extern int ext3_extent_tree_init(handle_t *, struct ext3_extents_tree *); + extern int ext3_ext_calc_credits_for_insert(struct ext3_extents_tree *, struct ext3_ext_path *); +Index: linux-2.6.9-full/fs/ext3/extents.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/extents.c 2007-03-23 15:57:00.000000000 +0300 ++++ linux-2.6.9-full/fs/ext3/extents.c 2007-03-26 22:07:37.000000000 +0400 +@@ -929,6 +929,150 @@ repeat: + } + + /* ++ * search the closest allocated block to the left for *logical ++ * and returns it at @logical + it's physical address at @phys ++ * if *logical is the smallest allocated block, the function ++ * returns 0 at @phys ++ * return value contains 0 (success) or error code ++ */ ++int ++ext3_ext_search_left(struct ext3_extents_tree *tree, struct ext3_ext_path *path, ++ unsigned long *logical, unsigned long *phys) ++{ ++ struct ext3_extent_idx *ix; ++ struct ext3_extent *ex; ++ int depth; ++ ++ BUG_ON(path == NULL); ++ depth = path->p_depth; ++ *phys = 0; ++ ++ if (depth == 0 && path->p_ext == NULL) ++ return 0; ++ ++ /* usually extent in the path covers blocks smaller ++ * then *logical, but it can be that extent is the ++ * first one in the file */ ++ ++ ex = path[depth].p_ext; ++ if (*logical < ex->ee_block) { ++ BUG_ON(EXT_FIRST_EXTENT(path[depth].p_hdr) != ex); ++ while (--depth >= 0) { ++ ix = path[depth].p_idx; ++ BUG_ON(ix != EXT_FIRST_INDEX(path[depth].p_hdr)); ++ } ++ return 0; ++ } ++ ++ BUG_ON(*logical < ex->ee_block + ex->ee_len); ++ ++ *logical = ex->ee_block + ex->ee_len - 1; ++ *phys = ex->ee_start + ex->ee_len - 1; ++ return 0; ++} ++EXPORT_SYMBOL(ext3_ext_search_left); ++ ++/* ++ * search the closest allocated block to the right for *logical ++ * and returns it at @logical + it's physical address at @phys ++ * if *logical is the smallest allocated block, the function ++ * returns 0 at @phys ++ * return value contains 0 (success) or error code ++ */ ++int ++ext3_ext_search_right(struct ext3_extents_tree *tree, struct ext3_ext_path *path, ++ unsigned long *logical, unsigned long *phys) ++{ ++ struct buffer_head *bh = NULL; ++ struct ext3_extent_header *eh; ++ struct ext3_extent_idx *ix; ++ struct ext3_extent *ex; ++ unsigned long block; ++ int depth; ++ ++ BUG_ON(path == NULL); ++ depth = path->p_depth; ++ *phys = 0; ++ ++ if (depth == 0 && path->p_ext == NULL) ++ return 0; ++ ++ /* usually extent in the path covers blocks smaller ++ * then *logical, but it can be that extent is the ++ * first one in the file */ ++ ++ ex = path[depth].p_ext; ++ if (*logical < ex->ee_block) { ++ BUG_ON(EXT_FIRST_EXTENT(path[depth].p_hdr) != ex); ++ while (--depth >= 0) { ++ ix = path[depth].p_idx; ++ BUG_ON(ix != EXT_FIRST_INDEX(path[depth].p_hdr)); ++ } ++ *logical = ex->ee_block; ++ *phys = ex->ee_start; ++ return 0; ++ } ++ ++ BUG_ON(*logical < ex->ee_block + ex->ee_len); ++ ++ if (ex != EXT_LAST_EXTENT(path[depth].p_hdr)) { ++ /* next allocated block in this leaf */ ++ ex++; ++ *logical = ex->ee_block; ++ *phys = ex->ee_start; ++ return 0; ++ } ++ ++ /* go up and search for index to the right */ ++ while (--depth >= 0) { ++ ix = path[depth].p_idx; ++ if (ix != EXT_LAST_INDEX(path[depth].p_hdr)) ++ break; ++ } ++ ++ if (depth < 0) { ++ /* we've gone up to the root and ++ * found no index to the right */ ++ return 0; ++ } ++ ++ /* we've found index to the right, let's ++ * follow it and find the closest allocated ++ * block to the right */ ++ ix++; ++ block = ix->ei_leaf; ++ while (++depth < path->p_depth) { ++ bh = sb_bread(tree->inode->i_sb, block); ++ if (bh == NULL) ++ return -EIO; ++ eh = EXT_BLOCK_HDR(bh); ++ if (ext3_ext_check_header(eh)) { ++ brelse(bh); ++ return -EIO; ++ } ++ ix = EXT_FIRST_INDEX(eh); ++ block = ix->ei_leaf; ++ brelse(bh); ++ } ++ ++ bh = sb_bread(tree->inode->i_sb, block); ++ if (bh == NULL) ++ return -EIO; ++ eh = EXT_BLOCK_HDR(bh); ++ if (ext3_ext_check_header(eh)) { ++ brelse(bh); ++ return -EIO; ++ } ++ ex = EXT_FIRST_EXTENT(eh); ++ *logical = ex->ee_block; ++ *phys = ex->ee_start; ++ brelse(bh); ++ return 0; ++ ++} ++EXPORT_SYMBOL(ext3_ext_search_right); ++ ++/* + * returns allocated block in subsequent extent or EXT_MAX_BLOCK + * NOTE: it consider block number from index entry as + * allocated block. thus, index entries have to be consistent diff --git a/ldiskfs/kernel_patches/patches/ext3-filterdata-sles10.patch b/ldiskfs/kernel_patches/patches/ext3-filterdata-sles10.patch new file mode 100644 index 0000000000000000000000000000000000000000..5f7c8c9b9917ed26bd36c72ef10bdb72b2fe1711 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-filterdata-sles10.patch @@ -0,0 +1,25 @@ +Index: linux-2.6.16.27-0.9-full/include/linux/ext3_fs_i.h +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/include/linux/ext3_fs_i.h 2007-03-28 16:03:20.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/include/linux/ext3_fs_i.h 2007-03-28 19:40:53.000000000 +0400 +@@ -139,6 +139,8 @@ struct ext3_inode_info { + /* mballoc */ + struct list_head i_prealloc_list; + spinlock_t i_prealloc_lock; ++ ++ void *i_filterdata; + }; + + #endif /* _LINUX_EXT3_FS_I */ +Index: linux-2.6.16.27-0.9-full/fs/ext3/super.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/super.c 2007-03-28 18:20:17.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/fs/ext3/super.c 2007-03-28 19:40:53.000000000 +0400 +@@ -462,6 +462,7 @@ static struct inode *ext3_alloc_inode(st + memset(&ei->i_cached_extent, 0, sizeof(ei->i_cached_extent)); + INIT_LIST_HEAD(&ei->i_prealloc_list); + spin_lock_init(&ei->i_prealloc_lock); ++ ei->i_filterdata = NULL; + return &ei->vfs_inode; + } + diff --git a/ldiskfs/kernel_patches/patches/ext3-inode-version-2.6-sles10.patch b/ldiskfs/kernel_patches/patches/ext3-inode-version-2.6-sles10.patch new file mode 100644 index 0000000000000000000000000000000000000000..7b6b1b8890a5cf8e473c8d6143e83f717a361258 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-inode-version-2.6-sles10.patch @@ -0,0 +1,426 @@ +Index: linux-2.6.16-sles10/fs/ext3/inode.c +=================================================================== +--- linux-2.6.16-sles10.orig/fs/ext3/inode.c ++++ linux-2.6.16-sles10/fs/ext3/inode.c +@@ -2558,6 +2558,13 @@ void ext3_read_inode(struct inode * inod + EXT3_INODE_GET_XTIME(i_atime, inode, raw_inode); + EXT3_EINODE_GET_XTIME(i_crtime, ei, raw_inode); + ++ ei->i_fs_version = le32_to_cpu(raw_inode->i_disk_version); ++ if (EXT3_INODE_SIZE(inode->i_sb) > EXT3_GOOD_OLD_INODE_SIZE) { ++ if (EXT3_FITS_IN_INODE(raw_inode, ei, i_version_hi)) ++ ei->i_fs_version |= (__u64)(le32_to_cpu(raw_inode->i_version_hi)) ++ << 32; ++ } ++ + if (S_ISREG(inode->i_mode)) { + inode->i_op = &ext3_file_inode_operations; + inode->i_fop = &ext3_file_operations; +@@ -2696,8 +2703,14 @@ static int ext3_do_update_inode(handle_t + } else for (block = 0; block < EXT3_N_BLOCKS; block++) + raw_inode->i_block[block] = ei->i_data[block]; + +- if (ei->i_extra_isize) ++ raw_inode->i_disk_version = cpu_to_le32(ei->i_fs_version); ++ if (ei->i_extra_isize) { ++ if (EXT3_FITS_IN_INODE(raw_inode, ei, i_version_hi)) { ++ raw_inode->i_version_hi = cpu_to_le32(ei->i_fs_version ++ >> 32); ++ } + raw_inode->i_extra_isize = cpu_to_le16(ei->i_extra_isize); ++ } + + BUFFER_TRACE(bh, "call ext3_journal_dirty_metadata"); + rc = ext3_journal_dirty_metadata(handle, bh); +@@ -2971,10 +2984,32 @@ ext3_reserve_inode_write(handle_t *handl + int ext3_mark_inode_dirty(handle_t *handle, struct inode *inode) + { + struct ext3_iloc iloc; +- int err; ++ int err, ret; ++ static int expand_message; + + might_sleep(); + err = ext3_reserve_inode_write(handle, inode, &iloc); ++ if (EXT3_I(inode)->i_extra_isize < ++ EXT3_SB(inode->i_sb)->s_want_extra_isize && ++ !(EXT3_I(inode)->i_state & EXT3_STATE_NO_EXPAND)) { ++ /* We need extra buffer credits since we may write into EA block ++ * with this same handle */ ++ if ((ext3_journal_extend(handle, ++ EXT3_DATA_TRANS_BLOCKS(inode->i_sb))) == 0) { ++ ret = ext3_expand_extra_isize(inode, ++ EXT3_SB(inode->i_sb)->s_want_extra_isize, ++ iloc, handle); ++ if (ret) { ++ EXT3_I(inode)->i_state |= EXT3_STATE_NO_EXPAND; ++ if (!expand_message) { ++ ext3_warning(inode->i_sb, __FUNCTION__, ++ "Unable to expand inode %lu. Delete some" ++ " EAs or run e2fsck.", inode->i_ino); ++ expand_message = 1; ++ } ++ } ++ } ++ } + if (!err) + err = ext3_mark_iloc_dirty(handle, inode, &iloc); + return err; +Index: linux-2.6.16-sles10/include/linux/ext3_fs.h +=================================================================== +--- linux-2.6.16-sles10.orig/include/linux/ext3_fs.h ++++ linux-2.6.16-sles10/include/linux/ext3_fs.h +@@ -205,6 +205,7 @@ struct ext3_group_desc + #define EXT3_STATE_JDATA 0x00000001 /* journaled data exists */ + #define EXT3_STATE_NEW 0x00000002 /* inode is newly created */ + #define EXT3_STATE_XATTR 0x00000004 /* has in-inode xattrs */ ++#define EXT3_STATE_NO_EXPAND 0x00000008 /* No space for expansion */ + + /* Used to pass group descriptor data when online resize is done */ + struct ext3_new_group_input { +@@ -281,7 +282,7 @@ struct ext3_inode { + __le32 i_flags; /* File flags */ + union { + struct { +- __u32 l_i_reserved1; ++ __u32 l_i_version; + } linux1; + struct { + __u32 h_i_translator; +@@ -326,6 +327,7 @@ struct ext3_inode { + __le32 i_atime_extra; /* extra Access time (nsec << 2 | epoch) */ + __le32 i_crtime; /* File Creation time */ + __le32 i_crtime_extra; /* extra File Creation time (nsec << 2 | epoch) */ ++ __le32 i_version_hi; /* high 32 bits for 64-bit version */ + }; + + #define i_size_high i_dir_acl +@@ -388,6 +390,8 @@ do { \ + raw_inode->xtime ## _extra); \ + } while (0) + ++#define i_disk_version osd1.linux1.l_i_version ++ + #if defined(__KERNEL__) || defined(__linux__) + #define i_reserved1 osd1.linux1.l_i_reserved1 + #define i_frag osd2.linux2.l_i_frag +Index: linux-2.6.16-sles10/include/linux/ext3_fs_i.h +=================================================================== +--- linux-2.6.16-sles10.orig/include/linux/ext3_fs_i.h ++++ linux-2.6.16-sles10/include/linux/ext3_fs_i.h +@@ -20,6 +20,8 @@ + #include <linux/rbtree.h> + #include <linux/seqlock.h> + ++#define HAVE_DISK_INODE_VERSION ++ + struct ext3_reserve_window { + __u32 _rsv_start; /* First byte reserved */ + __u32 _rsv_end; /* Last byte reserved or 0 */ +@@ -138,6 +140,8 @@ struct ext3_inode_info { + __u32 i_cached_extent[4]; + + void *i_filterdata; ++ ++ __u64 i_fs_version; + }; + + #endif /* _LINUX_EXT3_FS_I */ +Index: linux-2.6.16-sles10/fs/ext3/xattr.c +=================================================================== +--- linux-2.6.16-sles10.orig/fs/ext3/xattr.c ++++ linux-2.6.16-sles10/fs/ext3/xattr.c +@@ -505,6 +505,20 @@ ext3_xattr_release_block(handle_t *handl + } + } + ++static inline size_t ext3_xattr_free_space(struct ext3_xattr_entry *last, ++ size_t *min_offs, void *base, int *total) ++{ ++ for (; !IS_LAST_ENTRY(last); last = EXT3_XATTR_NEXT(last)) { ++ *total += EXT3_XATTR_LEN(last->e_name_len); ++ if (!last->e_value_block && last->e_value_size) { ++ size_t offs = le16_to_cpu(last->e_value_offs); ++ if (offs < *min_offs) ++ *min_offs = offs; ++ } ++ } ++ return (*min_offs - ((void *)last - base) - sizeof(__u32)); ++} ++ + struct ext3_xattr_info { + int name_index; + const char *name; +@@ -1007,6 +1021,8 @@ ext3_xattr_set_handle(handle_t *handle, + if (!error) { + ext3_xattr_update_super_block(handle, inode->i_sb); + inode->i_ctime = ext3_current_time(inode); ++ if (!value) ++ EXT3_I(inode)->i_state &= ~EXT3_STATE_NO_EXPAND; + ext3_mark_inode_dirty(handle, inode); + /* + * The bh is consumed by ext3_mark_iloc_dirty, even with +@@ -1059,6 +1075,249 @@ retry: + return error; + } + ++static void ext3_xattr_shift_entries(struct ext3_xattr_entry *entry, ++ int value_offs_shift, void *to, ++ void *from, size_t n, int blocksize) ++{ ++ struct ext3_xattr_entry *last = entry; ++ int new_offs; ++ ++ /* Adjust the value offsets of the entries */ ++ for (; !IS_LAST_ENTRY(last); last = EXT3_XATTR_NEXT(last)) { ++ if (!last->e_value_block && last->e_value_size) { ++ new_offs = le16_to_cpu(last->e_value_offs) + ++ value_offs_shift; ++ BUG_ON(new_offs + le32_to_cpu(last->e_value_size) > ++ blocksize); ++ last->e_value_offs = cpu_to_le16(new_offs); ++ } ++ } ++ /* Shift the entries by n bytes */ ++ memmove(to, from, n); ++} ++ ++/* Expand an inode by new_extra_isize bytes. ++ * Returns 0 on success or negative error number on failure. ++ */ ++int ext3_expand_extra_isize(struct inode *inode, int new_extra_isize, ++ struct ext3_iloc iloc, handle_t *handle) ++{ ++ struct ext3_inode *raw_inode; ++ struct ext3_xattr_ibody_header *header; ++ struct ext3_xattr_entry *entry, *last, *first; ++ struct buffer_head *bh = NULL; ++ struct ext3_xattr_ibody_find *is = NULL; ++ struct ext3_xattr_block_find *bs = NULL; ++ char *buffer = NULL, *b_entry_name = NULL; ++ size_t min_offs, free; ++ int total_ino, total_blk; ++ void *base, *start, *end; ++ int extra_isize = 0, error = 0, tried_min_extra_isize = 0; ++ int s_min_extra_isize = EXT3_SB(inode->i_sb)->s_es->s_min_extra_isize; ++ ++ down_write(&EXT3_I(inode)->xattr_sem); ++ ++retry: ++ if (EXT3_I(inode)->i_extra_isize >= new_extra_isize) { ++ up_write(&EXT3_I(inode)->xattr_sem); ++ return 0; ++ } ++ ++ raw_inode = ext3_raw_inode(&iloc); ++ ++ header = IHDR(inode, raw_inode); ++ entry = IFIRST(header); ++ ++ /* No extended attributes present */ ++ if (!(EXT3_I(inode)->i_state & EXT3_STATE_XATTR) || ++ header->h_magic != cpu_to_le32(EXT3_XATTR_MAGIC)) { ++ memset((void *)raw_inode + EXT3_GOOD_OLD_INODE_SIZE, 0, ++ new_extra_isize); ++ EXT3_I(inode)->i_extra_isize = new_extra_isize; ++ goto cleanup; ++ } ++ ++ /* ++ * Check if enough free space is available in the inode to shift the ++ * entries ahead by new_extra_isize. ++ */ ++ ++ base = start = entry; ++ end = (void *)raw_inode + EXT3_SB(inode->i_sb)->s_inode_size; ++ min_offs = end - base; ++ last = entry; ++ total_ino = sizeof(struct ext3_xattr_ibody_header); ++ ++ free = ext3_xattr_free_space(last, &min_offs, base, &total_ino); ++ if (free >= new_extra_isize) { ++ entry = IFIRST(header); ++ ext3_xattr_shift_entries(entry, EXT3_I(inode)->i_extra_isize - ++ new_extra_isize, (void *)raw_inode + ++ EXT3_GOOD_OLD_INODE_SIZE + new_extra_isize, ++ (void *)header, total_ino, ++ inode->i_sb->s_blocksize); ++ EXT3_I(inode)->i_extra_isize = new_extra_isize; ++ error = 0; ++ goto cleanup; ++ } ++ ++ /* ++ * Enough free space isn't available in the inode, check if ++ * EA block can hold new_extra_isize bytes. ++ */ ++ if (EXT3_I(inode)->i_file_acl) { ++ bh = sb_bread(inode->i_sb, EXT3_I(inode)->i_file_acl); ++ error = -EIO; ++ if (!bh) ++ goto cleanup; ++ if (ext3_xattr_check_block(bh)) { ++ ext3_error(inode->i_sb, __FUNCTION__, ++ "inode %lu: bad block %d", inode->i_ino, ++ EXT3_I(inode)->i_file_acl); ++ error = -EIO; ++ goto cleanup; ++ } ++ base = BHDR(bh); ++ first = BFIRST(bh); ++ end = bh->b_data + bh->b_size; ++ min_offs = end - base; ++ free = ext3_xattr_free_space(first, &min_offs, base, ++ &total_blk); ++ if (free < new_extra_isize) { ++ if (!tried_min_extra_isize && s_min_extra_isize) { ++ tried_min_extra_isize++; ++ new_extra_isize = s_min_extra_isize; ++ goto retry; ++ } ++ error = -1; ++ goto cleanup; ++ } ++ } else { ++ free = inode->i_sb->s_blocksize; ++ } ++ ++ while (new_extra_isize > 0) { ++ size_t offs, size, entry_size; ++ struct ext3_xattr_entry *small_entry = NULL; ++ struct ext3_xattr_info i = { ++ .value = NULL, ++ .value_len = 0, ++ }; ++ unsigned int total_size, shift_bytes, temp = ~0U; ++ ++ is = (struct ext3_xattr_ibody_find *) kmalloc(sizeof(struct ++ ext3_xattr_ibody_find), GFP_KERNEL); ++ bs = (struct ext3_xattr_block_find *) kmalloc(sizeof(struct ++ ext3_xattr_block_find), GFP_KERNEL); ++ memset((void *)is, 0, sizeof(struct ext3_xattr_ibody_find)); ++ memset((void *)bs, 0, sizeof(struct ext3_xattr_block_find)); ++ ++ is->s.not_found = bs->s.not_found = -ENODATA; ++ is->iloc.bh = NULL; ++ bs->bh = NULL; ++ ++ last = IFIRST(header); ++ /* Find the entry best suited to be pushed into EA block */ ++ entry = NULL; ++ for (; !IS_LAST_ENTRY(last); last = EXT3_XATTR_NEXT(last)) { ++ total_size = EXT3_XATTR_SIZE(le32_to_cpu(last->e_value_size)) + ++ EXT3_XATTR_LEN(last->e_name_len); ++ if (total_size <= free && total_size < temp) { ++ if (total_size < new_extra_isize) { ++ small_entry = last; ++ } else { ++ entry = last; ++ temp = total_size; ++ } ++ } ++ } ++ ++ if (entry == NULL) { ++ if (small_entry) { ++ entry = small_entry; ++ } else { ++ if (!tried_min_extra_isize && ++ s_min_extra_isize) { ++ tried_min_extra_isize++; ++ new_extra_isize = s_min_extra_isize; ++ goto retry; ++ } ++ error = -1; ++ goto cleanup; ++ } ++ } ++ offs = le16_to_cpu(entry->e_value_offs); ++ size = le32_to_cpu(entry->e_value_size); ++ entry_size = EXT3_XATTR_LEN(entry->e_name_len); ++ i.name_index = entry->e_name_index, ++ buffer = kmalloc(EXT3_XATTR_SIZE(size), GFP_KERNEL); ++ b_entry_name = kmalloc(entry->e_name_len + 1, GFP_KERNEL); ++ /* Save the entry name and the entry value */ ++ memcpy((void *)buffer, (void *)IFIRST(header) + offs, ++ EXT3_XATTR_SIZE(size)); ++ memcpy((void *)b_entry_name, (void *)entry->e_name, ++ entry->e_name_len); ++ b_entry_name[entry->e_name_len] = '\0'; ++ i.name = b_entry_name; ++ ++ error = ext3_get_inode_loc(inode, &is->iloc); ++ if (error) ++ goto cleanup; ++ ++ error = ext3_xattr_ibody_find(inode, &i, is); ++ if (error) ++ goto cleanup; ++ ++ /* Remove the chosen entry from the inode */ ++ error = ext3_xattr_ibody_set(handle, inode, &i, is); ++ ++ entry = IFIRST(header); ++ if (entry_size + EXT3_XATTR_SIZE(size) >= new_extra_isize) ++ shift_bytes = new_extra_isize; ++ else ++ shift_bytes = entry_size + size; ++ /* Adjust the offsets and shift the remaining entries ahead */ ++ ext3_xattr_shift_entries(entry, EXT3_I(inode)->i_extra_isize - ++ shift_bytes, (void *)raw_inode + ++ EXT3_GOOD_OLD_INODE_SIZE + extra_isize + shift_bytes, ++ (void *)header, total_ino - entry_size, ++ inode->i_sb->s_blocksize); ++ ++ extra_isize += shift_bytes; ++ new_extra_isize -= shift_bytes; ++ EXT3_I(inode)->i_extra_isize = extra_isize; ++ ++ i.name = b_entry_name; ++ i.value = buffer; ++ i.value_len = cpu_to_le32(size); ++ error = ext3_xattr_block_find(inode, &i, bs); ++ if (error) ++ goto cleanup; ++ ++ /* Add entry which was removed from the inode into the block */ ++ error = ext3_xattr_block_set(handle, inode, &i, bs); ++ if (error) ++ goto cleanup; ++ } ++ ++cleanup: ++ if (b_entry_name) ++ kfree(b_entry_name); ++ if (buffer) ++ kfree(buffer); ++ if (is) { ++ brelse(is->iloc.bh); ++ kfree(is); ++ } ++ if (bs) ++ kfree(bs); ++ brelse(bh); ++ up_write(&EXT3_I(inode)->xattr_sem); ++ return error; ++} ++ ++ ++ + /* + * ext3_xattr_delete_inode() + * +Index: linux-2.6.16-sles10/fs/ext3/xattr.h +=================================================================== +--- linux-2.6.16-sles10.orig/fs/ext3/xattr.h ++++ linux-2.6.16-sles10/fs/ext3/xattr.h +@@ -75,6 +75,9 @@ extern int ext3_xattr_set_handle(handle_ + extern void ext3_xattr_delete_inode(handle_t *, struct inode *); + extern void ext3_xattr_put_super(struct super_block *); + ++int ext3_expand_extra_isize(struct inode *inode, int new_extra_isize, ++ struct ext3_iloc iloc, handle_t *handle); ++ + extern int init_ext3_xattr(void); + extern void exit_ext3_xattr(void); + diff --git a/ldiskfs/kernel_patches/patches/ext3-inode-version-2.6.18-vanilla.patch b/ldiskfs/kernel_patches/patches/ext3-inode-version-2.6.18-vanilla.patch new file mode 100644 index 0000000000000000000000000000000000000000..26f71acf4542fc4bcbdc4cdd79e624a189f66984 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-inode-version-2.6.18-vanilla.patch @@ -0,0 +1,426 @@ +Index: linux-2.6.18/fs/ext3/inode.c +=================================================================== +--- linux-2.6.18.orig/fs/ext3/inode.c ++++ linux-2.6.18/fs/ext3/inode.c +@@ -2703,6 +2703,13 @@ void ext3_read_inode(struct inode * inod + EXT3_INODE_GET_XTIME(i_atime, inode, raw_inode); + EXT3_EINODE_GET_XTIME(i_crtime, ei, raw_inode); + ++ ei->i_fs_version = le32_to_cpu(raw_inode->i_disk_version); ++ if (EXT3_INODE_SIZE(inode->i_sb) > EXT3_GOOD_OLD_INODE_SIZE) { ++ if (EXT3_FITS_IN_INODE(raw_inode, ei, i_version_hi)) ++ ei->i_fs_version |= (__u64)(le32_to_cpu(raw_inode->i_version_hi)) ++ << 32; ++ } ++ + if (S_ISREG(inode->i_mode)) { + inode->i_op = &ext3_file_inode_operations; + inode->i_fop = &ext3_file_operations; +@@ -2841,8 +2848,14 @@ static int ext3_do_update_inode(handle_t + } else for (block = 0; block < EXT3_N_BLOCKS; block++) + raw_inode->i_block[block] = ei->i_data[block]; + +- if (ei->i_extra_isize) ++ raw_inode->i_disk_version = cpu_to_le32(ei->i_fs_version); ++ if (ei->i_extra_isize) { ++ if (EXT3_FITS_IN_INODE(raw_inode, ei, i_version_hi)) { ++ raw_inode->i_version_hi = cpu_to_le32(ei->i_fs_version ++ >> 32); ++ } + raw_inode->i_extra_isize = cpu_to_le16(ei->i_extra_isize); ++ } + + BUFFER_TRACE(bh, "call ext3_journal_dirty_metadata"); + rc = ext3_journal_dirty_metadata(handle, bh); +@@ -3116,10 +3129,32 @@ ext3_reserve_inode_write(handle_t *handl + int ext3_mark_inode_dirty(handle_t *handle, struct inode *inode) + { + struct ext3_iloc iloc; +- int err; ++ int err, ret; ++ static int expand_message; + + might_sleep(); + err = ext3_reserve_inode_write(handle, inode, &iloc); ++ if (EXT3_I(inode)->i_extra_isize < ++ EXT3_SB(inode->i_sb)->s_want_extra_isize && ++ !(EXT3_I(inode)->i_state & EXT3_STATE_NO_EXPAND)) { ++ /* We need extra buffer credits since we may write into EA block ++ * with this same handle */ ++ if ((ext3_journal_extend(handle, ++ EXT3_DATA_TRANS_BLOCKS(inode->i_sb))) == 0) { ++ ret = ext3_expand_extra_isize(inode, ++ EXT3_SB(inode->i_sb)->s_want_extra_isize, ++ iloc, handle); ++ if (ret) { ++ EXT3_I(inode)->i_state |= EXT3_STATE_NO_EXPAND; ++ if (!expand_message) { ++ ext3_warning(inode->i_sb, __FUNCTION__, ++ "Unable to expand inode %lu. Delete some" ++ " EAs or run e2fsck.", inode->i_ino); ++ expand_message = 1; ++ } ++ } ++ } ++ } + if (!err) + err = ext3_mark_iloc_dirty(handle, inode, &iloc); + return err; +Index: linux-2.6.18/include/linux/ext3_fs.h +=================================================================== +--- linux-2.6.18.orig/include/linux/ext3_fs.h ++++ linux-2.6.18/include/linux/ext3_fs.h +@@ -201,6 +201,7 @@ struct ext3_group_desc + #define EXT3_STATE_JDATA 0x00000001 /* journaled data exists */ + #define EXT3_STATE_NEW 0x00000002 /* inode is newly created */ + #define EXT3_STATE_XATTR 0x00000004 /* has in-inode xattrs */ ++#define EXT3_STATE_NO_EXPAND 0x00000008 /* No space for expansion */ + + /* Used to pass group descriptor data when online resize is done */ + struct ext3_new_group_input { +@@ -277,7 +278,7 @@ struct ext3_inode { + __le32 i_flags; /* File flags */ + union { + struct { +- __u32 l_i_reserved1; ++ __u32 l_i_version; + } linux1; + struct { + __u32 h_i_translator; +@@ -322,6 +323,7 @@ struct ext3_inode { + __le32 i_atime_extra; /* extra Access time (nsec << 2 | epoch) */ + __le32 i_crtime; /* File Creation time */ + __le32 i_crtime_extra; /* extra File Creation time (nsec << 2 | epoch) */ ++ __le32 i_version_hi; /* high 32 bits for 64-bit version */ + }; + + #define i_size_high i_dir_acl +@@ -384,6 +386,8 @@ do { \ + raw_inode->xtime ## _extra); \ + } while (0) + ++#define i_disk_version osd1.linux1.l_i_version ++ + #if defined(__KERNEL__) || defined(__linux__) + #define i_reserved1 osd1.linux1.l_i_reserved1 + #define i_frag osd2.linux2.l_i_frag +Index: linux-2.6.18/include/linux/ext3_fs_i.h +=================================================================== +--- linux-2.6.18.orig/include/linux/ext3_fs_i.h ++++ linux-2.6.18/include/linux/ext3_fs_i.h +@@ -21,6 +21,8 @@ + #include <linux/seqlock.h> + #include <linux/mutex.h> + ++#define HAVE_DISK_INODE_VERSION ++ + /* data type for block offset of block group */ + typedef int ext3_grpblk_t; + +@@ -147,6 +149,8 @@ struct ext3_inode_info { + struct timespec i_crtime; + + void *i_filterdata; ++ ++ __u64 i_fs_version; + }; + + #endif /* _LINUX_EXT3_FS_I */ +Index: linux-2.6.18/fs/ext3/xattr.c +=================================================================== +--- linux-2.6.18.orig/fs/ext3/xattr.c ++++ linux-2.6.18/fs/ext3/xattr.c +@@ -505,6 +505,20 @@ ext3_xattr_release_block(handle_t *handl + } + } + ++static inline size_t ext3_xattr_free_space(struct ext3_xattr_entry *last, ++ size_t *min_offs, void *base, int *total) ++{ ++ for (; !IS_LAST_ENTRY(last); last = EXT3_XATTR_NEXT(last)) { ++ *total += EXT3_XATTR_LEN(last->e_name_len); ++ if (!last->e_value_block && last->e_value_size) { ++ size_t offs = le16_to_cpu(last->e_value_offs); ++ if (offs < *min_offs) ++ *min_offs = offs; ++ } ++ } ++ return (*min_offs - ((void *)last - base) - sizeof(__u32)); ++} ++ + struct ext3_xattr_info { + int name_index; + const char *name; +@@ -1008,6 +1022,8 @@ ext3_xattr_set_handle(handle_t *handle, + if (!error) { + ext3_xattr_update_super_block(handle, inode->i_sb); + inode->i_ctime = ext3_current_time(inode); ++ if (!value) ++ EXT3_I(inode)->i_state &= ~EXT3_STATE_NO_EXPAND; + error = ext3_mark_iloc_dirty(handle, inode, &is.iloc); + /* + * The bh is consumed by ext3_mark_iloc_dirty, even with +@@ -1060,6 +1076,249 @@ retry: + return error; + } + ++static void ext3_xattr_shift_entries(struct ext3_xattr_entry *entry, ++ int value_offs_shift, void *to, ++ void *from, size_t n, int blocksize) ++{ ++ struct ext3_xattr_entry *last = entry; ++ int new_offs; ++ ++ /* Adjust the value offsets of the entries */ ++ for (; !IS_LAST_ENTRY(last); last = EXT3_XATTR_NEXT(last)) { ++ if (!last->e_value_block && last->e_value_size) { ++ new_offs = le16_to_cpu(last->e_value_offs) + ++ value_offs_shift; ++ BUG_ON(new_offs + le32_to_cpu(last->e_value_size) > ++ blocksize); ++ last->e_value_offs = cpu_to_le16(new_offs); ++ } ++ } ++ /* Shift the entries by n bytes */ ++ memmove(to, from, n); ++} ++ ++/* Expand an inode by new_extra_isize bytes. ++ * Returns 0 on success or negative error number on failure. ++ */ ++int ext3_expand_extra_isize(struct inode *inode, int new_extra_isize, ++ struct ext3_iloc iloc, handle_t *handle) ++{ ++ struct ext3_inode *raw_inode; ++ struct ext3_xattr_ibody_header *header; ++ struct ext3_xattr_entry *entry, *last, *first; ++ struct buffer_head *bh = NULL; ++ struct ext3_xattr_ibody_find *is = NULL; ++ struct ext3_xattr_block_find *bs = NULL; ++ char *buffer = NULL, *b_entry_name = NULL; ++ size_t min_offs, free; ++ int total_ino, total_blk; ++ void *base, *start, *end; ++ int extra_isize = 0, error = 0, tried_min_extra_isize = 0; ++ int s_min_extra_isize = EXT3_SB(inode->i_sb)->s_es->s_min_extra_isize; ++ ++ down_write(&EXT3_I(inode)->xattr_sem); ++ ++retry: ++ if (EXT3_I(inode)->i_extra_isize >= new_extra_isize) { ++ up_write(&EXT3_I(inode)->xattr_sem); ++ return 0; ++ } ++ ++ raw_inode = ext3_raw_inode(&iloc); ++ ++ header = IHDR(inode, raw_inode); ++ entry = IFIRST(header); ++ ++ /* No extended attributes present */ ++ if (!(EXT3_I(inode)->i_state & EXT3_STATE_XATTR) || ++ header->h_magic != cpu_to_le32(EXT3_XATTR_MAGIC)) { ++ memset((void *)raw_inode + EXT3_GOOD_OLD_INODE_SIZE, 0, ++ new_extra_isize); ++ EXT3_I(inode)->i_extra_isize = new_extra_isize; ++ goto cleanup; ++ } ++ ++ /* ++ * Check if enough free space is available in the inode to shift the ++ * entries ahead by new_extra_isize. ++ */ ++ ++ base = start = entry; ++ end = (void *)raw_inode + EXT3_SB(inode->i_sb)->s_inode_size; ++ min_offs = end - base; ++ last = entry; ++ total_ino = sizeof(struct ext3_xattr_ibody_header); ++ ++ free = ext3_xattr_free_space(last, &min_offs, base, &total_ino); ++ if (free >= new_extra_isize) { ++ entry = IFIRST(header); ++ ext3_xattr_shift_entries(entry, EXT3_I(inode)->i_extra_isize - ++ new_extra_isize, (void *)raw_inode + ++ EXT3_GOOD_OLD_INODE_SIZE + new_extra_isize, ++ (void *)header, total_ino, ++ inode->i_sb->s_blocksize); ++ EXT3_I(inode)->i_extra_isize = new_extra_isize; ++ error = 0; ++ goto cleanup; ++ } ++ ++ /* ++ * Enough free space isn't available in the inode, check if ++ * EA block can hold new_extra_isize bytes. ++ */ ++ if (EXT3_I(inode)->i_file_acl) { ++ bh = sb_bread(inode->i_sb, EXT3_I(inode)->i_file_acl); ++ error = -EIO; ++ if (!bh) ++ goto cleanup; ++ if (ext3_xattr_check_block(bh)) { ++ ext3_error(inode->i_sb, __FUNCTION__, ++ "inode %lu: bad block "E3FSBLK, inode->i_ino, ++ EXT3_I(inode)->i_file_acl); ++ error = -EIO; ++ goto cleanup; ++ } ++ base = BHDR(bh); ++ first = BFIRST(bh); ++ end = bh->b_data + bh->b_size; ++ min_offs = end - base; ++ free = ext3_xattr_free_space(first, &min_offs, base, ++ &total_blk); ++ if (free < new_extra_isize) { ++ if (!tried_min_extra_isize && s_min_extra_isize) { ++ tried_min_extra_isize++; ++ new_extra_isize = s_min_extra_isize; ++ goto retry; ++ } ++ error = -1; ++ goto cleanup; ++ } ++ } else { ++ free = inode->i_sb->s_blocksize; ++ } ++ ++ while (new_extra_isize > 0) { ++ size_t offs, size, entry_size; ++ struct ext3_xattr_entry *small_entry = NULL; ++ struct ext3_xattr_info i = { ++ .value = NULL, ++ .value_len = 0, ++ }; ++ unsigned int total_size, shift_bytes, temp = ~0U; ++ ++ is = (struct ext3_xattr_ibody_find *) kmalloc(sizeof(struct ++ ext3_xattr_ibody_find), GFP_KERNEL); ++ bs = (struct ext3_xattr_block_find *) kmalloc(sizeof(struct ++ ext3_xattr_block_find), GFP_KERNEL); ++ memset((void *)is, 0, sizeof(struct ext3_xattr_ibody_find)); ++ memset((void *)bs, 0, sizeof(struct ext3_xattr_block_find)); ++ ++ is->s.not_found = bs->s.not_found = -ENODATA; ++ is->iloc.bh = NULL; ++ bs->bh = NULL; ++ ++ last = IFIRST(header); ++ /* Find the entry best suited to be pushed into EA block */ ++ entry = NULL; ++ for (; !IS_LAST_ENTRY(last); last = EXT3_XATTR_NEXT(last)) { ++ total_size = EXT3_XATTR_SIZE(le32_to_cpu(last->e_value_size)) + ++ EXT3_XATTR_LEN(last->e_name_len); ++ if (total_size <= free && total_size < temp) { ++ if (total_size < new_extra_isize) { ++ small_entry = last; ++ } else { ++ entry = last; ++ temp = total_size; ++ } ++ } ++ } ++ ++ if (entry == NULL) { ++ if (small_entry) { ++ entry = small_entry; ++ } else { ++ if (!tried_min_extra_isize && ++ s_min_extra_isize) { ++ tried_min_extra_isize++; ++ new_extra_isize = s_min_extra_isize; ++ goto retry; ++ } ++ error = -1; ++ goto cleanup; ++ } ++ } ++ offs = le16_to_cpu(entry->e_value_offs); ++ size = le32_to_cpu(entry->e_value_size); ++ entry_size = EXT3_XATTR_LEN(entry->e_name_len); ++ i.name_index = entry->e_name_index, ++ buffer = kmalloc(EXT3_XATTR_SIZE(size), GFP_KERNEL); ++ b_entry_name = kmalloc(entry->e_name_len + 1, GFP_KERNEL); ++ /* Save the entry name and the entry value */ ++ memcpy((void *)buffer, (void *)IFIRST(header) + offs, ++ EXT3_XATTR_SIZE(size)); ++ memcpy((void *)b_entry_name, (void *)entry->e_name, ++ entry->e_name_len); ++ b_entry_name[entry->e_name_len] = '\0'; ++ i.name = b_entry_name; ++ ++ error = ext3_get_inode_loc(inode, &is->iloc); ++ if (error) ++ goto cleanup; ++ ++ error = ext3_xattr_ibody_find(inode, &i, is); ++ if (error) ++ goto cleanup; ++ ++ /* Remove the chosen entry from the inode */ ++ error = ext3_xattr_ibody_set(handle, inode, &i, is); ++ ++ entry = IFIRST(header); ++ if (entry_size + EXT3_XATTR_SIZE(size) >= new_extra_isize) ++ shift_bytes = new_extra_isize; ++ else ++ shift_bytes = entry_size + size; ++ /* Adjust the offsets and shift the remaining entries ahead */ ++ ext3_xattr_shift_entries(entry, EXT3_I(inode)->i_extra_isize - ++ shift_bytes, (void *)raw_inode + ++ EXT3_GOOD_OLD_INODE_SIZE + extra_isize + shift_bytes, ++ (void *)header, total_ino - entry_size, ++ inode->i_sb->s_blocksize); ++ ++ extra_isize += shift_bytes; ++ new_extra_isize -= shift_bytes; ++ EXT3_I(inode)->i_extra_isize = extra_isize; ++ ++ i.name = b_entry_name; ++ i.value = buffer; ++ i.value_len = cpu_to_le32(size); ++ error = ext3_xattr_block_find(inode, &i, bs); ++ if (error) ++ goto cleanup; ++ ++ /* Add entry which was removed from the inode into the block */ ++ error = ext3_xattr_block_set(handle, inode, &i, bs); ++ if (error) ++ goto cleanup; ++ } ++ ++cleanup: ++ if (b_entry_name) ++ kfree(b_entry_name); ++ if (buffer) ++ kfree(buffer); ++ if (is) { ++ brelse(is->iloc.bh); ++ kfree(is); ++ } ++ if (bs) ++ kfree(bs); ++ brelse(bh); ++ up_write(&EXT3_I(inode)->xattr_sem); ++ return error; ++} ++ ++ ++ + /* + * ext3_xattr_delete_inode() + * +Index: linux-2.6.18/fs/ext3/xattr.h +=================================================================== +--- linux-2.6.18.orig/fs/ext3/xattr.h ++++ linux-2.6.18/fs/ext3/xattr.h +@@ -74,6 +74,9 @@ extern int ext3_xattr_set_handle(handle_ + extern void ext3_xattr_delete_inode(handle_t *, struct inode *); + extern void ext3_xattr_put_super(struct super_block *); + ++int ext3_expand_extra_isize(struct inode *inode, int new_extra_isize, ++ struct ext3_iloc iloc, handle_t *handle); ++ + extern int init_ext3_xattr(void); + extern void exit_ext3_xattr(void); + diff --git a/ldiskfs/kernel_patches/patches/ext3-mballoc2-2.6-suse.patch b/ldiskfs/kernel_patches/patches/ext3-mballoc2-2.6-suse.patch deleted file mode 100644 index cec18779bd6b6235ce7ca21b9c039af494b1827e..0000000000000000000000000000000000000000 --- a/ldiskfs/kernel_patches/patches/ext3-mballoc2-2.6-suse.patch +++ /dev/null @@ -1,3108 +0,0 @@ -Index: linux-2.6.5-7.282-full/include/linux/ext3_fs.h -=================================================================== ---- linux-2.6.5-7.282-full.orig/include/linux/ext3_fs.h 2006-10-24 22:18:28.000000000 +0400 -+++ linux-2.6.5-7.282-full/include/linux/ext3_fs.h 2006-10-24 22:18:28.000000000 +0400 -@@ -57,6 +57,14 @@ struct statfs; - #define ext3_debug(f, a...) do {} while (0) - #endif - -+#define EXT3_MULTIBLOCK_ALLOCATOR 1 -+ -+#define EXT3_MB_HINT_MERGE 1 -+#define EXT3_MB_HINT_RESERVED 2 -+#define EXT3_MB_HINT_METADATA 4 -+#define EXT3_MB_HINT_FIRST 8 -+#define EXT3_MB_HINT_BEST 16 -+ - /* - * Special inodes numbers - */ -@@ -339,6 +347,7 @@ struct ext3_inode { - #define EXT3_MOUNT_IOPEN_NOPRIV 0x100000/* Make iopen world-readable */ - #define EXT3_MOUNT_EXTENTS 0x200000/* Extents support */ - #define EXT3_MOUNT_EXTDEBUG 0x400000/* Extents debug */ -+#define EXT3_MOUNT_MBALLOC 0x800000/* Buddy allocation support */ - - /* Compatibility, for having both ext2_fs.h and ext3_fs.h included at once */ - #ifndef clear_opt -@@ -361,6 +370,14 @@ struct ext3_inode { - #define ext3_find_first_zero_bit ext2_find_first_zero_bit - #define ext3_find_next_zero_bit ext2_find_next_zero_bit - -+#ifndef ext2_find_next_le_bit -+#ifdef __LITTLE_ENDIAN -+#define ext2_find_next_le_bit(addr, size, off) find_next_bit((addr), (size), (off)) -+#else -+#error "mballoc needs a patch for big-endian systems - CFS bug 10634" -+#endif /* __LITTLE_ENDIAN */ -+#endif /* !ext2_find_next_le_bit */ -+ - /* - * Maximal mount counts between two filesystem checks - */ -@@ -700,7 +717,10 @@ extern int ext3_bg_has_super(struct supe - extern unsigned long ext3_bg_num_gdb(struct super_block *sb, int group); - extern int ext3_new_block (handle_t *, struct inode *, unsigned long, int *); - extern void ext3_free_blocks (handle_t *, struct inode *, unsigned long, -- unsigned long); -+ unsigned long, int); -+extern int ext3_new_block_old(handle_t *, struct inode *, unsigned long, int *); -+extern void ext3_free_blocks_old(handle_t *, struct inode *, unsigned long, -+ unsigned long); - extern unsigned long ext3_count_free_blocks (struct super_block *); - extern void ext3_check_blocks_bitmap (struct super_block *); - extern struct ext3_group_desc * ext3_get_group_desc(struct super_block * sb, -@@ -824,6 +843,17 @@ extern void ext3_extents_initialize_bloc - extern int ext3_ext_ioctl(struct inode *inode, struct file *filp, - unsigned int cmd, unsigned long arg); - -+/* mballoc.c */ -+extern long ext3_mb_stats; -+extern long ext3_mb_max_to_scan; -+extern int ext3_mb_init(struct super_block *, int); -+extern int ext3_mb_release(struct super_block *); -+extern int ext3_mb_new_blocks(handle_t *, struct inode *, unsigned long, int *, int, int *); -+extern int ext3_mb_reserve_blocks(struct super_block *, int); -+extern void ext3_mb_release_blocks(struct super_block *, int); -+int __init init_ext3_proc(void); -+void exit_ext3_proc(void); -+ - #endif /* __KERNEL__ */ - - #define EXT3_IOC_CREATE_INUM _IOW('f', 5, long) -Index: linux-2.6.5-7.282-full/include/linux/ext3_fs_sb.h -=================================================================== ---- linux-2.6.5-7.282-full.orig/include/linux/ext3_fs_sb.h 2006-10-24 22:18:28.000000000 +0400 -+++ linux-2.6.5-7.282-full/include/linux/ext3_fs_sb.h 2006-10-24 22:18:28.000000000 +0400 -@@ -23,9 +23,15 @@ - #define EXT_INCLUDE - #include <linux/blockgroup_lock.h> - #include <linux/percpu_counter.h> -+#include <linux/list.h> - #endif - #endif - #include <linux/rbtree.h> -+#include <linux/proc_fs.h> -+ -+struct ext3_buddy_group_blocks; -+struct ext3_mb_history; -+#define EXT3_BB_MAX_BLOCKS - - /* - * third extended-fs super-block data in memory -@@ -78,6 +84,43 @@ struct ext3_sb_info { - struct timer_list turn_ro_timer; /* For turning read-only (crash simulation) */ - wait_queue_head_t ro_wait_queue; /* For people waiting for the fs to go read-only */ - #endif -+ -+ /* for buddy allocator */ -+ struct ext3_group_info ***s_group_info; -+ struct inode *s_buddy_cache; -+ long s_blocks_reserved; -+ spinlock_t s_reserve_lock; -+ struct list_head s_active_transaction; -+ struct list_head s_closed_transaction; -+ struct list_head s_committed_transaction; -+ spinlock_t s_md_lock; -+ tid_t s_last_transaction; -+ int s_mb_factor; -+ unsigned short *s_mb_offsets, *s_mb_maxs; -+ unsigned long s_stripe; -+ -+ /* history to debug policy */ -+ struct ext3_mb_history *s_mb_history; -+ int s_mb_history_cur; -+ int s_mb_history_max; -+ struct proc_dir_entry *s_mb_proc; -+ spinlock_t s_mb_history_lock; -+ -+ /* stats for buddy allocator */ -+ atomic_t s_bal_reqs; /* number of reqs with len > 1 */ -+ atomic_t s_bal_success; /* we found long enough chunks */ -+ atomic_t s_bal_allocated; /* in blocks */ -+ atomic_t s_bal_ex_scanned; /* total extents scanned */ -+ atomic_t s_bal_goals; /* goal hits */ -+ atomic_t s_bal_breaks; /* too long searches */ -+ atomic_t s_bal_2orders; /* 2^order hits */ -+ spinlock_t s_bal_lock; -+ unsigned long s_mb_buddies_generated; -+ unsigned long long s_mb_generation_time; - }; - -+#define EXT3_GROUP_INFO(sb, group) \ -+ EXT3_SB(sb)->s_group_info[(group) >> EXT3_DESC_PER_BLOCK_BITS(sb)] \ -+ [(group) & (EXT3_DESC_PER_BLOCK(sb) - 1)] -+ - #endif /* _LINUX_EXT3_FS_SB */ -Index: linux-2.6.5-7.282-full/fs/ext3/super.c -=================================================================== ---- linux-2.6.5-7.282-full.orig/fs/ext3/super.c 2006-10-24 22:18:28.000000000 +0400 -+++ linux-2.6.5-7.282-full/fs/ext3/super.c 2006-10-24 22:18:28.000000000 +0400 -@@ -389,6 +389,7 @@ void ext3_put_super (struct super_block - struct ext3_super_block *es = sbi->s_es; - int i; - -+ ext3_mb_release(sb); - ext3_ext_release(sb); - ext3_xattr_put_super(sb); - journal_destroy(sbi->s_journal); -@@ -588,6 +589,7 @@ enum { - Opt_err, - Opt_iopen, Opt_noiopen, Opt_iopen_nopriv, - Opt_extents, Opt_noextents, Opt_extdebug, -+ Opt_mballoc, Opt_nomballoc, Opt_stripe, - }; - - static match_table_t tokens = { -@@ -634,6 +636,9 @@ static match_table_t tokens = { - {Opt_extents, "extents"}, - {Opt_noextents, "noextents"}, - {Opt_extdebug, "extdebug"}, -+ {Opt_mballoc, "mballoc"}, -+ {Opt_nomballoc, "nomballoc"}, -+ {Opt_stripe, "stripe=%u"}, - {Opt_barrier, "barrier=%u"}, - {Opt_err, NULL} - }; -@@ -859,6 +864,19 @@ static int parse_options (char * options - case Opt_extdebug: - set_opt (sbi->s_mount_opt, EXTDEBUG); - break; -+ case Opt_mballoc: -+ set_opt(sbi->s_mount_opt, MBALLOC); -+ break; -+ case Opt_nomballoc: -+ clear_opt(sbi->s_mount_opt, MBALLOC); -+ break; -+ case Opt_stripe: -+ if (match_int(&args[0], &option)) -+ return 0; -+ if (option < 0) -+ return 0; -+ sbi->s_stripe = option; -+ break; - default: - printk (KERN_ERR - "EXT3-fs: Unrecognized mount option \"%s\" " -@@ -1512,6 +1530,7 @@ static int ext3_fill_super (struct super - ext3_count_dirs(sb)); - - ext3_ext_init(sb); -+ ext3_mb_init(sb, needs_recovery); - - return 0; - -@@ -2160,7 +2179,13 @@ static struct file_system_type ext3_fs_t - - static int __init init_ext3_fs(void) - { -- int err = init_ext3_xattr(); -+ int err; -+ -+ err = init_ext3_proc(); -+ if (err) -+ return err; -+ -+ err = init_ext3_xattr(); - if (err) - return err; - err = init_inodecache(); -@@ -2189,6 +2214,7 @@ static void __exit exit_ext3_fs(void) - unregister_filesystem(&ext3_fs_type); - destroy_inodecache(); - exit_ext3_xattr(); -+ exit_ext3_proc(); - } - - int ext3_prep_san_write(struct inode *inode, long *blocks, -Index: linux-2.6.5-7.282-full/fs/ext3/extents.c -=================================================================== ---- linux-2.6.5-7.282-full.orig/fs/ext3/extents.c 2006-10-24 22:18:28.000000000 +0400 -+++ linux-2.6.5-7.282-full/fs/ext3/extents.c 2006-10-24 22:18:28.000000000 +0400 -@@ -779,7 +779,7 @@ cleanup: - for (i = 0; i < depth; i++) { - if (!ablocks[i]) - continue; -- ext3_free_blocks(handle, tree->inode, ablocks[i], 1); -+ ext3_free_blocks(handle, tree->inode, ablocks[i], 1, 1); - } - } - kfree(ablocks); -@@ -1438,7 +1438,7 @@ int ext3_ext_rm_idx(handle_t *handle, st - path->p_idx->ei_leaf); - bh = sb_find_get_block(tree->inode->i_sb, path->p_idx->ei_leaf); - ext3_forget(handle, 1, tree->inode, bh, path->p_idx->ei_leaf); -- ext3_free_blocks(handle, tree->inode, path->p_idx->ei_leaf, 1); -+ ext3_free_blocks(handle, tree->inode, path->p_idx->ei_leaf, 1, 1); - return err; - } - -@@ -1923,10 +1923,12 @@ ext3_remove_blocks(struct ext3_extents_t - int needed = ext3_remove_blocks_credits(tree, ex, from, to); - handle_t *handle = ext3_journal_start(tree->inode, needed); - struct buffer_head *bh; -- int i; -+ int i, metadata = 0; - - if (IS_ERR(handle)) - return PTR_ERR(handle); -+ if (S_ISDIR(tree->inode->i_mode) || S_ISLNK(tree->inode->i_mode)) -+ metadata = 1; - if (from >= ex->ee_block && to == ex->ee_block + ex->ee_len - 1) { - /* tail removal */ - unsigned long num, start; -@@ -1938,7 +1940,7 @@ ext3_remove_blocks(struct ext3_extents_t - bh = sb_find_get_block(tree->inode->i_sb, start + i); - ext3_forget(handle, 0, tree->inode, bh, start + i); - } -- ext3_free_blocks(handle, tree->inode, start, num); -+ ext3_free_blocks(handle, tree->inode, start, num, metadata); - } else if (from == ex->ee_block && to <= ex->ee_block + ex->ee_len - 1) { - printk("strange request: removal %lu-%lu from %u:%u\n", - from, to, ex->ee_block, ex->ee_len); -Index: linux-2.6.5-7.282-full/fs/ext3/inode.c -=================================================================== ---- linux-2.6.5-7.282-full.orig/fs/ext3/inode.c 2006-10-24 22:18:28.000000000 +0400 -+++ linux-2.6.5-7.282-full/fs/ext3/inode.c 2006-10-24 22:18:28.000000000 +0400 -@@ -574,7 +574,7 @@ static int ext3_alloc_branch(handle_t *h - ext3_journal_forget(handle, branch[i].bh); - } - for (i = 0; i < keys; i++) -- ext3_free_blocks(handle, inode, le32_to_cpu(branch[i].key), 1); -+ ext3_free_blocks(handle, inode, le32_to_cpu(branch[i].key), 1, 1); - return err; - } - -@@ -675,7 +675,7 @@ err_out: - if (err == -EAGAIN) - for (i = 0; i < num; i++) - ext3_free_blocks(handle, inode, -- le32_to_cpu(where[i].key), 1); -+ le32_to_cpu(where[i].key), 1, 1); - return err; - } - -@@ -1837,7 +1837,7 @@ ext3_clear_blocks(handle_t *handle, stru - } - } - -- ext3_free_blocks(handle, inode, block_to_free, count); -+ ext3_free_blocks(handle, inode, block_to_free, count, 1); - } - - /** -@@ -2008,7 +2008,7 @@ static void ext3_free_branches(handle_t - ext3_journal_test_restart(handle, inode); - } - -- ext3_free_blocks(handle, inode, nr, 1); -+ ext3_free_blocks(handle, inode, nr, 1, 1); - - if (parent_bh) { - /* -Index: linux-2.6.5-7.282-full/fs/ext3/balloc.c -=================================================================== ---- linux-2.6.5-7.282-full.orig/fs/ext3/balloc.c 2006-08-30 18:12:13.000000000 +0400 -+++ linux-2.6.5-7.282-full/fs/ext3/balloc.c 2006-10-24 22:18:28.000000000 +0400 -@@ -78,7 +78,7 @@ struct ext3_group_desc * ext3_get_group_ - * - * Return buffer_head on success or NULL in case of failure. - */ --static struct buffer_head * -+struct buffer_head * - read_block_bitmap(struct super_block *sb, unsigned int block_group) - { - struct ext3_group_desc * desc; -@@ -274,7 +274,7 @@ void ext3_discard_reservation(struct ino - } - - /* Free given blocks, update quota and i_blocks field */ --void ext3_free_blocks(handle_t *handle, struct inode *inode, -+void ext3_free_blocks_old(handle_t *handle, struct inode *inode, - unsigned long block, unsigned long count) - { - struct buffer_head *bitmap_bh = NULL; -@@ -1142,7 +1142,7 @@ int ext3_should_retry_alloc(struct super - * bitmap, and then for any free bit if that fails. - * This function also updates quota and i_blocks field. - */ --int ext3_new_block(handle_t *handle, struct inode *inode, -+int ext3_new_block_old(handle_t *handle, struct inode *inode, - unsigned long goal, int *errp) - { - struct buffer_head *bitmap_bh = NULL; -Index: linux-2.6.5-7.282-full/fs/ext3/xattr.c -=================================================================== ---- linux-2.6.5-7.282-full.orig/fs/ext3/xattr.c 2006-10-24 22:18:28.000000000 +0400 -+++ linux-2.6.5-7.282-full/fs/ext3/xattr.c 2006-10-24 22:18:28.000000000 +0400 -@@ -1371,7 +1371,7 @@ ext3_xattr_set_handle2(handle_t *handle, - new_bh = sb_getblk(sb, block); - if (!new_bh) { - getblk_failed: -- ext3_free_blocks(handle, inode, block, 1); -+ ext3_free_blocks(handle, inode, block, 1, 1); - error = -EIO; - goto cleanup; - } -@@ -1411,7 +1411,7 @@ getblk_failed: - if (HDR(old_bh)->h_refcount == cpu_to_le32(1)) { - /* Free the old block. */ - ea_bdebug(old_bh, "freeing"); -- ext3_free_blocks(handle, inode, old_bh->b_blocknr, 1); -+ ext3_free_blocks(handle, inode, old_bh->b_blocknr, 1, 1); - - /* ext3_forget() calls bforget() for us, but we - let our caller release old_bh, so we need to -@@ -1519,7 +1519,7 @@ ext3_xattr_delete_inode(handle_t *handle - mb_cache_entry_free(ce); - ce = NULL; - } -- ext3_free_blocks(handle, inode, EXT3_I(inode)->i_file_acl, 1); -+ ext3_free_blocks(handle, inode, EXT3_I(inode)->i_file_acl, 1, 1); - get_bh(bh); - ext3_forget(handle, 1, inode, bh, EXT3_I(inode)->i_file_acl); - } else { -Index: linux-2.6.5-7.282-full/fs/ext3/mballoc.c -=================================================================== ---- linux-2.6.5-7.282-full.orig/fs/ext3/mballoc.c 2006-10-23 18:07:54.821533176 +0400 -+++ linux-2.6.5-7.282-full/fs/ext3/mballoc.c 2006-10-24 22:20:45.000000000 +0400 -@@ -0,0 +1,2726 @@ -+/* -+ * Copyright (c) 2003-2005, Cluster File Systems, Inc, info@clusterfs.com -+ * Written by Alex Tomas <alex@clusterfs.com> -+ * -+ * This program is free software; you can redistribute it and/or modify -+ * it under the terms of the GNU General Public License version 2 as -+ * published by the Free Software Foundation. -+ * -+ * This program is distributed in the hope that it will be useful, -+ * but WITHOUT ANY WARRANTY; without even the implied warranty of -+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -+ * GNU General Public License for more details. -+ * -+ * You should have received a copy of the GNU General Public Licens -+ * along with this program; if not, write to the Free Software -+ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111- -+ */ -+ -+ -+/* -+ * mballoc.c contains the multiblocks allocation routines -+ */ -+ -+#include <linux/config.h> -+#include <linux/time.h> -+#include <linux/fs.h> -+#include <linux/namei.h> -+#include <linux/jbd.h> -+#include <linux/ext3_fs.h> -+#include <linux/ext3_jbd.h> -+#include <linux/quotaops.h> -+#include <linux/buffer_head.h> -+#include <linux/module.h> -+#include <linux/swap.h> -+#include <linux/proc_fs.h> -+#include <linux/pagemap.h> -+#include <linux/seq_file.h> -+ -+/* -+ * TODO: -+ * - bitmap read-ahead (proposed by Oleg Drokin aka green) -+ * - track min/max extents in each group for better group selection -+ * - mb_mark_used() may allocate chunk right after splitting buddy -+ * - special flag to advice allocator to look for requested + N blocks -+ * this may improve interaction between extents and mballoc -+ * - tree of groups sorted by number of free blocks -+ * - percpu reservation code (hotpath) -+ * - error handling -+ */ -+ -+/* -+ * with AGRESSIVE_CHECK allocator runs consistency checks over -+ * structures. these checks slow things down a lot -+ */ -+#define AGGRESSIVE_CHECK__ -+ -+/* -+ */ -+#define MB_DEBUG__ -+#ifdef MB_DEBUG -+#define mb_debug(fmt,a...) printk(fmt, ##a) -+#else -+#define mb_debug(fmt,a...) -+#endif -+ -+/* -+ * with EXT3_MB_HISTORY mballoc stores last N allocations in memory -+ * and you can monitor it in /proc/fs/ext3/<dev>/mb_history -+ */ -+#define EXT3_MB_HISTORY -+ -+/* -+ * How long mballoc can look for a best extent (in found extents) -+ */ -+long ext3_mb_max_to_scan = 500; -+ -+/* -+ * How long mballoc must look for a best extent -+ */ -+long ext3_mb_min_to_scan = 30; -+ -+/* -+ * with 'ext3_mb_stats' allocator will collect stats that will be -+ * shown at umount. The collecting costs though! -+ */ -+ -+long ext3_mb_stats = 1; -+ -+/* -+ * for which requests use 2^N search using buddies -+ */ -+long ext3_mb_order2_reqs = 8; -+ -+#ifdef EXT3_BB_MAX_BLOCKS -+#undef EXT3_BB_MAX_BLOCKS -+#endif -+#define EXT3_BB_MAX_BLOCKS 30 -+ -+struct ext3_free_metadata { -+ unsigned short group; -+ unsigned short num; -+ unsigned short blocks[EXT3_BB_MAX_BLOCKS]; -+ struct list_head list; -+}; -+ -+struct ext3_group_info { -+ unsigned long bb_state; -+ unsigned long bb_tid; -+ struct ext3_free_metadata *bb_md_cur; -+ unsigned short bb_first_free; -+ unsigned short bb_free; -+ unsigned short bb_fragments; -+ unsigned short bb_counters[]; -+}; -+ -+ -+#define EXT3_GROUP_INFO_NEED_INIT_BIT 0 -+#define EXT3_GROUP_INFO_LOCKED_BIT 1 -+ -+#define EXT3_MB_GRP_NEED_INIT(grp) \ -+ (test_bit(EXT3_GROUP_INFO_NEED_INIT_BIT, &(grp)->bb_state)) -+ -+struct ext3_free_extent { -+ __u16 fe_start; -+ __u16 fe_len; -+ __u16 fe_group; -+}; -+ -+struct ext3_allocation_context { -+ struct super_block *ac_sb; -+ -+ /* search goals */ -+ struct ext3_free_extent ac_g_ex; -+ -+ /* the best found extent */ -+ struct ext3_free_extent ac_b_ex; -+ -+ /* number of iterations done. we have to track to limit searching */ -+ unsigned long ac_ex_scanned; -+ __u16 ac_groups_scanned; -+ __u16 ac_found; -+ __u16 ac_tail; -+ __u16 ac_buddy; -+ __u8 ac_status; -+ __u8 ac_flags; /* allocation hints */ -+ __u8 ac_criteria; -+ __u8 ac_repeats; -+ __u8 ac_2order; /* if request is to allocate 2^N blocks and -+ * N > 0, the field stores N, otherwise 0 */ -+ -+ struct page *ac_buddy_page; -+ struct page *ac_bitmap_page; -+}; -+ -+#define AC_STATUS_CONTINUE 1 -+#define AC_STATUS_FOUND 2 -+#define AC_STATUS_BREAK 3 -+ -+struct ext3_mb_history { -+ struct ext3_free_extent goal; /* goal allocation */ -+ struct ext3_free_extent result; /* result allocation */ -+ unsigned pid; -+ unsigned ino; -+ __u16 found; /* how many extents have been found */ -+ __u16 groups; /* how many groups have been scanned */ -+ __u16 tail; /* what tail broke some buddy */ -+ __u16 buddy; /* buddy the tail ^^^ broke */ -+ __u8 cr; /* which phase the result extent was found at */ -+ __u8 merged; -+}; -+ -+struct ext3_buddy { -+ struct page *bd_buddy_page; -+ void *bd_buddy; -+ struct page *bd_bitmap_page; -+ void *bd_bitmap; -+ struct ext3_group_info *bd_info; -+ struct super_block *bd_sb; -+ __u16 bd_blkbits; -+ __u16 bd_group; -+}; -+#define EXT3_MB_BITMAP(e3b) ((e3b)->bd_bitmap) -+#define EXT3_MB_BUDDY(e3b) ((e3b)->bd_buddy) -+ -+#ifndef EXT3_MB_HISTORY -+#define ext3_mb_store_history(sb,ino,ac) -+#else -+static void ext3_mb_store_history(struct super_block *, unsigned ino, -+ struct ext3_allocation_context *ac); -+#endif -+ -+#define in_range(b, first, len) ((b) >= (first) && (b) <= (first) + (len) - 1) -+ -+static struct proc_dir_entry *proc_root_ext3; -+ -+struct buffer_head * read_block_bitmap(struct super_block *, unsigned int); -+void ext3_mb_poll_new_transaction(struct super_block *, handle_t *); -+void ext3_mb_free_committed_blocks(struct super_block *); -+ -+#if BITS_PER_LONG == 64 -+#define mb_correct_addr_and_bit(bit,addr) \ -+{ \ -+ bit += ((unsigned long) addr & 7UL) << 3; \ -+ addr = (void *) ((unsigned long) addr & ~7UL); \ -+} -+#elif BITS_PER_LONG == 32 -+#define mb_correct_addr_and_bit(bit,addr) \ -+{ \ -+ bit += ((unsigned long) addr & 3UL) << 3; \ -+ addr = (void *) ((unsigned long) addr & ~3UL); \ -+} -+#else -+#error "how many bits you are?!" -+#endif -+ -+static inline int mb_test_bit(int bit, void *addr) -+{ -+ mb_correct_addr_and_bit(bit,addr); -+ return ext2_test_bit(bit, addr); -+} -+ -+static inline void mb_set_bit(int bit, void *addr) -+{ -+ mb_correct_addr_and_bit(bit,addr); -+ ext2_set_bit(bit, addr); -+} -+ -+static inline void mb_set_bit_atomic(int bit, void *addr) -+{ -+ mb_correct_addr_and_bit(bit,addr); -+ ext2_set_bit_atomic(NULL, bit, addr); -+} -+ -+static inline void mb_clear_bit(int bit, void *addr) -+{ -+ mb_correct_addr_and_bit(bit,addr); -+ ext2_clear_bit(bit, addr); -+} -+ -+static inline void mb_clear_bit_atomic(int bit, void *addr) -+{ -+ mb_correct_addr_and_bit(bit,addr); -+ ext2_clear_bit_atomic(NULL, bit, addr); -+} -+ -+static inline int mb_find_next_zero_bit(void *addr, int max, int start) -+{ -+ int fix; -+#if BITS_PER_LONG == 64 -+ fix = ((unsigned long) addr & 7UL) << 3; -+ addr = (void *) ((unsigned long) addr & ~7UL); -+#elif BITS_PER_LONG == 32 -+ fix = ((unsigned long) addr & 3UL) << 3; -+ addr = (void *) ((unsigned long) addr & ~3UL); -+#else -+#error "how many bits you are?!" -+#endif -+ max += fix; -+ start += fix; -+ return ext2_find_next_zero_bit(addr, max, start) - fix; -+} -+ -+static inline void *mb_find_buddy(struct ext3_buddy *e3b, int order, int *max) -+{ -+ char *bb; -+ -+ J_ASSERT(EXT3_MB_BITMAP(e3b) != EXT3_MB_BUDDY(e3b)); -+ J_ASSERT(max != NULL); -+ -+ if (order > e3b->bd_blkbits + 1) { -+ *max = 0; -+ return NULL; -+ } -+ -+ /* at order 0 we see each particular block */ -+ *max = 1 << (e3b->bd_blkbits + 3); -+ if (order == 0) -+ return EXT3_MB_BITMAP(e3b); -+ -+ bb = EXT3_MB_BUDDY(e3b) + EXT3_SB(e3b->bd_sb)->s_mb_offsets[order]; -+ *max = EXT3_SB(e3b->bd_sb)->s_mb_maxs[order]; -+ -+ return bb; -+} -+ -+#ifdef AGGRESSIVE_CHECK -+ -+static void mb_check_buddy(struct ext3_buddy *e3b) -+{ -+ int order = e3b->bd_blkbits + 1; -+ int max, max2, i, j, k, count; -+ int fragments = 0, fstart; -+ void *buddy, *buddy2; -+ -+ if (!test_opt(e3b->bd_sb, MBALLOC)) -+ return; -+ -+ { -+ static int mb_check_counter = 0; -+ if (mb_check_counter++ % 300 != 0) -+ return; -+ } -+ -+ while (order > 1) { -+ buddy = mb_find_buddy(e3b, order, &max); -+ J_ASSERT(buddy); -+ buddy2 = mb_find_buddy(e3b, order - 1, &max2); -+ J_ASSERT(buddy2); -+ J_ASSERT(buddy != buddy2); -+ J_ASSERT(max * 2 == max2); -+ -+ count = 0; -+ for (i = 0; i < max; i++) { -+ -+ if (mb_test_bit(i, buddy)) { -+ /* only single bit in buddy2 may be 1 */ -+ if (!mb_test_bit(i << 1, buddy2)) -+ J_ASSERT(mb_test_bit((i<<1)+1, buddy2)); -+ else if (!mb_test_bit((i << 1) + 1, buddy2)) -+ J_ASSERT(mb_test_bit(i << 1, buddy2)); -+ continue; -+ } -+ -+ /* both bits in buddy2 must be 0 */ -+ J_ASSERT(mb_test_bit(i << 1, buddy2)); -+ J_ASSERT(mb_test_bit((i << 1) + 1, buddy2)); -+ -+ for (j = 0; j < (1 << order); j++) { -+ k = (i * (1 << order)) + j; -+ J_ASSERT(!mb_test_bit(k, EXT3_MB_BITMAP(e3b))); -+ } -+ count++; -+ } -+ J_ASSERT(e3b->bd_info->bb_counters[order] == count); -+ order--; -+ } -+ -+ fstart = -1; -+ buddy = mb_find_buddy(e3b, 0, &max); -+ for (i = 0; i < max; i++) { -+ if (!mb_test_bit(i, buddy)) { -+ J_ASSERT(i >= e3b->bd_info->bb_first_free); -+ if (fstart == -1) { -+ fragments++; -+ fstart = i; -+ } -+ continue; -+ } -+ fstart = -1; -+ /* check used bits only */ -+ for (j = 0; j < e3b->bd_blkbits + 1; j++) { -+ buddy2 = mb_find_buddy(e3b, j, &max2); -+ k = i >> j; -+ J_ASSERT(k < max2); -+ J_ASSERT(mb_test_bit(k, buddy2)); -+ } -+ } -+ J_ASSERT(!EXT3_MB_GRP_NEED_INIT(e3b->bd_info)); -+ J_ASSERT(e3b->bd_info->bb_fragments == fragments); -+} -+ -+#else -+#define mb_check_buddy(e3b) -+#endif -+ -+/* find most significant bit */ -+static int inline fmsb(unsigned short word) -+{ -+ int order; -+ -+ if (word > 255) { -+ order = 7; -+ word >>= 8; -+ } else { -+ order = -1; -+ } -+ -+ do { -+ order++; -+ word >>= 1; -+ } while (word != 0); -+ -+ return order; -+} -+ -+static void inline -+ext3_mb_mark_free_simple(struct super_block *sb, void *buddy, unsigned first, -+ int len, struct ext3_group_info *grp) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ unsigned short min, max, chunk, border; -+ -+ mb_debug("mark %u/%u free\n", first, len); -+ J_ASSERT(len < EXT3_BLOCKS_PER_GROUP(sb)); -+ -+ border = 2 << sb->s_blocksize_bits; -+ -+ while (len > 0) { -+ /* find how many blocks can be covered since this position */ -+ max = ffs(first | border) - 1; -+ -+ /* find how many blocks of power 2 we need to mark */ -+ min = fmsb(len); -+ -+ mb_debug(" %u/%u -> max %u, min %u\n", -+ first & ((2 << sb->s_blocksize_bits) - 1), -+ len, max, min); -+ -+ if (max < min) -+ min = max; -+ chunk = 1 << min; -+ -+ /* mark multiblock chunks only */ -+ grp->bb_counters[min]++; -+ if (min > 0) { -+ mb_debug(" set %u at %u \n", first >> min, -+ sbi->s_mb_offsets[min]); -+ mb_clear_bit(first >> min, buddy + sbi->s_mb_offsets[min]); -+ } -+ -+ len -= chunk; -+ first += chunk; -+ } -+} -+ -+static void -+ext3_mb_generate_buddy(struct super_block *sb, void *buddy, void *bitmap, -+ int group) -+{ -+ struct ext3_group_info *grp = EXT3_GROUP_INFO(sb, group); -+ unsigned short max = EXT3_BLOCKS_PER_GROUP(sb); -+ unsigned short i = 0, first, len; -+ unsigned free = 0, fragments = 0; -+ unsigned long long period = get_cycles(); -+ -+ i = mb_find_next_zero_bit(bitmap, max, 0); -+ grp->bb_first_free = i; -+ while (i < max) { -+ fragments++; -+ first = i; -+ i = ext2_find_next_le_bit(bitmap, max, i); -+ len = i - first; -+ free += len; -+ if (len > 1) -+ ext3_mb_mark_free_simple(sb, buddy, first, len, grp); -+ else -+ grp->bb_counters[0]++; -+ if (i < max) -+ i = mb_find_next_zero_bit(bitmap, max, i); -+ } -+ grp->bb_fragments = fragments; -+ -+ /* bb_state shouldn't being modified because all -+ * others waits for init completion on page lock */ -+ clear_bit(EXT3_GROUP_INFO_NEED_INIT_BIT, &grp->bb_state); -+ if (free != grp->bb_free) { -+ printk("EXT3-fs: group %u: %u blocks in bitmap, %u in gd\n", -+ group, free, grp->bb_free); -+ grp->bb_free = free; -+ } -+ -+ period = get_cycles() - period; -+ spin_lock(&EXT3_SB(sb)->s_bal_lock); -+ EXT3_SB(sb)->s_mb_buddies_generated++; -+ EXT3_SB(sb)->s_mb_generation_time += period; -+ spin_unlock(&EXT3_SB(sb)->s_bal_lock); -+} -+ -+static int ext3_mb_init_cache(struct page *page) -+{ -+ int blocksize, blocks_per_page, groups_per_page; -+ int err = 0, i, first_group, first_block; -+ struct super_block *sb; -+ struct buffer_head *bhs; -+ struct buffer_head **bh; -+ struct inode *inode; -+ char *data, *bitmap; -+ -+ mb_debug("init page %lu\n", page->index); -+ -+ inode = page->mapping->host; -+ sb = inode->i_sb; -+ blocksize = 1 << inode->i_blkbits; -+ blocks_per_page = PAGE_CACHE_SIZE / blocksize; -+ -+ groups_per_page = blocks_per_page >> 1; -+ if (groups_per_page == 0) -+ groups_per_page = 1; -+ -+ /* allocate buffer_heads to read bitmaps */ -+ if (groups_per_page > 1) { -+ err = -ENOMEM; -+ i = sizeof(struct buffer_head *) * groups_per_page; -+ bh = kmalloc(i, GFP_NOFS); -+ if (bh == NULL) -+ goto out; -+ memset(bh, 0, i); -+ } else -+ bh = &bhs; -+ -+ first_group = page->index * blocks_per_page / 2; -+ -+ /* read all groups the page covers into the cache */ -+ for (i = 0; i < groups_per_page; i++) { -+ struct ext3_group_desc * desc; -+ -+ if (first_group + i >= EXT3_SB(sb)->s_groups_count) -+ break; -+ -+ err = -EIO; -+ desc = ext3_get_group_desc(sb, first_group + i, NULL); -+ if (desc == NULL) -+ goto out; -+ -+ err = -ENOMEM; -+ bh[i] = sb_getblk(sb, le32_to_cpu(desc->bg_block_bitmap)); -+ if (bh[i] == NULL) -+ goto out; -+ -+ if (buffer_uptodate(bh[i])) -+ continue; -+ -+ lock_buffer(bh[i]); -+ if (buffer_uptodate(bh[i])) { -+ unlock_buffer(bh[i]); -+ continue; -+ } -+ -+ get_bh(bh[i]); -+ bh[i]->b_end_io = end_buffer_read_sync; -+ submit_bh(READ, bh[i]); -+ mb_debug("read bitmap for group %u\n", first_group + i); -+ } -+ -+ /* wait for I/O completion */ -+ for (i = 0; i < groups_per_page && bh[i]; i++) -+ wait_on_buffer(bh[i]); -+ -+ err = -EIO; -+ for (i = 0; i < groups_per_page && bh[i]; i++) -+ if (!buffer_uptodate(bh[i])) -+ goto out; -+ -+ first_block = page->index * blocks_per_page; -+ for (i = 0; i < blocks_per_page; i++) { -+ int group; -+ -+ group = (first_block + i) >> 1; -+ if (group >= EXT3_SB(sb)->s_groups_count) -+ break; -+ -+ data = page_address(page) + (i * blocksize); -+ bitmap = bh[group - first_group]->b_data; -+ -+ if ((first_block + i) & 1) { -+ /* this is block of buddy */ -+ mb_debug("put buddy for group %u in page %lu/%x\n", -+ group, page->index, i * blocksize); -+ memset(data, 0xff, blocksize); -+ EXT3_GROUP_INFO(sb, group)->bb_fragments = 0; -+ memset(EXT3_GROUP_INFO(sb, group)->bb_counters, 0, -+ sizeof(unsigned short)*(sb->s_blocksize_bits+2)); -+ ext3_mb_generate_buddy(sb, data, bitmap, group); -+ } else { -+ /* this is block of bitmap */ -+ mb_debug("put bitmap for group %u in page %lu/%x\n", -+ group, page->index, i * blocksize); -+ memcpy(data, bitmap, blocksize); -+ } -+ } -+ SetPageUptodate(page); -+ -+out: -+ if (bh) { -+ for (i = 0; i < groups_per_page && bh[i]; i++) -+ brelse(bh[i]); -+ if (bh != &bhs) -+ kfree(bh); -+ } -+ return err; -+} -+ -+static int ext3_mb_load_buddy(struct super_block *sb, int group, -+ struct ext3_buddy *e3b) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ struct inode *inode = sbi->s_buddy_cache; -+ int blocks_per_page, block, pnum, poff; -+ struct page *page; -+ -+ mb_debug("load group %u\n", group); -+ -+ blocks_per_page = PAGE_CACHE_SIZE / sb->s_blocksize; -+ -+ e3b->bd_blkbits = sb->s_blocksize_bits; -+ e3b->bd_info = EXT3_GROUP_INFO(sb, group); -+ e3b->bd_sb = sb; -+ e3b->bd_group = group; -+ e3b->bd_buddy_page = NULL; -+ e3b->bd_bitmap_page = NULL; -+ -+ block = group * 2; -+ pnum = block / blocks_per_page; -+ poff = block % blocks_per_page; -+ -+ /* we could use find_or_create_page(), but it locks page -+ * what we'd like to avoid in fast path ... */ -+ page = find_get_page(inode->i_mapping, pnum); -+ if (page == NULL || !PageUptodate(page)) { -+ if (page) -+ page_cache_release(page); -+ page = find_or_create_page(inode->i_mapping, pnum, GFP_NOFS); -+ if (page) { -+ BUG_ON(page->mapping != inode->i_mapping); -+ if (!PageUptodate(page)) -+ ext3_mb_init_cache(page); -+ unlock_page(page); -+ } -+ } -+ if (page == NULL || !PageUptodate(page)) -+ goto err; -+ e3b->bd_bitmap_page = page; -+ e3b->bd_bitmap = page_address(page) + (poff * sb->s_blocksize); -+ mark_page_accessed(page); -+ -+ block++; -+ pnum = block / blocks_per_page; -+ poff = block % blocks_per_page; -+ -+ page = find_get_page(inode->i_mapping, pnum); -+ if (page == NULL || !PageUptodate(page)) { -+ if (page) -+ page_cache_release(page); -+ page = find_or_create_page(inode->i_mapping, pnum, GFP_NOFS); -+ if (page) { -+ BUG_ON(page->mapping != inode->i_mapping); -+ if (!PageUptodate(page)) -+ ext3_mb_init_cache(page); -+ unlock_page(page); -+ } -+ } -+ if (page == NULL || !PageUptodate(page)) -+ goto err; -+ e3b->bd_buddy_page = page; -+ e3b->bd_buddy = page_address(page) + (poff * sb->s_blocksize); -+ mark_page_accessed(page); -+ -+ J_ASSERT(e3b->bd_bitmap_page != NULL); -+ J_ASSERT(e3b->bd_buddy_page != NULL); -+ -+ return 0; -+ -+err: -+ if (e3b->bd_bitmap_page) -+ page_cache_release(e3b->bd_bitmap_page); -+ if (e3b->bd_buddy_page) -+ page_cache_release(e3b->bd_buddy_page); -+ e3b->bd_buddy = NULL; -+ e3b->bd_bitmap = NULL; -+ return -EIO; -+} -+ -+static void ext3_mb_release_desc(struct ext3_buddy *e3b) -+{ -+ if (e3b->bd_bitmap_page) -+ page_cache_release(e3b->bd_bitmap_page); -+ if (e3b->bd_buddy_page) -+ page_cache_release(e3b->bd_buddy_page); -+} -+ -+ -+static inline void -+ext3_lock_group(struct super_block *sb, int group) -+{ -+ bit_spin_lock(EXT3_GROUP_INFO_LOCKED_BIT, -+ &EXT3_GROUP_INFO(sb, group)->bb_state); -+} -+ -+static inline void -+ext3_unlock_group(struct super_block *sb, int group) -+{ -+ bit_spin_unlock(EXT3_GROUP_INFO_LOCKED_BIT, -+ &EXT3_GROUP_INFO(sb, group)->bb_state); -+} -+ -+static int mb_find_order_for_block(struct ext3_buddy *e3b, int block) -+{ -+ int order = 1; -+ void *bb; -+ -+ J_ASSERT(EXT3_MB_BITMAP(e3b) != EXT3_MB_BUDDY(e3b)); -+ J_ASSERT(block < (1 << (e3b->bd_blkbits + 3))); -+ -+ bb = EXT3_MB_BUDDY(e3b); -+ while (order <= e3b->bd_blkbits + 1) { -+ block = block >> 1; -+ if (!mb_test_bit(block, bb)) { -+ /* this block is part of buddy of order 'order' */ -+ return order; -+ } -+ bb += 1 << (e3b->bd_blkbits - order); -+ order++; -+ } -+ return 0; -+} -+ -+static inline void mb_clear_bits(void *bm, int cur, int len) -+{ -+ __u32 *addr; -+ -+ len = cur + len; -+ while (cur < len) { -+ if ((cur & 31) == 0 && (len - cur) >= 32) { -+ /* fast path: clear whole word at once */ -+ addr = bm + (cur >> 3); -+ *addr = 0; -+ cur += 32; -+ continue; -+ } -+ mb_clear_bit_atomic(cur, bm); -+ cur++; -+ } -+} -+ -+static inline void mb_set_bits(void *bm, int cur, int len) -+{ -+ __u32 *addr; -+ -+ len = cur + len; -+ while (cur < len) { -+ if ((cur & 31) == 0 && (len - cur) >= 32) { -+ /* fast path: clear whole word at once */ -+ addr = bm + (cur >> 3); -+ *addr = 0xffffffff; -+ cur += 32; -+ continue; -+ } -+ mb_set_bit_atomic(cur, bm); -+ cur++; -+ } -+} -+ -+static int mb_free_blocks(struct ext3_buddy *e3b, int first, int count) -+{ -+ int block = 0, max = 0, order; -+ void *buddy, *buddy2; -+ -+ mb_check_buddy(e3b); -+ -+ e3b->bd_info->bb_free += count; -+ if (first < e3b->bd_info->bb_first_free) -+ e3b->bd_info->bb_first_free = first; -+ -+ /* let's maintain fragments counter */ -+ if (first != 0) -+ block = !mb_test_bit(first - 1, EXT3_MB_BITMAP(e3b)); -+ if (first + count < EXT3_SB(e3b->bd_sb)->s_mb_maxs[0]) -+ max = !mb_test_bit(first + count, EXT3_MB_BITMAP(e3b)); -+ if (block && max) -+ e3b->bd_info->bb_fragments--; -+ else if (!block && !max) -+ e3b->bd_info->bb_fragments++; -+ -+ /* let's maintain buddy itself */ -+ while (count-- > 0) { -+ block = first++; -+ order = 0; -+ -+ J_ASSERT(mb_test_bit(block, EXT3_MB_BITMAP(e3b))); -+ mb_clear_bit(block, EXT3_MB_BITMAP(e3b)); -+ e3b->bd_info->bb_counters[order]++; -+ -+ /* start of the buddy */ -+ buddy = mb_find_buddy(e3b, order, &max); -+ -+ do { -+ block &= ~1UL; -+ if (mb_test_bit(block, buddy) || -+ mb_test_bit(block + 1, buddy)) -+ break; -+ -+ /* both the buddies are free, try to coalesce them */ -+ buddy2 = mb_find_buddy(e3b, order + 1, &max); -+ -+ if (!buddy2) -+ break; -+ -+ if (order > 0) { -+ /* for special purposes, we don't set -+ * free bits in bitmap */ -+ mb_set_bit(block, buddy); -+ mb_set_bit(block + 1, buddy); -+ } -+ e3b->bd_info->bb_counters[order]--; -+ e3b->bd_info->bb_counters[order]--; -+ -+ block = block >> 1; -+ order++; -+ e3b->bd_info->bb_counters[order]++; -+ -+ mb_clear_bit(block, buddy2); -+ buddy = buddy2; -+ } while (1); -+ } -+ mb_check_buddy(e3b); -+ -+ return 0; -+} -+ -+static int mb_find_extent(struct ext3_buddy *e3b, int order, int block, -+ int needed, struct ext3_free_extent *ex) -+{ -+ int next = block, max, ord; -+ void *buddy; -+ -+ J_ASSERT(ex != NULL); -+ -+ buddy = mb_find_buddy(e3b, order, &max); -+ J_ASSERT(buddy); -+ J_ASSERT(block < max); -+ if (mb_test_bit(block, buddy)) { -+ ex->fe_len = 0; -+ ex->fe_start = 0; -+ ex->fe_group = 0; -+ return 0; -+ } -+ -+ if (likely(order == 0)) { -+ /* find actual order */ -+ order = mb_find_order_for_block(e3b, block); -+ block = block >> order; -+ } -+ -+ ex->fe_len = 1 << order; -+ ex->fe_start = block << order; -+ ex->fe_group = e3b->bd_group; -+ -+ /* calc difference from given start */ -+ next = next - ex->fe_start; -+ ex->fe_len -= next; -+ ex->fe_start += next; -+ -+ while (needed > ex->fe_len && (buddy = mb_find_buddy(e3b, order, &max))) { -+ -+ if (block + 1 >= max) -+ break; -+ -+ next = (block + 1) * (1 << order); -+ if (mb_test_bit(next, EXT3_MB_BITMAP(e3b))) -+ break; -+ -+ ord = mb_find_order_for_block(e3b, next); -+ -+ order = ord; -+ block = next >> order; -+ ex->fe_len += 1 << order; -+ } -+ -+ J_ASSERT(ex->fe_start + ex->fe_len <= (1 << (e3b->bd_blkbits + 3))); -+ return ex->fe_len; -+} -+ -+static int mb_mark_used(struct ext3_buddy *e3b, struct ext3_free_extent *ex) -+{ -+ int ord, mlen = 0, max = 0, cur; -+ int start = ex->fe_start; -+ int len = ex->fe_len; -+ unsigned ret = 0; -+ int len0 = len; -+ void *buddy; -+ -+ mb_check_buddy(e3b); -+ -+ e3b->bd_info->bb_free -= len; -+ if (e3b->bd_info->bb_first_free == start) -+ e3b->bd_info->bb_first_free += len; -+ -+ /* let's maintain fragments counter */ -+ if (start != 0) -+ mlen = !mb_test_bit(start - 1, EXT3_MB_BITMAP(e3b)); -+ if (start + len < EXT3_SB(e3b->bd_sb)->s_mb_maxs[0]) -+ max = !mb_test_bit(start + len, EXT3_MB_BITMAP(e3b)); -+ if (mlen && max) -+ e3b->bd_info->bb_fragments++; -+ else if (!mlen && !max) -+ e3b->bd_info->bb_fragments--; -+ -+ /* let's maintain buddy itself */ -+ while (len) { -+ ord = mb_find_order_for_block(e3b, start); -+ -+ if (((start >> ord) << ord) == start && len >= (1 << ord)) { -+ /* the whole chunk may be allocated at once! */ -+ mlen = 1 << ord; -+ buddy = mb_find_buddy(e3b, ord, &max); -+ J_ASSERT((start >> ord) < max); -+ mb_set_bit(start >> ord, buddy); -+ e3b->bd_info->bb_counters[ord]--; -+ start += mlen; -+ len -= mlen; -+ J_ASSERT(len >= 0); -+ continue; -+ } -+ -+ /* store for history */ -+ if (ret == 0) -+ ret = len | (ord << 16); -+ -+ /* we have to split large buddy */ -+ J_ASSERT(ord > 0); -+ buddy = mb_find_buddy(e3b, ord, &max); -+ mb_set_bit(start >> ord, buddy); -+ e3b->bd_info->bb_counters[ord]--; -+ -+ ord--; -+ cur = (start >> ord) & ~1U; -+ buddy = mb_find_buddy(e3b, ord, &max); -+ mb_clear_bit(cur, buddy); -+ mb_clear_bit(cur + 1, buddy); -+ e3b->bd_info->bb_counters[ord]++; -+ e3b->bd_info->bb_counters[ord]++; -+ } -+ -+ /* now drop all the bits in bitmap */ -+ mb_set_bits(EXT3_MB_BITMAP(e3b), ex->fe_start, len0); -+ -+ mb_check_buddy(e3b); -+ -+ return ret; -+} -+ -+/* -+ * Must be called under group lock! -+ */ -+static void ext3_mb_use_best_found(struct ext3_allocation_context *ac, -+ struct ext3_buddy *e3b) -+{ -+ unsigned long ret; -+ -+ ac->ac_b_ex.fe_len = min(ac->ac_b_ex.fe_len, ac->ac_g_ex.fe_len); -+ ret = mb_mark_used(e3b, &ac->ac_b_ex); -+ -+ ac->ac_status = AC_STATUS_FOUND; -+ ac->ac_tail = ret & 0xffff; -+ ac->ac_buddy = ret >> 16; -+ -+ /* hold in-core structures until allocated -+ * blocks are marked non-free in on-disk bitmap */ -+ ac->ac_buddy_page = e3b->bd_buddy_page; -+ page_cache_get(e3b->bd_buddy_page); -+ ac->ac_bitmap_page = e3b->bd_bitmap_page; -+ page_cache_get(e3b->bd_bitmap_page); -+} -+ -+/* -+ * The routine checks whether found extent is good enough. If it is, -+ * then the extent gets marked used and flag is set to the context -+ * to stop scanning. Otherwise, the extent is compared with the -+ * previous found extent and if new one is better, then it's stored -+ * in the context. Later, the best found extent will be used, if -+ * mballoc can't find good enough extent. -+ * -+ * FIXME: real allocation policy is to be designed yet! -+ */ -+static void ext3_mb_measure_extent(struct ext3_allocation_context *ac, -+ struct ext3_free_extent *ex, -+ struct ext3_buddy *e3b) -+{ -+ struct ext3_free_extent *bex = &ac->ac_b_ex; -+ struct ext3_free_extent *gex = &ac->ac_g_ex; -+ -+ J_ASSERT(ex->fe_len > 0); -+ J_ASSERT(ex->fe_len < (1 << ac->ac_sb->s_blocksize_bits) * 8); -+ J_ASSERT(ex->fe_start < (1 << ac->ac_sb->s_blocksize_bits) * 8); -+ -+ ac->ac_found++; -+ -+ /* -+ * The special case - take what you catch first -+ */ -+ if (unlikely(ac->ac_flags & EXT3_MB_HINT_FIRST)) { -+ *bex = *ex; -+ ext3_mb_use_best_found(ac, e3b); -+ return; -+ } -+ -+ /* -+ * Let's check whether the chunk is good enough -+ */ -+ if (ex->fe_len == gex->fe_len) { -+ *bex = *ex; -+ ext3_mb_use_best_found(ac, e3b); -+ return; -+ } -+ -+ /* -+ * If this is first found extent, just store it in the context -+ */ -+ if (bex->fe_len == 0) { -+ *bex = *ex; -+ return; -+ } -+ -+ /* -+ * If new found extent is better, store it in the context -+ */ -+ if (bex->fe_len < gex->fe_len) { -+ /* if the request isn't satisfied, any found extent -+ * larger than previous best one is better */ -+ if (ex->fe_len > bex->fe_len) -+ *bex = *ex; -+ } else if (ex->fe_len > gex->fe_len) { -+ /* if the request is satisfied, then we try to find -+ * an extent that still satisfy the request, but is -+ * smaller than previous one */ -+ *bex = *ex; -+ } -+ -+ /* -+ * Let's scan at least few extents and don't pick up a first one -+ */ -+ if (bex->fe_len > gex->fe_len && ac->ac_found > ext3_mb_min_to_scan) -+ ac->ac_status = AC_STATUS_BREAK; -+ -+ /* -+ * We don't want to scan for a whole year -+ */ -+ if (ac->ac_found > ext3_mb_max_to_scan) -+ ac->ac_status = AC_STATUS_BREAK; -+} -+ -+static int ext3_mb_try_best_found(struct ext3_allocation_context *ac, -+ struct ext3_buddy *e3b) -+{ -+ struct ext3_free_extent ex = ac->ac_b_ex; -+ int group = ex.fe_group, max, err; -+ -+ J_ASSERT(ex.fe_len > 0); -+ err = ext3_mb_load_buddy(ac->ac_sb, group, e3b); -+ if (err) -+ return err; -+ -+ ext3_lock_group(ac->ac_sb, group); -+ max = mb_find_extent(e3b, 0, ex.fe_start, ex.fe_len, &ex); -+ -+ if (max > 0) { -+ ac->ac_b_ex = ex; -+ ext3_mb_use_best_found(ac, e3b); -+ } -+ -+ ext3_unlock_group(ac->ac_sb, group); -+ -+ ext3_mb_release_desc(e3b); -+ -+ return 0; -+} -+ -+static int ext3_mb_find_by_goal(struct ext3_allocation_context *ac, -+ struct ext3_buddy *e3b) -+{ -+ int group = ac->ac_g_ex.fe_group, max, err; -+ struct ext3_sb_info *sbi = EXT3_SB(ac->ac_sb); -+ struct ext3_super_block *es = sbi->s_es; -+ struct ext3_free_extent ex; -+ -+ err = ext3_mb_load_buddy(ac->ac_sb, group, e3b); -+ if (err) -+ return err; -+ -+ ext3_lock_group(ac->ac_sb, group); -+ max = mb_find_extent(e3b, 0, ac->ac_g_ex.fe_start, -+ ac->ac_g_ex.fe_len, &ex); -+ -+ if (max >= ac->ac_g_ex.fe_len && ac->ac_g_ex.fe_len == sbi->s_stripe) { -+ unsigned long start; -+ start = (e3b->bd_group * EXT3_BLOCKS_PER_GROUP(ac->ac_sb) + -+ ex.fe_start + le32_to_cpu(es->s_first_data_block)); -+ if (start % sbi->s_stripe == 0) { -+ ac->ac_found++; -+ ac->ac_b_ex = ex; -+ ext3_mb_use_best_found(ac, e3b); -+ } -+ } else if (max >= ac->ac_g_ex.fe_len) { -+ J_ASSERT(ex.fe_len > 0); -+ J_ASSERT(ex.fe_group == ac->ac_g_ex.fe_group); -+ J_ASSERT(ex.fe_start == ac->ac_g_ex.fe_start); -+ ac->ac_found++; -+ ac->ac_b_ex = ex; -+ ext3_mb_use_best_found(ac, e3b); -+ } else if (max > 0 && (ac->ac_flags & EXT3_MB_HINT_MERGE)) { -+ /* Sometimes, caller may want to merge even small -+ * number of blocks to an existing extent */ -+ J_ASSERT(ex.fe_len > 0); -+ J_ASSERT(ex.fe_group == ac->ac_g_ex.fe_group); -+ J_ASSERT(ex.fe_start == ac->ac_g_ex.fe_start); -+ ac->ac_found++; -+ ac->ac_b_ex = ex; -+ ext3_mb_use_best_found(ac, e3b); -+ } -+ ext3_unlock_group(ac->ac_sb, group); -+ -+ ext3_mb_release_desc(e3b); -+ -+ return 0; -+} -+ -+/* -+ * The routine scans buddy structures (not bitmap!) from given order -+ * to max order and tries to find big enough chunk to satisfy the req -+ */ -+static void ext3_mb_simple_scan_group(struct ext3_allocation_context *ac, -+ struct ext3_buddy *e3b) -+{ -+ struct super_block *sb = ac->ac_sb; -+ struct ext3_group_info *grp = e3b->bd_info; -+ void *buddy; -+ int i, k, max; -+ -+ J_ASSERT(ac->ac_2order > 0); -+ for (i = ac->ac_2order; i <= sb->s_blocksize_bits + 1; i++) { -+ if (grp->bb_counters[i] == 0) -+ continue; -+ -+ buddy = mb_find_buddy(e3b, i, &max); -+ if (buddy == NULL) { -+ printk(KERN_ALERT "looking for wrong order?\n"); -+ break; -+ } -+ -+ k = mb_find_next_zero_bit(buddy, max, 0); -+ J_ASSERT(k < max); -+ -+ ac->ac_found++; -+ -+ ac->ac_b_ex.fe_len = 1 << i; -+ ac->ac_b_ex.fe_start = k << i; -+ ac->ac_b_ex.fe_group = e3b->bd_group; -+ -+ ext3_mb_use_best_found(ac, e3b); -+ J_ASSERT(ac->ac_b_ex.fe_len == ac->ac_g_ex.fe_len); -+ -+ if (unlikely(ext3_mb_stats)) -+ atomic_inc(&EXT3_SB(sb)->s_bal_2orders); -+ -+ break; -+ } -+} -+ -+/* -+ * The routine scans the group and measures all found extents. -+ * In order to optimize scanning, caller must pass number of -+ * free blocks in the group, so the routine can know upper limit. -+ */ -+static void ext3_mb_complex_scan_group(struct ext3_allocation_context *ac, -+ struct ext3_buddy *e3b) -+{ -+ struct super_block *sb = ac->ac_sb; -+ void *bitmap = EXT3_MB_BITMAP(e3b); -+ struct ext3_free_extent ex; -+ int i, free; -+ -+ free = e3b->bd_info->bb_free; -+ J_ASSERT(free > 0); -+ -+ i = e3b->bd_info->bb_first_free; -+ -+ while (free && ac->ac_status == AC_STATUS_CONTINUE) { -+ i = mb_find_next_zero_bit(bitmap, sb->s_blocksize * 8, i); -+ if (i >= sb->s_blocksize * 8) { -+ J_ASSERT(free == 0); -+ break; -+ } -+ -+ mb_find_extent(e3b, 0, i, ac->ac_g_ex.fe_len, &ex); -+ J_ASSERT(ex.fe_len > 0); -+ J_ASSERT(free >= ex.fe_len); -+ -+ ext3_mb_measure_extent(ac, &ex, e3b); -+ -+ i += ex.fe_len; -+ free -= ex.fe_len; -+ } -+} -+ -+/* -+ * This is a special case for storages like raid5 -+ * we try to find stripe-aligned chunks for stripe-size requests -+ */ -+static void ext3_mb_scan_aligned(struct ext3_allocation_context *ac, -+ struct ext3_buddy *e3b) -+{ -+ struct super_block *sb = ac->ac_sb; -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ void *bitmap = EXT3_MB_BITMAP(e3b); -+ struct ext3_free_extent ex; -+ unsigned long i, max; -+ -+ J_ASSERT(sbi->s_stripe != 0); -+ -+ /* find first stripe-aligned block */ -+ i = e3b->bd_group * EXT3_BLOCKS_PER_GROUP(sb) -+ + le32_to_cpu(sbi->s_es->s_first_data_block); -+ i = ((i + sbi->s_stripe - 1) / sbi->s_stripe) * sbi->s_stripe; -+ i = (i - le32_to_cpu(sbi->s_es->s_first_data_block)) -+ % EXT3_BLOCKS_PER_GROUP(sb); -+ -+ while (i < sb->s_blocksize * 8) { -+ if (!mb_test_bit(i, bitmap)) { -+ max = mb_find_extent(e3b, 0, i, sbi->s_stripe, &ex); -+ if (max >= sbi->s_stripe) { -+ ac->ac_found++; -+ ac->ac_b_ex = ex; -+ ext3_mb_use_best_found(ac, e3b); -+ break; -+ } -+ } -+ i += sbi->s_stripe; -+ } -+} -+ -+static int ext3_mb_good_group(struct ext3_allocation_context *ac, -+ int group, int cr) -+{ -+ struct ext3_group_info *grp = EXT3_GROUP_INFO(ac->ac_sb, group); -+ unsigned free, fragments, i, bits; -+ -+ J_ASSERT(cr >= 0 && cr < 4); -+ J_ASSERT(!EXT3_MB_GRP_NEED_INIT(grp)); -+ -+ free = grp->bb_free; -+ fragments = grp->bb_fragments; -+ if (free == 0) -+ return 0; -+ if (fragments == 0) -+ return 0; -+ -+ switch (cr) { -+ case 0: -+ J_ASSERT(ac->ac_2order != 0); -+ bits = ac->ac_sb->s_blocksize_bits + 1; -+ for (i = ac->ac_2order; i <= bits; i++) -+ if (grp->bb_counters[i] > 0) -+ return 1; -+ break; -+ case 1: -+ if ((free / fragments) >= ac->ac_g_ex.fe_len) -+ return 1; -+ break; -+ case 2: -+ if (free >= ac->ac_g_ex.fe_len) -+ return 1; -+ break; -+ case 3: -+ return 1; -+ default: -+ BUG(); -+ } -+ -+ return 0; -+} -+ -+int ext3_mb_new_blocks(handle_t *handle, struct inode *inode, -+ unsigned long goal, int *len, int flags, int *errp) -+{ -+ struct buffer_head *bitmap_bh = NULL; -+ struct ext3_allocation_context ac; -+ int i, group, block, cr, err = 0; -+ struct ext3_group_desc *gdp; -+ struct ext3_super_block *es; -+ struct buffer_head *gdp_bh; -+ struct ext3_sb_info *sbi; -+ struct super_block *sb; -+ struct ext3_buddy e3b; -+ -+ J_ASSERT(len != NULL); -+ J_ASSERT(*len > 0); -+ -+ sb = inode->i_sb; -+ if (!sb) { -+ printk("ext3_mb_new_nblocks: nonexistent device"); -+ return 0; -+ } -+ -+ if (!test_opt(sb, MBALLOC)) { -+ static int ext3_mballoc_warning = 0; -+ if (ext3_mballoc_warning == 0) { -+ printk(KERN_ERR "EXT3-fs: multiblock request with " -+ "mballoc disabled!\n"); -+ ext3_mballoc_warning++; -+ } -+ *len = 1; -+ err = ext3_new_block_old(handle, inode, goal, errp); -+ return err; -+ } -+ -+ ext3_mb_poll_new_transaction(sb, handle); -+ -+ sbi = EXT3_SB(sb); -+ es = EXT3_SB(sb)->s_es; -+ -+ /* -+ * We can't allocate > group size -+ */ -+ if (*len >= EXT3_BLOCKS_PER_GROUP(sb) - 10) -+ *len = EXT3_BLOCKS_PER_GROUP(sb) - 10; -+ -+ if (!(flags & EXT3_MB_HINT_RESERVED)) { -+ /* someone asks for non-reserved blocks */ -+ BUG_ON(*len > 1); -+ err = ext3_mb_reserve_blocks(sb, 1); -+ if (err) { -+ *errp = err; -+ return 0; -+ } -+ } -+ -+ ac.ac_buddy_page = NULL; -+ ac.ac_bitmap_page = NULL; -+ -+ /* -+ * Check quota for allocation of this blocks. -+ */ -+ while (*len && DQUOT_ALLOC_BLOCK(inode, *len)) -+ *len -= 1; -+ if (*len == 0) { -+ *errp = -EDQUOT; -+ block = 0; -+ goto out; -+ } -+ -+ /* start searching from the goal */ -+ if (goal < le32_to_cpu(es->s_first_data_block) || -+ goal >= le32_to_cpu(es->s_blocks_count)) -+ goal = le32_to_cpu(es->s_first_data_block); -+ group = (goal - le32_to_cpu(es->s_first_data_block)) / -+ EXT3_BLOCKS_PER_GROUP(sb); -+ block = ((goal - le32_to_cpu(es->s_first_data_block)) % -+ EXT3_BLOCKS_PER_GROUP(sb)); -+ -+ /* set up allocation goals */ -+ ac.ac_b_ex.fe_group = 0; -+ ac.ac_b_ex.fe_start = 0; -+ ac.ac_b_ex.fe_len = 0; -+ ac.ac_status = AC_STATUS_CONTINUE; -+ ac.ac_groups_scanned = 0; -+ ac.ac_ex_scanned = 0; -+ ac.ac_found = 0; -+ ac.ac_sb = inode->i_sb; -+ ac.ac_g_ex.fe_group = group; -+ ac.ac_g_ex.fe_start = block; -+ ac.ac_g_ex.fe_len = *len; -+ ac.ac_flags = flags; -+ ac.ac_2order = 0; -+ ac.ac_criteria = 0; -+ -+ if (*len == 1 && sbi->s_stripe) { -+ /* looks like a metadata, let's use a dirty hack for raid5 -+ * move all metadata in first groups in hope to hit cached -+ * sectors and thus avoid read-modify cycles in raid5 */ -+ ac.ac_g_ex.fe_group = group = 0; -+ } -+ -+ /* probably, the request is for 2^8+ blocks (1/2/3/... MB) */ -+ i = ffs(*len); -+ if (i >= ext3_mb_order2_reqs) { -+ i--; -+ if ((*len & (~(1 << i))) == 0) -+ ac.ac_2order = i; -+ } -+ -+ /* first, try the goal */ -+ err = ext3_mb_find_by_goal(&ac, &e3b); -+ if (err) -+ goto out_err; -+ if (ac.ac_status == AC_STATUS_FOUND) -+ goto found; -+ -+ /* Let's just scan groups to find more-less suitable blocks */ -+ cr = ac.ac_2order ? 0 : 1; -+repeat: -+ for (; cr < 4 && ac.ac_status == AC_STATUS_CONTINUE; cr++) { -+ ac.ac_criteria = cr; -+ for (i = 0; i < EXT3_SB(sb)->s_groups_count; group++, i++) { -+ if (group == EXT3_SB(sb)->s_groups_count) -+ group = 0; -+ -+ if (EXT3_MB_GRP_NEED_INIT(EXT3_GROUP_INFO(sb, group))) { -+ /* we need full data about the group -+ * to make a good selection */ -+ err = ext3_mb_load_buddy(ac.ac_sb, group, &e3b); -+ if (err) -+ goto out_err; -+ ext3_mb_release_desc(&e3b); -+ } -+ -+ /* check is group good for our criteries */ -+ if (!ext3_mb_good_group(&ac, group, cr)) -+ continue; -+ -+ err = ext3_mb_load_buddy(ac.ac_sb, group, &e3b); -+ if (err) -+ goto out_err; -+ -+ ext3_lock_group(sb, group); -+ if (!ext3_mb_good_group(&ac, group, cr)) { -+ /* someone did allocation from this group */ -+ ext3_unlock_group(sb, group); -+ ext3_mb_release_desc(&e3b); -+ continue; -+ } -+ -+ ac.ac_groups_scanned++; -+ if (cr == 0) -+ ext3_mb_simple_scan_group(&ac, &e3b); -+ else if (cr == 1 && *len == sbi->s_stripe) -+ ext3_mb_scan_aligned(&ac, &e3b); -+ else -+ ext3_mb_complex_scan_group(&ac, &e3b); -+ -+ ext3_unlock_group(sb, group); -+ -+ ext3_mb_release_desc(&e3b); -+ -+ if (ac.ac_status != AC_STATUS_CONTINUE) -+ break; -+ } -+ } -+ -+ if (ac.ac_b_ex.fe_len > 0 && ac.ac_status != AC_STATUS_FOUND && -+ !(ac.ac_flags & EXT3_MB_HINT_FIRST)) { -+ /* -+ * We've been searching too long. Let's try to allocate -+ * the best chunk we've found so far -+ */ -+ -+ /*if (ac.ac_found > ext3_mb_max_to_scan) -+ printk(KERN_DEBUG "EXT3-fs: too long searching at " -+ "%u (%d/%d)\n", cr, ac.ac_b_ex.fe_len, -+ ac.ac_g_ex.fe_len);*/ -+ ext3_mb_try_best_found(&ac, &e3b); -+ if (ac.ac_status != AC_STATUS_FOUND) { -+ /* -+ * Someone more lucky has already allocated it. -+ * The only thing we can do is just take first -+ * found block(s) -+ printk(KERN_DEBUG "EXT3-fs: someone won our chunk\n"); -+ */ -+ ac.ac_b_ex.fe_group = 0; -+ ac.ac_b_ex.fe_start = 0; -+ ac.ac_b_ex.fe_len = 0; -+ ac.ac_status = AC_STATUS_CONTINUE; -+ ac.ac_flags |= EXT3_MB_HINT_FIRST; -+ cr = 3; -+ goto repeat; -+ } -+ } -+ -+ if (ac.ac_status != AC_STATUS_FOUND) { -+ /* -+ * We aren't lucky definitely -+ */ -+ DQUOT_FREE_BLOCK(inode, *len); -+ *errp = -ENOSPC; -+ block = 0; -+#if 1 -+ printk(KERN_ERR "EXT3-fs: can't allocate: status %d flags %d\n", -+ ac.ac_status, ac.ac_flags); -+ printk(KERN_ERR "EXT3-fs: goal %d, best found %d/%d/%d cr %d\n", -+ ac.ac_g_ex.fe_len, ac.ac_b_ex.fe_group, -+ ac.ac_b_ex.fe_start, ac.ac_b_ex.fe_len, cr); -+ printk(KERN_ERR "EXT3-fs: %lu block reserved, %d found\n", -+ sbi->s_blocks_reserved, ac.ac_found); -+ printk("EXT3-fs: groups: "); -+ for (i = 0; i < EXT3_SB(sb)->s_groups_count; i++) -+ printk("%d: %d ", i, EXT3_GROUP_INFO(sb, i)->bb_free); -+ printk("\n"); -+#endif -+ goto out; -+ } -+ -+found: -+ J_ASSERT(ac.ac_b_ex.fe_len > 0); -+ -+ /* good news - free block(s) have been found. now it's time -+ * to mark block(s) in good old journaled bitmap */ -+ block = ac.ac_b_ex.fe_group * EXT3_BLOCKS_PER_GROUP(sb) -+ + ac.ac_b_ex.fe_start -+ + le32_to_cpu(es->s_first_data_block); -+ -+ /* we made a desicion, now mark found blocks in good old -+ * bitmap to be journaled */ -+ -+ ext3_debug("using block group %d(%d)\n", -+ ac.ac_b_group.group, gdp->bg_free_blocks_count); -+ -+ bitmap_bh = read_block_bitmap(sb, ac.ac_b_ex.fe_group); -+ if (!bitmap_bh) { -+ *errp = -EIO; -+ goto out_err; -+ } -+ -+ err = ext3_journal_get_write_access(handle, bitmap_bh); -+ if (err) { -+ *errp = err; -+ goto out_err; -+ } -+ -+ gdp = ext3_get_group_desc(sb, ac.ac_b_ex.fe_group, &gdp_bh); -+ if (!gdp) { -+ *errp = -EIO; -+ goto out_err; -+ } -+ -+ err = ext3_journal_get_write_access(handle, gdp_bh); -+ if (err) -+ goto out_err; -+ -+ block = ac.ac_b_ex.fe_group * EXT3_BLOCKS_PER_GROUP(sb) -+ + ac.ac_b_ex.fe_start -+ + le32_to_cpu(es->s_first_data_block); -+ -+ if (block == le32_to_cpu(gdp->bg_block_bitmap) || -+ block == le32_to_cpu(gdp->bg_inode_bitmap) || -+ in_range(block, le32_to_cpu(gdp->bg_inode_table), -+ EXT3_SB(sb)->s_itb_per_group)) -+ ext3_error(sb, "ext3_new_block", -+ "Allocating block in system zone - " -+ "block = %u", block); -+#ifdef AGGRESSIVE_CHECK -+ for (i = 0; i < ac.ac_b_ex.fe_len; i++) -+ J_ASSERT(!mb_test_bit(ac.ac_b_ex.fe_start + i, bitmap_bh->b_data)); -+#endif -+ mb_set_bits(bitmap_bh->b_data, ac.ac_b_ex.fe_start, ac.ac_b_ex.fe_len); -+ -+ spin_lock(sb_bgl_lock(sbi, ac.ac_b_ex.fe_group)); -+ gdp->bg_free_blocks_count = -+ cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) -+ - ac.ac_b_ex.fe_len); -+ spin_unlock(sb_bgl_lock(sbi, ac.ac_b_ex.fe_group)); -+ percpu_counter_mod(&sbi->s_freeblocks_counter, - ac.ac_b_ex.fe_len); -+ -+ err = ext3_journal_dirty_metadata(handle, bitmap_bh); -+ if (err) -+ goto out_err; -+ err = ext3_journal_dirty_metadata(handle, gdp_bh); -+ if (err) -+ goto out_err; -+ -+ sb->s_dirt = 1; -+ *errp = 0; -+ brelse(bitmap_bh); -+ -+ /* drop non-allocated, but dquote'd blocks */ -+ J_ASSERT(*len >= ac.ac_b_ex.fe_len); -+ DQUOT_FREE_BLOCK(inode, *len - ac.ac_b_ex.fe_len); -+ -+ *len = ac.ac_b_ex.fe_len; -+ J_ASSERT(*len > 0); -+ J_ASSERT(block != 0); -+ goto out; -+ -+out_err: -+ /* if we've already allocated something, roll it back */ -+ if (ac.ac_status == AC_STATUS_FOUND) { -+ /* FIXME: free blocks here */ -+ } -+ -+ DQUOT_FREE_BLOCK(inode, *len); -+ brelse(bitmap_bh); -+ *errp = err; -+ block = 0; -+out: -+ if (ac.ac_buddy_page) -+ page_cache_release(ac.ac_buddy_page); -+ if (ac.ac_bitmap_page) -+ page_cache_release(ac.ac_bitmap_page); -+ -+ if (!(flags & EXT3_MB_HINT_RESERVED)) { -+ /* block wasn't reserved before and we reserved it -+ * at the beginning of allocation. it doesn't matter -+ * whether we allocated anything or we failed: time -+ * to release reservation. NOTE: because I expect -+ * any multiblock request from delayed allocation -+ * path only, here is single block always */ -+ ext3_mb_release_blocks(sb, 1); -+ } -+ -+ if (unlikely(ext3_mb_stats) && ac.ac_g_ex.fe_len > 1) { -+ atomic_inc(&sbi->s_bal_reqs); -+ atomic_add(*len, &sbi->s_bal_allocated); -+ if (*len >= ac.ac_g_ex.fe_len) -+ atomic_inc(&sbi->s_bal_success); -+ atomic_add(ac.ac_found, &sbi->s_bal_ex_scanned); -+ if (ac.ac_g_ex.fe_start == ac.ac_b_ex.fe_start && -+ ac.ac_g_ex.fe_group == ac.ac_b_ex.fe_group) -+ atomic_inc(&sbi->s_bal_goals); -+ if (ac.ac_found > ext3_mb_max_to_scan) -+ atomic_inc(&sbi->s_bal_breaks); -+ } -+ -+ ext3_mb_store_history(sb, inode->i_ino, &ac); -+ -+ return block; -+} -+EXPORT_SYMBOL(ext3_mb_new_blocks); -+ -+#ifdef EXT3_MB_HISTORY -+struct ext3_mb_proc_session { -+ struct ext3_mb_history *history; -+ struct super_block *sb; -+ int start; -+ int max; -+}; -+ -+static void *ext3_mb_history_skip_empty(struct ext3_mb_proc_session *s, -+ struct ext3_mb_history *hs, -+ int first) -+{ -+ if (hs == s->history + s->max) -+ hs = s->history; -+ if (!first && hs == s->history + s->start) -+ return NULL; -+ while (hs->goal.fe_len == 0) { -+ hs++; -+ if (hs == s->history + s->max) -+ hs = s->history; -+ if (hs == s->history + s->start) -+ return NULL; -+ } -+ return hs; -+} -+ -+static void *ext3_mb_seq_history_start(struct seq_file *seq, loff_t *pos) -+{ -+ struct ext3_mb_proc_session *s = seq->private; -+ struct ext3_mb_history *hs; -+ int l = *pos; -+ -+ if (l == 0) -+ return SEQ_START_TOKEN; -+ hs = ext3_mb_history_skip_empty(s, s->history + s->start, 1); -+ if (!hs) -+ return NULL; -+ while (--l && (hs = ext3_mb_history_skip_empty(s, ++hs, 0)) != NULL); -+ return hs; -+} -+ -+static void *ext3_mb_seq_history_next(struct seq_file *seq, void *v, loff_t *pos) -+{ -+ struct ext3_mb_proc_session *s = seq->private; -+ struct ext3_mb_history *hs = v; -+ -+ ++*pos; -+ if (v == SEQ_START_TOKEN) -+ return ext3_mb_history_skip_empty(s, s->history + s->start, 1); -+ else -+ return ext3_mb_history_skip_empty(s, ++hs, 0); -+} -+ -+static int ext3_mb_seq_history_show(struct seq_file *seq, void *v) -+{ -+ struct ext3_mb_history *hs = v; -+ char buf[20], buf2[20]; -+ -+ if (v == SEQ_START_TOKEN) { -+ seq_printf(seq, "%-5s %-8s %-17s %-17s %-5s %-5s %-2s %-5s %-5s %-6s\n", -+ "pid", "inode", "goal", "result", "found", "grps", "cr", -+ "merge", "tail", "broken"); -+ return 0; -+ } -+ -+ sprintf(buf, "%u/%u/%u", hs->goal.fe_group, -+ hs->goal.fe_start, hs->goal.fe_len); -+ sprintf(buf2, "%u/%u/%u", hs->result.fe_group, -+ hs->result.fe_start, hs->result.fe_len); -+ seq_printf(seq, "%-5u %-8u %-17s %-17s %-5u %-5u %-2u %-5s %-5u %-6u\n", -+ hs->pid, hs->ino, buf, buf2, hs->found, hs->groups, -+ hs->cr, hs->merged ? "M" : "", hs->tail, -+ hs->buddy ? 1 << hs->buddy : 0); -+ return 0; -+} -+ -+static void ext3_mb_seq_history_stop(struct seq_file *seq, void *v) -+{ -+} -+ -+static struct seq_operations ext3_mb_seq_history_ops = { -+ .start = ext3_mb_seq_history_start, -+ .next = ext3_mb_seq_history_next, -+ .stop = ext3_mb_seq_history_stop, -+ .show = ext3_mb_seq_history_show, -+}; -+ -+static int ext3_mb_seq_history_open(struct inode *inode, struct file *file) -+{ -+ struct super_block *sb = PDE(inode)->data; -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ struct ext3_mb_proc_session *s; -+ int rc, size; -+ -+ s = kmalloc(sizeof(*s), GFP_KERNEL); -+ if (s == NULL) -+ return -EIO; -+ size = sizeof(struct ext3_mb_history) * sbi->s_mb_history_max; -+ s->history = kmalloc(size, GFP_KERNEL); -+ if (s == NULL) { -+ kfree(s); -+ return -EIO; -+ } -+ -+ spin_lock(&sbi->s_mb_history_lock); -+ memcpy(s->history, sbi->s_mb_history, size); -+ s->max = sbi->s_mb_history_max; -+ s->start = sbi->s_mb_history_cur % s->max; -+ spin_unlock(&sbi->s_mb_history_lock); -+ -+ rc = seq_open(file, &ext3_mb_seq_history_ops); -+ if (rc == 0) { -+ struct seq_file *m = (struct seq_file *)file->private_data; -+ m->private = s; -+ } else { -+ kfree(s->history); -+ kfree(s); -+ } -+ return rc; -+ -+} -+ -+static int ext3_mb_seq_history_release(struct inode *inode, struct file *file) -+{ -+ struct seq_file *seq = (struct seq_file *)file->private_data; -+ struct ext3_mb_proc_session *s = seq->private; -+ kfree(s->history); -+ kfree(s); -+ return seq_release(inode, file); -+} -+ -+static struct file_operations ext3_mb_seq_history_fops = { -+ .owner = THIS_MODULE, -+ .open = ext3_mb_seq_history_open, -+ .read = seq_read, -+ .llseek = seq_lseek, -+ .release = ext3_mb_seq_history_release, -+}; -+ -+static void *ext3_mb_seq_groups_start(struct seq_file *seq, loff_t *pos) -+{ -+ struct super_block *sb = seq->private; -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ long group; -+ -+ if (*pos < 0 || *pos >= sbi->s_groups_count) -+ return NULL; -+ -+ group = *pos + 1; -+ return (void *) group; -+} -+ -+static void *ext3_mb_seq_groups_next(struct seq_file *seq, void *v, loff_t *pos) -+{ -+ struct super_block *sb = seq->private; -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ long group; -+ -+ ++*pos; -+ if (*pos < 0 || *pos >= sbi->s_groups_count) -+ return NULL; -+ group = *pos + 1; -+ return (void *) group;; -+} -+ -+static int ext3_mb_seq_groups_show(struct seq_file *seq, void *v) -+{ -+ struct super_block *sb = seq->private; -+ long group = (long) v, i; -+ struct sg { -+ struct ext3_group_info info; -+ unsigned short counters[16]; -+ } sg; -+ -+ group--; -+ if (group == 0) -+ seq_printf(seq, "#%-5s: %-5s %-5s %-5s [ %-5s %-5s %-5s %-5s %-5s %-5s %-5s %-5s %-5s %-5s %-5s %-5s %-5s %-5s ]\n", -+ "group", "free", "frags", "first", "2^0", "2^1", "2^2", -+ "2^3", "2^4", "2^5", "2^6", "2^7", "2^8", "2^9", "2^10", -+ "2^11", "2^12", "2^13"); -+ -+ i = (sb->s_blocksize_bits + 2) * sizeof(sg.info.bb_counters[0]) + -+ sizeof(struct ext3_group_info); -+ ext3_lock_group(sb, group); -+ memcpy(&sg, EXT3_GROUP_INFO(sb, group), i); -+ ext3_unlock_group(sb, group); -+ -+ if (EXT3_MB_GRP_NEED_INIT(&sg.info)) -+ return 0; -+ -+ seq_printf(seq, "#%-5lu: %-5u %-5u %-5u [", group, sg.info.bb_free, -+ sg.info.bb_fragments, sg.info.bb_first_free); -+ for (i = 0; i <= 13; i++) -+ seq_printf(seq, " %-5u", i <= sb->s_blocksize_bits + 1 ? -+ sg.info.bb_counters[i] : 0); -+ seq_printf(seq, " ]\n"); -+ -+ return 0; -+} -+ -+static void ext3_mb_seq_groups_stop(struct seq_file *seq, void *v) -+{ -+} -+ -+static struct seq_operations ext3_mb_seq_groups_ops = { -+ .start = ext3_mb_seq_groups_start, -+ .next = ext3_mb_seq_groups_next, -+ .stop = ext3_mb_seq_groups_stop, -+ .show = ext3_mb_seq_groups_show, -+}; -+ -+static int ext3_mb_seq_groups_open(struct inode *inode, struct file *file) -+{ -+ struct super_block *sb = PDE(inode)->data; -+ int rc; -+ -+ rc = seq_open(file, &ext3_mb_seq_groups_ops); -+ if (rc == 0) { -+ struct seq_file *m = (struct seq_file *)file->private_data; -+ m->private = sb; -+ } -+ return rc; -+ -+} -+ -+static struct file_operations ext3_mb_seq_groups_fops = { -+ .owner = THIS_MODULE, -+ .open = ext3_mb_seq_groups_open, -+ .read = seq_read, -+ .llseek = seq_lseek, -+ .release = seq_release, -+}; -+ -+static void ext3_mb_history_release(struct super_block *sb) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ char name[64]; -+ -+ snprintf(name, sizeof(name) - 1, "%s", bdevname(sb->s_bdev, name)); -+ remove_proc_entry("mb_groups", sbi->s_mb_proc); -+ remove_proc_entry("mb_history", sbi->s_mb_proc); -+ remove_proc_entry(name, proc_root_ext3); -+ -+ if (sbi->s_mb_history) -+ kfree(sbi->s_mb_history); -+} -+ -+static void ext3_mb_history_init(struct super_block *sb) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ char name[64]; -+ int i; -+ -+ snprintf(name, sizeof(name) - 1, "%s", bdevname(sb->s_bdev, name)); -+ sbi->s_mb_proc = proc_mkdir(name, proc_root_ext3); -+ if (sbi->s_mb_proc != NULL) { -+ struct proc_dir_entry *p; -+ p = create_proc_entry("mb_history", S_IRUGO, sbi->s_mb_proc); -+ if (p) { -+ p->proc_fops = &ext3_mb_seq_history_fops; -+ p->data = sb; -+ } -+ p = create_proc_entry("mb_groups", S_IRUGO, sbi->s_mb_proc); -+ if (p) { -+ p->proc_fops = &ext3_mb_seq_groups_fops; -+ p->data = sb; -+ } -+ } -+ -+ sbi->s_mb_history_max = 1000; -+ sbi->s_mb_history_cur = 0; -+ spin_lock_init(&sbi->s_mb_history_lock); -+ i = sbi->s_mb_history_max * sizeof(struct ext3_mb_history); -+ sbi->s_mb_history = kmalloc(i, GFP_KERNEL); -+ memset(sbi->s_mb_history, 0, i); -+ /* if we can't allocate history, then we simple won't use it */ -+} -+ -+static void -+ext3_mb_store_history(struct super_block *sb, unsigned ino, -+ struct ext3_allocation_context *ac) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ struct ext3_mb_history h; -+ -+ if (likely(sbi->s_mb_history == NULL)) -+ return; -+ -+ h.pid = current->pid; -+ h.ino = ino; -+ h.goal = ac->ac_g_ex; -+ h.result = ac->ac_b_ex; -+ h.found = ac->ac_found; -+ h.cr = ac->ac_criteria; -+ h.groups = ac->ac_groups_scanned; -+ h.tail = ac->ac_tail; -+ h.buddy = ac->ac_buddy; -+ h.merged = 0; -+ if (ac->ac_g_ex.fe_start == ac->ac_b_ex.fe_start && -+ ac->ac_g_ex.fe_group == ac->ac_b_ex.fe_group) -+ h.merged = 1; -+ -+ spin_lock(&sbi->s_mb_history_lock); -+ memcpy(sbi->s_mb_history + sbi->s_mb_history_cur, &h, sizeof(h)); -+ if (++sbi->s_mb_history_cur >= sbi->s_mb_history_max) -+ sbi->s_mb_history_cur = 0; -+ spin_unlock(&sbi->s_mb_history_lock); -+} -+ -+#else -+#define ext3_mb_history_release(sb) -+#define ext3_mb_history_init(sb) -+#endif -+ -+int ext3_mb_init_backend(struct super_block *sb) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ int i, j, len, metalen; -+ int num_meta_group_infos = -+ (sbi->s_groups_count + EXT3_DESC_PER_BLOCK(sb) - 1) >> -+ EXT3_DESC_PER_BLOCK_BITS(sb); -+ struct ext3_group_info **meta_group_info; -+ -+ /* An 8TB filesystem with 64-bit pointers requires a 4096 byte -+ * kmalloc. A 128kb malloc should suffice for a 256TB filesystem. -+ * So a two level scheme suffices for now. */ -+ sbi->s_group_info = kmalloc(sizeof(*sbi->s_group_info) * -+ num_meta_group_infos, GFP_KERNEL); -+ if (sbi->s_group_info == NULL) { -+ printk(KERN_ERR "EXT3-fs: can't allocate buddy meta group\n"); -+ return -ENOMEM; -+ } -+ sbi->s_buddy_cache = new_inode(sb); -+ if (sbi->s_buddy_cache == NULL) { -+ printk(KERN_ERR "EXT3-fs: can't get new inode\n"); -+ goto err_freesgi; -+ } -+ -+ metalen = sizeof(*meta_group_info) << EXT3_DESC_PER_BLOCK_BITS(sb); -+ for (i = 0; i < num_meta_group_infos; i++) { -+ if ((i + 1) == num_meta_group_infos) -+ metalen = sizeof(*meta_group_info) * -+ (sbi->s_groups_count - -+ (i << EXT3_DESC_PER_BLOCK_BITS(sb))); -+ meta_group_info = kmalloc(metalen, GFP_KERNEL); -+ if (meta_group_info == NULL) { -+ printk(KERN_ERR "EXT3-fs: can't allocate mem for a " -+ "buddy group\n"); -+ goto err_freemeta; -+ } -+ sbi->s_group_info[i] = meta_group_info; -+ } -+ -+ /* -+ * calculate needed size. if change bb_counters size, -+ * don't forget about ext3_mb_generate_buddy() -+ */ -+ len = sizeof(struct ext3_group_info); -+ len += sizeof(unsigned short) * (sb->s_blocksize_bits + 2); -+ for (i = 0; i < sbi->s_groups_count; i++) { -+ struct ext3_group_desc * desc; -+ -+ meta_group_info = -+ sbi->s_group_info[i >> EXT3_DESC_PER_BLOCK_BITS(sb)]; -+ j = i & (EXT3_DESC_PER_BLOCK(sb) - 1); -+ -+ meta_group_info[j] = kmalloc(len, GFP_KERNEL); -+ if (meta_group_info[j] == NULL) { -+ printk(KERN_ERR "EXT3-fs: can't allocate buddy mem\n"); -+ i--; -+ goto err_freebuddy; -+ } -+ desc = ext3_get_group_desc(sb, i, NULL); -+ if (desc == NULL) { -+ printk(KERN_ERR"EXT3-fs: can't read descriptor %u\n",i); -+ goto err_freebuddy; -+ } -+ memset(meta_group_info[j], 0, len); -+ set_bit(EXT3_GROUP_INFO_NEED_INIT_BIT, -+ &meta_group_info[j]->bb_state); -+ meta_group_info[j]->bb_free = -+ le16_to_cpu(desc->bg_free_blocks_count); -+ } -+ -+ return 0; -+ -+err_freebuddy: -+ while (i >= 0) { -+ kfree(EXT3_GROUP_INFO(sb, i)); -+ i--; -+ } -+ i = num_meta_group_infos; -+err_freemeta: -+ while (--i >= 0) -+ kfree(sbi->s_group_info[i]); -+ iput(sbi->s_buddy_cache); -+err_freesgi: -+ kfree(sbi->s_group_info); -+ return -ENOMEM; -+} -+ -+int ext3_mb_init(struct super_block *sb, int needs_recovery) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ struct inode *root = sb->s_root->d_inode; -+ unsigned i, offset, max; -+ struct dentry *dentry; -+ -+ if (!test_opt(sb, MBALLOC)) -+ return 0; -+ -+ i = (sb->s_blocksize_bits + 2) * sizeof(unsigned short); -+ -+ sbi->s_mb_offsets = kmalloc(i, GFP_KERNEL); -+ if (sbi->s_mb_offsets == NULL) { -+ clear_opt(sbi->s_mount_opt, MBALLOC); -+ return -ENOMEM; -+ } -+ sbi->s_mb_maxs = kmalloc(i, GFP_KERNEL); -+ if (sbi->s_mb_maxs == NULL) { -+ clear_opt(sbi->s_mount_opt, MBALLOC); -+ kfree(sbi->s_mb_maxs); -+ return -ENOMEM; -+ } -+ -+ /* order 0 is regular bitmap */ -+ sbi->s_mb_maxs[0] = sb->s_blocksize << 3; -+ sbi->s_mb_offsets[0] = 0; -+ -+ i = 1; -+ offset = 0; -+ max = sb->s_blocksize << 2; -+ do { -+ sbi->s_mb_offsets[i] = offset; -+ sbi->s_mb_maxs[i] = max; -+ offset += 1 << (sb->s_blocksize_bits - i); -+ max = max >> 1; -+ i++; -+ } while (i <= sb->s_blocksize_bits + 1); -+ -+ /* init file for buddy data */ -+ if ((i = ext3_mb_init_backend(sb))) { -+ clear_opt(sbi->s_mount_opt, MBALLOC); -+ kfree(sbi->s_mb_offsets); -+ kfree(sbi->s_mb_maxs); -+ return i; -+ } -+ -+ spin_lock_init(&sbi->s_reserve_lock); -+ spin_lock_init(&sbi->s_md_lock); -+ INIT_LIST_HEAD(&sbi->s_active_transaction); -+ INIT_LIST_HEAD(&sbi->s_closed_transaction); -+ INIT_LIST_HEAD(&sbi->s_committed_transaction); -+ spin_lock_init(&sbi->s_bal_lock); -+ -+ /* remove old on-disk buddy file */ -+ down(&root->i_sem); -+ dentry = lookup_one_len(".buddy", sb->s_root, strlen(".buddy")); -+ if (dentry->d_inode != NULL) { -+ i = vfs_unlink(root, dentry); -+ if (i != 0) -+ printk("EXT3-fs: can't remove .buddy file: %d\n", i); -+ } -+ dput(dentry); -+ up(&root->i_sem); -+ -+ ext3_mb_history_init(sb); -+ -+ printk("EXT3-fs: mballoc enabled\n"); -+ return 0; -+} -+ -+int ext3_mb_release(struct super_block *sb) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ int i, num_meta_group_infos; -+ -+ if (!test_opt(sb, MBALLOC)) -+ return 0; -+ -+ /* release freed, non-committed blocks */ -+ spin_lock(&sbi->s_md_lock); -+ list_splice_init(&sbi->s_closed_transaction, -+ &sbi->s_committed_transaction); -+ list_splice_init(&sbi->s_active_transaction, -+ &sbi->s_committed_transaction); -+ spin_unlock(&sbi->s_md_lock); -+ ext3_mb_free_committed_blocks(sb); -+ -+ if (sbi->s_group_info) { -+ for (i = 0; i < sbi->s_groups_count; i++) -+ kfree(EXT3_GROUP_INFO(sb, i)); -+ num_meta_group_infos = (sbi->s_groups_count + -+ EXT3_DESC_PER_BLOCK(sb) - 1) >> -+ EXT3_DESC_PER_BLOCK_BITS(sb); -+ for (i = 0; i < num_meta_group_infos; i++) -+ kfree(sbi->s_group_info[i]); -+ kfree(sbi->s_group_info); -+ } -+ if (sbi->s_mb_offsets) -+ kfree(sbi->s_mb_offsets); -+ if (sbi->s_mb_maxs) -+ kfree(sbi->s_mb_maxs); -+ if (sbi->s_buddy_cache) -+ iput(sbi->s_buddy_cache); -+ if (sbi->s_blocks_reserved) -+ printk("ext3-fs: %ld blocks being reserved at umount!\n", -+ sbi->s_blocks_reserved); -+ if (ext3_mb_stats) { -+ printk("EXT3-fs: mballoc: %u blocks %u reqs (%u success)\n", -+ atomic_read(&sbi->s_bal_allocated), -+ atomic_read(&sbi->s_bal_reqs), -+ atomic_read(&sbi->s_bal_success)); -+ printk("EXT3-fs: mballoc: %u extents scanned, %u goal hits, " -+ "%u 2^N hits, %u breaks\n", -+ atomic_read(&sbi->s_bal_ex_scanned), -+ atomic_read(&sbi->s_bal_goals), -+ atomic_read(&sbi->s_bal_2orders), -+ atomic_read(&sbi->s_bal_breaks)); -+ printk("EXT3-fs: mballoc: %lu generated and it took %Lu\n", -+ sbi->s_mb_buddies_generated++, -+ sbi->s_mb_generation_time); -+ } -+ -+ ext3_mb_history_release(sb); -+ -+ return 0; -+} -+ -+void ext3_mb_free_committed_blocks(struct super_block *sb) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ int err, i, count = 0, count2 = 0; -+ struct ext3_free_metadata *md; -+ struct ext3_buddy e3b; -+ -+ if (list_empty(&sbi->s_committed_transaction)) -+ return; -+ -+ /* there is committed blocks to be freed yet */ -+ do { -+ /* get next array of blocks */ -+ md = NULL; -+ spin_lock(&sbi->s_md_lock); -+ if (!list_empty(&sbi->s_committed_transaction)) { -+ md = list_entry(sbi->s_committed_transaction.next, -+ struct ext3_free_metadata, list); -+ list_del(&md->list); -+ } -+ spin_unlock(&sbi->s_md_lock); -+ -+ if (md == NULL) -+ break; -+ -+ mb_debug("gonna free %u blocks in group %u (0x%p):", -+ md->num, md->group, md); -+ -+ err = ext3_mb_load_buddy(sb, md->group, &e3b); -+ /* we expect to find existing buddy because it's pinned */ -+ BUG_ON(err != 0); -+ -+ /* there are blocks to put in buddy to make them really free */ -+ count += md->num; -+ count2++; -+ ext3_lock_group(sb, md->group); -+ for (i = 0; i < md->num; i++) { -+ mb_debug(" %u", md->blocks[i]); -+ mb_free_blocks(&e3b, md->blocks[i], 1); -+ } -+ mb_debug("\n"); -+ ext3_unlock_group(sb, md->group); -+ -+ /* balance refcounts from ext3_mb_free_metadata() */ -+ page_cache_release(e3b.bd_buddy_page); -+ page_cache_release(e3b.bd_bitmap_page); -+ -+ kfree(md); -+ ext3_mb_release_desc(&e3b); -+ -+ } while (md); -+ mb_debug("freed %u blocks in %u structures\n", count, count2); -+} -+ -+void ext3_mb_poll_new_transaction(struct super_block *sb, handle_t *handle) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ -+ if (sbi->s_last_transaction == handle->h_transaction->t_tid) -+ return; -+ -+ /* new transaction! time to close last one and free blocks for -+ * committed transaction. we know that only transaction can be -+ * active, so previos transaction can be being logged and we -+ * know that transaction before previous is known to be already -+ * logged. this means that now we may free blocks freed in all -+ * transactions before previous one. hope I'm clear enough ... */ -+ -+ spin_lock(&sbi->s_md_lock); -+ if (sbi->s_last_transaction != handle->h_transaction->t_tid) { -+ mb_debug("new transaction %lu, old %lu\n", -+ (unsigned long) handle->h_transaction->t_tid, -+ (unsigned long) sbi->s_last_transaction); -+ list_splice_init(&sbi->s_closed_transaction, -+ &sbi->s_committed_transaction); -+ list_splice_init(&sbi->s_active_transaction, -+ &sbi->s_closed_transaction); -+ sbi->s_last_transaction = handle->h_transaction->t_tid; -+ } -+ spin_unlock(&sbi->s_md_lock); -+ -+ ext3_mb_free_committed_blocks(sb); -+} -+ -+int ext3_mb_free_metadata(handle_t *handle, struct ext3_buddy *e3b, -+ int group, int block, int count) -+{ -+ struct ext3_group_info *db = e3b->bd_info; -+ struct super_block *sb = e3b->bd_sb; -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ struct ext3_free_metadata *md; -+ int i; -+ -+ J_ASSERT(e3b->bd_bitmap_page != NULL); -+ J_ASSERT(e3b->bd_buddy_page != NULL); -+ -+ ext3_lock_group(sb, group); -+ for (i = 0; i < count; i++) { -+ md = db->bb_md_cur; -+ if (md && db->bb_tid != handle->h_transaction->t_tid) { -+ db->bb_md_cur = NULL; -+ md = NULL; -+ } -+ -+ if (md == NULL) { -+ ext3_unlock_group(sb, group); -+ md = kmalloc(sizeof(*md), GFP_KERNEL); -+ if (md == NULL) -+ return -ENOMEM; -+ md->num = 0; -+ md->group = group; -+ -+ ext3_lock_group(sb, group); -+ if (db->bb_md_cur == NULL) { -+ spin_lock(&sbi->s_md_lock); -+ list_add(&md->list, &sbi->s_active_transaction); -+ spin_unlock(&sbi->s_md_lock); -+ /* protect buddy cache from being freed, -+ * otherwise we'll refresh it from -+ * on-disk bitmap and lose not-yet-available -+ * blocks */ -+ page_cache_get(e3b->bd_buddy_page); -+ page_cache_get(e3b->bd_bitmap_page); -+ db->bb_md_cur = md; -+ db->bb_tid = handle->h_transaction->t_tid; -+ mb_debug("new md 0x%p for group %u\n", -+ md, md->group); -+ } else { -+ kfree(md); -+ md = db->bb_md_cur; -+ } -+ } -+ -+ BUG_ON(md->num >= EXT3_BB_MAX_BLOCKS); -+ md->blocks[md->num] = block + i; -+ md->num++; -+ if (md->num == EXT3_BB_MAX_BLOCKS) { -+ /* no more space, put full container on a sb's list */ -+ db->bb_md_cur = NULL; -+ } -+ } -+ ext3_unlock_group(sb, group); -+ return 0; -+} -+ -+void ext3_mb_free_blocks(handle_t *handle, struct inode *inode, -+ unsigned long block, unsigned long count, -+ int metadata, int *freed) -+{ -+ struct buffer_head *bitmap_bh = NULL; -+ struct ext3_group_desc *gdp; -+ struct ext3_super_block *es; -+ unsigned long bit, overflow; -+ struct buffer_head *gd_bh; -+ unsigned long block_group; -+ struct ext3_sb_info *sbi; -+ struct super_block *sb; -+ struct ext3_buddy e3b; -+ int err = 0, ret; -+ -+ *freed = 0; -+ sb = inode->i_sb; -+ if (!sb) { -+ printk ("ext3_free_blocks: nonexistent device"); -+ return; -+ } -+ -+ ext3_mb_poll_new_transaction(sb, handle); -+ -+ sbi = EXT3_SB(sb); -+ es = EXT3_SB(sb)->s_es; -+ if (block < le32_to_cpu(es->s_first_data_block) || -+ block + count < block || -+ block + count > le32_to_cpu(es->s_blocks_count)) { -+ ext3_error (sb, "ext3_free_blocks", -+ "Freeing blocks not in datazone - " -+ "block = %lu, count = %lu", block, count); -+ goto error_return; -+ } -+ -+ ext3_debug("freeing block %lu\n", block); -+ -+do_more: -+ overflow = 0; -+ block_group = (block - le32_to_cpu(es->s_first_data_block)) / -+ EXT3_BLOCKS_PER_GROUP(sb); -+ bit = (block - le32_to_cpu(es->s_first_data_block)) % -+ EXT3_BLOCKS_PER_GROUP(sb); -+ /* -+ * Check to see if we are freeing blocks across a group -+ * boundary. -+ */ -+ if (bit + count > EXT3_BLOCKS_PER_GROUP(sb)) { -+ overflow = bit + count - EXT3_BLOCKS_PER_GROUP(sb); -+ count -= overflow; -+ } -+ brelse(bitmap_bh); -+ bitmap_bh = read_block_bitmap(sb, block_group); -+ if (!bitmap_bh) -+ goto error_return; -+ gdp = ext3_get_group_desc (sb, block_group, &gd_bh); -+ if (!gdp) -+ goto error_return; -+ -+ if (in_range (le32_to_cpu(gdp->bg_block_bitmap), block, count) || -+ in_range (le32_to_cpu(gdp->bg_inode_bitmap), block, count) || -+ in_range (block, le32_to_cpu(gdp->bg_inode_table), -+ EXT3_SB(sb)->s_itb_per_group) || -+ in_range (block + count - 1, le32_to_cpu(gdp->bg_inode_table), -+ EXT3_SB(sb)->s_itb_per_group)) -+ ext3_error (sb, "ext3_free_blocks", -+ "Freeing blocks in system zones - " -+ "Block = %lu, count = %lu", -+ block, count); -+ -+ BUFFER_TRACE(bitmap_bh, "getting write access"); -+ err = ext3_journal_get_write_access(handle, bitmap_bh); -+ if (err) -+ goto error_return; -+ -+ /* -+ * We are about to modify some metadata. Call the journal APIs -+ * to unshare ->b_data if a currently-committing transaction is -+ * using it -+ */ -+ BUFFER_TRACE(gd_bh, "get_write_access"); -+ err = ext3_journal_get_write_access(handle, gd_bh); -+ if (err) -+ goto error_return; -+ -+ err = ext3_mb_load_buddy(sb, block_group, &e3b); -+ if (err) -+ goto error_return; -+ -+#ifdef AGGRESSIVE_CHECK -+ { -+ int i; -+ for (i = 0; i < count; i++) -+ J_ASSERT(mb_test_bit(bit + i, bitmap_bh->b_data)); -+ } -+#endif -+ mb_clear_bits(bitmap_bh->b_data, bit, count); -+ -+ /* We dirtied the bitmap block */ -+ BUFFER_TRACE(bitmap_bh, "dirtied bitmap block"); -+ err = ext3_journal_dirty_metadata(handle, bitmap_bh); -+ -+ if (metadata) { -+ /* blocks being freed are metadata. these blocks shouldn't -+ * be used until this transaction is committed */ -+ ext3_mb_free_metadata(handle, &e3b, block_group, bit, count); -+ } else { -+ ext3_lock_group(sb, block_group); -+ mb_free_blocks(&e3b, bit, count); -+ ext3_unlock_group(sb, block_group); -+ } -+ -+ spin_lock(sb_bgl_lock(sbi, block_group)); -+ gdp->bg_free_blocks_count = -+ cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) + count); -+ spin_unlock(sb_bgl_lock(sbi, block_group)); -+ percpu_counter_mod(&sbi->s_freeblocks_counter, count); -+ -+ ext3_mb_release_desc(&e3b); -+ -+ *freed = count; -+ -+ /* And the group descriptor block */ -+ BUFFER_TRACE(gd_bh, "dirtied group descriptor block"); -+ ret = ext3_journal_dirty_metadata(handle, gd_bh); -+ if (!err) err = ret; -+ -+ if (overflow && !err) { -+ block += count; -+ count = overflow; -+ goto do_more; -+ } -+ sb->s_dirt = 1; -+error_return: -+ brelse(bitmap_bh); -+ ext3_std_error(sb, err); -+ return; -+} -+ -+int ext3_mb_reserve_blocks(struct super_block *sb, int blocks) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ int free, ret = -ENOSPC; -+ -+ BUG_ON(blocks < 0); -+ spin_lock(&sbi->s_reserve_lock); -+ free = percpu_counter_read_positive(&sbi->s_freeblocks_counter); -+ if (blocks <= free - sbi->s_blocks_reserved) { -+ sbi->s_blocks_reserved += blocks; -+ ret = 0; -+ } -+ spin_unlock(&sbi->s_reserve_lock); -+ return ret; -+} -+ -+void ext3_mb_release_blocks(struct super_block *sb, int blocks) -+{ -+ struct ext3_sb_info *sbi = EXT3_SB(sb); -+ -+ BUG_ON(blocks < 0); -+ spin_lock(&sbi->s_reserve_lock); -+ sbi->s_blocks_reserved -= blocks; -+ WARN_ON(sbi->s_blocks_reserved < 0); -+ if (sbi->s_blocks_reserved < 0) -+ sbi->s_blocks_reserved = 0; -+ spin_unlock(&sbi->s_reserve_lock); -+} -+ -+int ext3_new_block(handle_t *handle, struct inode *inode, -+ unsigned long goal, int *errp) -+{ -+ int ret, len; -+ -+ if (!test_opt(inode->i_sb, MBALLOC)) { -+ ret = ext3_new_block_old(handle, inode, goal, errp); -+ goto out; -+ } -+ len = 1; -+ ret = ext3_mb_new_blocks(handle, inode, goal, &len, 0, errp); -+out: -+ return ret; -+} -+ -+ -+void ext3_free_blocks(handle_t *handle, struct inode * inode, -+ unsigned long block, unsigned long count, int metadata) -+{ -+ struct super_block *sb; -+ int freed; -+ -+ sb = inode->i_sb; -+ if (!test_opt(sb, MBALLOC) || !EXT3_SB(sb)->s_group_info) -+ ext3_free_blocks_old(handle, inode, block, count); -+ else { -+ ext3_mb_free_blocks(handle, inode, block, count, metadata, &freed); -+ if (freed) -+ DQUOT_FREE_BLOCK(inode, freed); -+ } -+ return; -+} -+ -+#define EXT3_ROOT "ext3" -+#define EXT3_MB_STATS_NAME "mb_stats" -+#define EXT3_MB_MAX_TO_SCAN_NAME "mb_max_to_scan" -+#define EXT3_MB_MIN_TO_SCAN_NAME "mb_min_to_scan" -+#define EXT3_MB_ORDER2_REQ "mb_order2_req" -+ -+static int ext3_mb_stats_read(char *page, char **start, off_t off, -+ int count, int *eof, void *data) -+{ -+ int len; -+ -+ *eof = 1; -+ if (off != 0) -+ return 0; -+ -+ len = sprintf(page, "%ld\n", ext3_mb_stats); -+ *start = page; -+ return len; -+} -+ -+static int ext3_mb_stats_write(struct file *file, const char *buffer, -+ unsigned long count, void *data) -+{ -+ char str[32]; -+ -+ if (count >= sizeof(str)) { -+ printk(KERN_ERR "EXT3-fs: %s string too long, max %u bytes\n", -+ EXT3_MB_STATS_NAME, (int)sizeof(str)); -+ return -EOVERFLOW; -+ } -+ -+ if (copy_from_user(str, buffer, count)) -+ return -EFAULT; -+ -+ /* Only set to 0 or 1 respectively; zero->0; non-zero->1 */ -+ ext3_mb_stats = (simple_strtol(str, NULL, 0) != 0); -+ return count; -+} -+ -+static int ext3_mb_max_to_scan_read(char *page, char **start, off_t off, -+ int count, int *eof, void *data) -+{ -+ int len; -+ -+ *eof = 1; -+ if (off != 0) -+ return 0; -+ -+ len = sprintf(page, "%ld\n", ext3_mb_max_to_scan); -+ *start = page; -+ return len; -+} -+ -+static int ext3_mb_max_to_scan_write(struct file *file, const char *buffer, -+ unsigned long count, void *data) -+{ -+ char str[32]; -+ long value; -+ -+ if (count >= sizeof(str)) { -+ printk(KERN_ERR "EXT3-fs: %s string too long, max %u bytes\n", -+ EXT3_MB_MAX_TO_SCAN_NAME, (int)sizeof(str)); -+ return -EOVERFLOW; -+ } -+ -+ if (copy_from_user(str, buffer, count)) -+ return -EFAULT; -+ -+ /* Only set to 0 or 1 respectively; zero->0; non-zero->1 */ -+ value = simple_strtol(str, NULL, 0); -+ if (value <= 0) -+ return -ERANGE; -+ -+ ext3_mb_max_to_scan = value; -+ -+ return count; -+} -+ -+static int ext3_mb_min_to_scan_read(char *page, char **start, off_t off, -+ int count, int *eof, void *data) -+{ -+ int len; -+ -+ *eof = 1; -+ if (off != 0) -+ return 0; -+ -+ len = sprintf(page, "%ld\n", ext3_mb_min_to_scan); -+ *start = page; -+ return len; -+} -+ -+static int ext3_mb_min_to_scan_write(struct file *file, const char *buffer, -+ unsigned long count, void *data) -+{ -+ char str[32]; -+ long value; -+ -+ if (count >= sizeof(str)) { -+ printk(KERN_ERR "EXT3-fs: %s string too long, max %u bytes\n", -+ EXT3_MB_MIN_TO_SCAN_NAME, (int)sizeof(str)); -+ return -EOVERFLOW; -+ } -+ -+ if (copy_from_user(str, buffer, count)) -+ return -EFAULT; -+ -+ /* Only set to 0 or 1 respectively; zero->0; non-zero->1 */ -+ value = simple_strtol(str, NULL, 0); -+ if (value <= 0) -+ return -ERANGE; -+ -+ ext3_mb_min_to_scan = value; -+ -+ return count; -+} -+ -+static int ext3_mb_order2_req_read(char *page, char **start, off_t off, -+ int count, int *eof, void *data) -+{ -+ int len; -+ -+ *eof = 1; -+ if (off != 0) -+ return 0; -+ -+ len = sprintf(page, "%ld\n", ext3_mb_order2_reqs); -+ *start = page; -+ return len; -+} -+ -+static int ext3_mb_order2_req_write(struct file *file, const char *buffer, -+ unsigned long count, void *data) -+{ -+ char str[32]; -+ long value; -+ -+ if (count >= sizeof(str)) { -+ printk(KERN_ERR "EXT3-fs: %s string too long, max %u bytes\n", -+ EXT3_MB_MIN_TO_SCAN_NAME, (int)sizeof(str)); -+ return -EOVERFLOW; -+ } -+ -+ if (copy_from_user(str, buffer, count)) -+ return -EFAULT; -+ -+ /* Only set to 0 or 1 respectively; zero->0; non-zero->1 */ -+ value = simple_strtol(str, NULL, 0); -+ if (value <= 0) -+ return -ERANGE; -+ -+ ext3_mb_order2_reqs = value; -+ -+ return count; -+} -+ -+int __init init_ext3_proc(void) -+{ -+ struct proc_dir_entry *proc_ext3_mb_stats; -+ struct proc_dir_entry *proc_ext3_mb_max_to_scan; -+ struct proc_dir_entry *proc_ext3_mb_min_to_scan; -+ struct proc_dir_entry *proc_ext3_mb_order2_req; -+ -+ proc_root_ext3 = proc_mkdir(EXT3_ROOT, proc_root_fs); -+ if (proc_root_ext3 == NULL) { -+ printk(KERN_ERR "EXT3-fs: Unable to create %s\n", EXT3_ROOT); -+ return -EIO; -+ } -+ -+ /* Initialize EXT3_MB_STATS_NAME */ -+ proc_ext3_mb_stats = create_proc_entry(EXT3_MB_STATS_NAME, -+ S_IFREG | S_IRUGO | S_IWUSR, proc_root_ext3); -+ if (proc_ext3_mb_stats == NULL) { -+ printk(KERN_ERR "EXT3-fs: Unable to create %s\n", -+ EXT3_MB_STATS_NAME); -+ remove_proc_entry(EXT3_ROOT, proc_root_fs); -+ return -EIO; -+ } -+ -+ proc_ext3_mb_stats->data = NULL; -+ proc_ext3_mb_stats->read_proc = ext3_mb_stats_read; -+ proc_ext3_mb_stats->write_proc = ext3_mb_stats_write; -+ -+ /* Initialize EXT3_MAX_TO_SCAN_NAME */ -+ proc_ext3_mb_max_to_scan = create_proc_entry( -+ EXT3_MB_MAX_TO_SCAN_NAME, -+ S_IFREG | S_IRUGO | S_IWUSR, proc_root_ext3); -+ if (proc_ext3_mb_max_to_scan == NULL) { -+ printk(KERN_ERR "EXT3-fs: Unable to create %s\n", -+ EXT3_MB_MAX_TO_SCAN_NAME); -+ remove_proc_entry(EXT3_MB_STATS_NAME, proc_root_ext3); -+ remove_proc_entry(EXT3_ROOT, proc_root_fs); -+ return -EIO; -+ } -+ -+ proc_ext3_mb_max_to_scan->data = NULL; -+ proc_ext3_mb_max_to_scan->read_proc = ext3_mb_max_to_scan_read; -+ proc_ext3_mb_max_to_scan->write_proc = ext3_mb_max_to_scan_write; -+ -+ /* Initialize EXT3_MIN_TO_SCAN_NAME */ -+ proc_ext3_mb_min_to_scan = create_proc_entry( -+ EXT3_MB_MIN_TO_SCAN_NAME, -+ S_IFREG | S_IRUGO | S_IWUSR, proc_root_ext3); -+ if (proc_ext3_mb_min_to_scan == NULL) { -+ printk(KERN_ERR "EXT3-fs: Unable to create %s\n", -+ EXT3_MB_MIN_TO_SCAN_NAME); -+ remove_proc_entry(EXT3_MB_MAX_TO_SCAN_NAME, proc_root_ext3); -+ remove_proc_entry(EXT3_MB_STATS_NAME, proc_root_ext3); -+ remove_proc_entry(EXT3_ROOT, proc_root_fs); -+ return -EIO; -+ } -+ -+ proc_ext3_mb_min_to_scan->data = NULL; -+ proc_ext3_mb_min_to_scan->read_proc = ext3_mb_min_to_scan_read; -+ proc_ext3_mb_min_to_scan->write_proc = ext3_mb_min_to_scan_write; -+ -+ /* Initialize EXT3_ORDER2_REQ */ -+ proc_ext3_mb_order2_req = create_proc_entry( -+ EXT3_MB_ORDER2_REQ, -+ S_IFREG | S_IRUGO | S_IWUSR, proc_root_ext3); -+ if (proc_ext3_mb_order2_req == NULL) { -+ printk(KERN_ERR "EXT3-fs: Unable to create %s\n", -+ EXT3_MB_ORDER2_REQ); -+ remove_proc_entry(EXT3_MB_MIN_TO_SCAN_NAME, proc_root_ext3); -+ remove_proc_entry(EXT3_MB_MAX_TO_SCAN_NAME, proc_root_ext3); -+ remove_proc_entry(EXT3_MB_STATS_NAME, proc_root_ext3); -+ remove_proc_entry(EXT3_ROOT, proc_root_fs); -+ return -EIO; -+ } -+ -+ proc_ext3_mb_order2_req->data = NULL; -+ proc_ext3_mb_order2_req->read_proc = ext3_mb_order2_req_read; -+ proc_ext3_mb_order2_req->write_proc = ext3_mb_order2_req_write; -+ -+ return 0; -+} -+ -+void exit_ext3_proc(void) -+{ -+ remove_proc_entry(EXT3_MB_STATS_NAME, proc_root_ext3); -+ remove_proc_entry(EXT3_MB_MAX_TO_SCAN_NAME, proc_root_ext3); -+ remove_proc_entry(EXT3_MB_MIN_TO_SCAN_NAME, proc_root_ext3); -+ remove_proc_entry(EXT3_MB_ORDER2_REQ, proc_root_ext3); -+ remove_proc_entry(EXT3_ROOT, proc_root_fs); -+} -Index: linux-2.6.5-7.282-full/fs/ext3/Makefile -=================================================================== ---- linux-2.6.5-7.282-full.orig/fs/ext3/Makefile 2006-10-24 22:18:28.000000000 +0400 -+++ linux-2.6.5-7.282-full/fs/ext3/Makefile 2006-10-24 22:18:28.000000000 +0400 -@@ -6,7 +6,7 @@ obj-$(CONFIG_EXT3_FS) += ext3.o - - ext3-y := balloc.o bitmap.o dir.o file.o fsync.o ialloc.o inode.o iopen.o \ - ioctl.o namei.o super.o symlink.o hash.o \ -- extents.o -+ extents.o mballoc.o - - ext3-$(CONFIG_EXT3_FS_XATTR) += xattr.o xattr_user.o xattr_trusted.o - ext3-$(CONFIG_EXT3_FS_POSIX_ACL) += acl.o diff --git a/ldiskfs/kernel_patches/patches/ext3-mballoc3-core.patch b/ldiskfs/kernel_patches/patches/ext3-mballoc3-core.patch new file mode 100644 index 0000000000000000000000000000000000000000..a6033d1d19151df3f57cc701c141af6f7fffa5e9 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-mballoc3-core.patch @@ -0,0 +1,4528 @@ +Index: linux-2.6.9-full/include/linux/ext3_fs_sb.h +=================================================================== +--- linux-2.6.9-full.orig/include/linux/ext3_fs_sb.h 2006-05-18 23:57:04.000000000 +0400 ++++ linux-2.6.9-full/include/linux/ext3_fs_sb.h 2007-03-28 15:42:16.000000000 +0400 +@@ -81,6 +81,61 @@ struct ext3_sb_info { + char *s_qf_names[MAXQUOTAS]; /* Names of quota files with journalled quota */ + int s_jquota_fmt; /* Format of quota to use */ + #endif ++ ++ /* for buddy allocator */ ++ struct ext3_group_info ***s_group_info; ++ struct inode *s_buddy_cache; ++ long s_blocks_reserved; ++ spinlock_t s_reserve_lock; ++ struct list_head s_active_transaction; ++ struct list_head s_closed_transaction; ++ struct list_head s_committed_transaction; ++ spinlock_t s_md_lock; ++ tid_t s_last_transaction; ++ unsigned short *s_mb_offsets, *s_mb_maxs; ++ ++ /* tunables */ ++ unsigned long s_mb_factor; ++ unsigned long s_stripe; ++ unsigned long s_mb_stream_request; ++ unsigned long s_mb_max_to_scan; ++ unsigned long s_mb_min_to_scan; ++ unsigned long s_mb_max_groups_to_scan; ++ unsigned long s_mb_stats; ++ unsigned long s_mb_order2_reqs; ++ ++ /* history to debug policy */ ++ struct ext3_mb_history *s_mb_history; ++ int s_mb_history_cur; ++ int s_mb_history_max; ++ int s_mb_history_num; ++ struct proc_dir_entry *s_mb_proc; ++ spinlock_t s_mb_history_lock; ++ int s_mb_history_filter; ++ ++ /* stats for buddy allocator */ ++ spinlock_t s_mb_pa_lock; ++ atomic_t s_bal_reqs; /* number of reqs with len > 1 */ ++ atomic_t s_bal_success; /* we found long enough chunks */ ++ atomic_t s_bal_allocated; /* in blocks */ ++ atomic_t s_bal_ex_scanned; /* total extents scanned */ ++ atomic_t s_bal_goals; /* goal hits */ ++ atomic_t s_bal_breaks; /* too long searches */ ++ atomic_t s_bal_2orders; /* 2^order hits */ ++ spinlock_t s_bal_lock; ++ unsigned long s_mb_buddies_generated; ++ unsigned long long s_mb_generation_time; ++ atomic_t s_mb_lost_chunks; ++ atomic_t s_mb_preallocated; ++ atomic_t s_mb_discarded; ++ ++ /* locality groups */ ++ struct ext3_locality_group *s_locality_groups; ++ + }; + ++#define EXT3_GROUP_INFO(sb, group) \ ++ EXT3_SB(sb)->s_group_info[(group) >> EXT3_DESC_PER_BLOCK_BITS(sb)] \ ++ [(group) & (EXT3_DESC_PER_BLOCK(sb) - 1)] ++ + #endif /* _LINUX_EXT3_FS_SB */ +Index: linux-2.6.9-full/include/linux/ext3_fs.h +=================================================================== +--- linux-2.6.9-full.orig/include/linux/ext3_fs.h 2007-03-28 01:29:39.000000000 +0400 ++++ linux-2.6.9-full/include/linux/ext3_fs.h 2007-03-28 15:45:07.000000000 +0400 +@@ -57,6 +57,30 @@ struct statfs; + #define ext3_debug(f, a...) do {} while (0) + #endif + ++#define EXT3_MULTIBLOCK_ALLOCATOR 1 ++ ++#define EXT3_MB_HINT_MERGE 1 /* prefer goal again. length */ ++#define EXT3_MB_HINT_RESERVED 2 /* blocks already reserved */ ++#define EXT3_MB_HINT_METADATA 4 /* metadata is being allocated */ ++#define EXT3_MB_HINT_FIRST 8 /* first blocks in the file */ ++#define EXT3_MB_HINT_BEST 16 /* search for the best chunk */ ++#define EXT3_MB_HINT_DATA 32 /* data is being allocated */ ++#define EXT3_MB_HINT_NOPREALLOC 64 /* don't preallocate (for tails) */ ++#define EXT3_MB_HINT_GROUP_ALLOC 128 /* allocate for locality group */ ++#define EXT3_MB_HINT_GOAL_ONLY 256 /* allocate goal blocks or none */ ++ ++struct ext3_allocation_request { ++ struct inode *inode; /* target inode for block we're allocating */ ++ unsigned long logical; /* logical block in target inode */ ++ unsigned long goal; /* phys. target (a hint) */ ++ unsigned long lleft; /* the closest logical allocated block to the left */ ++ unsigned long pleft; /* phys. block for ^^^ */ ++ unsigned long lright; /* the closest logical allocated block to the right */ ++ unsigned long pright; /* phys. block for ^^^ */ ++ unsigned long len; /* how many blocks we want to allocate */ ++ unsigned long flags; /* flags. see above EXT3_MB_HINT_* */ ++}; ++ + /* + * Special inodes numbers + */ +@@ -404,6 +413,14 @@ + #define ext3_find_first_zero_bit ext2_find_first_zero_bit + #define ext3_find_next_zero_bit ext2_find_next_zero_bit + ++#ifndef ext2_find_next_le_bit ++#ifdef __LITTLE_ENDIAN ++#define ext2_find_next_le_bit(addr, size, off) find_next_bit((addr), (size), (off)) ++#else ++#error "mballoc needs a patch for big-endian systems - CFS bug 10634" ++#endif /* __LITTLE_ENDIAN */ ++#endif /* !ext2_find_next_le_bit */ ++ + /* + * Maximal mount counts between two filesystem checks + */ +@@ -763,6 +787,20 @@ extern unsigned long ext3_count_dirs (st + extern void ext3_check_inodes_bitmap (struct super_block *); + extern unsigned long ext3_count_free (struct buffer_head *, unsigned); + ++/* mballoc.c */ ++extern long ext3_mb_stats; ++extern long ext3_mb_max_to_scan; ++extern int ext3_mb_init(struct super_block *, int); ++extern int ext3_mb_release(struct super_block *); ++extern unsigned long ext3_mb_new_blocks(handle_t *, struct ext3_allocation_request *, int *); ++extern int ext3_mb_reserve_blocks(struct super_block *, int); ++extern void ext3_mb_release_blocks(struct super_block *, int); ++extern void ext3_mb_release_blocks(struct super_block *, int); ++extern void ext3_mb_discard_inode_preallocations(struct inode *); ++extern int __init init_ext3_proc(void); ++extern void exit_ext3_proc(void); ++extern void ext3_mb_free_blocks(handle_t *, struct inode *, unsigned long, unsigned long, int, int *); ++ + + /* inode.c */ + extern int ext3_block_truncate_page(handle_t *, struct page *, +Index: linux-2.6.9-full/fs/ext3/super.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/super.c 2007-03-28 01:29:38.000000000 +0400 ++++ linux-2.6.9-full/fs/ext3/super.c 2007-03-28 15:42:16.000000000 +0400 +@@ -394,6 +394,7 @@ void ext3_put_super (struct super_block + struct ext3_super_block *es = sbi->s_es; + int i; + ++ ext3_mb_release(sb); + ext3_ext_release(sb); + ext3_xattr_put_super(sb); + journal_destroy(sbi->s_journal); +@@ -463,6 +464,8 @@ static struct inode *ext3_alloc_inode(st + ei->vfs_inode.i_version = 1; + + memset(&ei->i_cached_extent, 0, sizeof(ei->i_cached_extent)); ++ INIT_LIST_HEAD(&ei->i_prealloc_list); ++ spin_lock_init(&ei->i_prealloc_lock); + return &ei->vfs_inode; + } + +@@ -2433,7 +2436,13 @@ static struct file_system_type ext3_fs_t + + static int __init init_ext3_fs(void) + { +- int err = init_ext3_xattr(); ++ int err; ++ ++ err = init_ext3_proc(); ++ if (err) ++ return err; ++ ++ err = init_ext3_xattr(); + if (err) + return err; + err = init_inodecache(); +@@ -2455,6 +2464,7 @@ static void __exit exit_ext3_fs(void) + unregister_filesystem(&ext3_fs_type); + destroy_inodecache(); + exit_ext3_xattr(); ++ exit_ext3_proc(); + } + + int ext3_prep_san_write(struct inode *inode, long *blocks, +Index: linux-2.6.9-full/fs/ext3/mballoc.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/mballoc.c 2007-02-13 18:39:59.640066087 +0300 ++++ linux-2.6.9-full/fs/ext3/mballoc.c 2007-03-29 00:28:40.000000000 +0400 +@@ -0,0 +1,4342 @@ ++/* ++ * Copyright (c) 2003-2006, Cluster File Systems, Inc, info@clusterfs.com ++ * Written by Alex Tomas <alex@clusterfs.com> ++ * ++ * This program is free software; you can redistribute it and/or modify ++ * it under the terms of the GNU General Public License version 2 as ++ * published by the Free Software Foundation. ++ * ++ * This program is distributed in the hope that it will be useful, ++ * but WITHOUT ANY WARRANTY; without even the implied warranty of ++ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the ++ * GNU General Public License for more details. ++ * ++ * You should have received a copy of the GNU General Public Licens ++ * along with this program; if not, write to the Free Software ++ * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111- ++ */ ++ ++ ++/* ++ * mballoc.c contains the multiblocks allocation routines ++ */ ++ ++#include <linux/time.h> ++#include <linux/fs.h> ++#include <linux/namei.h> ++#include <linux/ext3_jbd.h> ++#include <linux/jbd.h> ++#include <linux/ext3_fs.h> ++#include <linux/quotaops.h> ++#include <linux/buffer_head.h> ++#include <linux/module.h> ++#include <linux/swap.h> ++#include <linux/proc_fs.h> ++#include <linux/pagemap.h> ++#include <linux/seq_file.h> ++#include <linux/version.h> ++ ++/* ++ * MUSTDO: ++ * - test ext3_ext_search_left() and ext3_ext_search_right() ++ * - search for metadata in few groups ++ * ++ * TODO v4: ++ * - normalization should take into account whether file is still open ++ * - discard preallocations if no free space left (policy?) ++ * - don't normalize tails ++ * - quota ++ * - reservation for superuser ++ * ++ * TODO v3: ++ * - bitmap read-ahead (proposed by Oleg Drokin aka green) ++ * - track min/max extents in each group for better group selection ++ * - mb_mark_used() may allocate chunk right after splitting buddy ++ * - tree of groups sorted by number of free blocks ++ * - error handling ++ */ ++ ++/* ++ * mballoc operates on the following data: ++ * - on-disk bitmap ++ * - in-core buddy (actually includes buddy and bitmap) ++ * - preallocation descriptors (PAs) ++ * ++ * there are two types of preallocations: ++ * - inode ++ * assiged to specific inode and can be used for this inode only. ++ * it describes part of inode's space preallocated to specific ++ * physical blocks. any block from that preallocated can be used ++ * independent. the descriptor just tracks number of blocks left ++ * unused. so, before taking some block from descriptor, one must ++ * make sure corresponded logical block isn't allocated yet. this ++ * also means that freeing any block within descriptor's range ++ * must discard all preallocated blocks. ++ * - locality group ++ * assigned to specific locality group which does not translate to ++ * permanent set of inodes: inode can join and leave group. space ++ * from this type of preallocation can be used for any inode. thus ++ * it's consumed from the beginning to the end. ++ * ++ * relation between them can be expressed as: ++ * in-core buddy = on-disk bitmap + preallocation descriptors ++ * ++ * this mean blocks mballoc considers used are: ++ * - allocated blocks (persistent) ++ * - preallocated blocks (non-persistent) ++ * ++ * consistency in mballoc world means that at any time a block is either ++ * free or used in ALL structures. notice: "any time" should not be read ++ * literally -- time is discrete and delimited by locks. ++ * ++ * to keep it simple, we don't use block numbers, instead we count number of ++ * blocks: how many blocks marked used/free in on-disk bitmap, buddy and PA. ++ * ++ * all operations can be expressed as: ++ * - init buddy: buddy = on-disk + PAs ++ * - new PA: buddy += N; PA = N ++ * - use inode PA: on-disk += N; PA -= N ++ * - discard inode PA buddy -= on-disk - PA; PA = 0 ++ * - use locality group PA on-disk += N; PA -= N ++ * - discard locality group PA buddy -= PA; PA = 0 ++ * note: 'buddy -= on-disk - PA' is used to show that on-disk bitmap ++ * is used in real operation because we can't know actual used ++ * bits from PA, only from on-disk bitmap ++ * ++ * if we follow this strict logic, then all operations above should be atomic. ++ * given some of them can block, we'd have to use something like semaphores ++ * killing performance on high-end SMP hardware. let's try to relax it using ++ * the following knowledge: ++ * 1) if buddy is referenced, it's already initialized ++ * 2) while block is used in buddy and the buddy is referenced, ++ * nobody can re-allocate that block ++ * 3) we work on bitmaps and '+' actually means 'set bits'. if on-disk has ++ * bit set and PA claims same block, it's OK. IOW, one can set bit in ++ * on-disk bitmap if buddy has same bit set or/and PA covers corresponded ++ * block ++ * ++ * so, now we're building a concurrency table: ++ * - init buddy vs. ++ * - new PA ++ * blocks for PA are allocated in the buddy, buddy must be referenced ++ * until PA is linked to allocation group to avoid concurrent buddy init ++ * - use inode PA ++ * we need to make sure that either on-disk bitmap or PA has uptodate data ++ * given (3) we care that PA-=N operation doesn't interfere with init ++ * - discard inode PA ++ * the simplest way would be to have buddy initialized by the discard ++ * - use locality group PA ++ * again PA-=N must be serialized with init ++ * - discard locality group PA ++ * the simplest way would be to have buddy initialized by the discard ++ * - new PA vs. ++ * - use inode PA ++ * i_truncate_mutex serializes them ++ * - discard inode PA ++ * discard process must wait until PA isn't used by another process ++ * - use locality group PA ++ * some mutex should serialize them ++ * - discard locality group PA ++ * discard process must wait until PA isn't used by another process ++ * - use inode PA ++ * - use inode PA ++ * i_truncate_mutex or another mutex should serializes them ++ * - discard inode PA ++ * discard process must wait until PA isn't used by another process ++ * - use locality group PA ++ * nothing wrong here -- they're different PAs covering different blocks ++ * - discard locality group PA ++ * discard process must wait until PA isn't used by another process ++ * ++ * now we're ready to make few consequences: ++ * - PA is referenced and while it is no discard is possible ++ * - PA is referenced until block isn't marked in on-disk bitmap ++ * - PA changes only after on-disk bitmap ++ * - discard must not compete with init. either init is done before ++ * any discard or they're serialized somehow ++ * - buddy init as sum of on-disk bitmap and PAs is done atomically ++ * ++ * a special case when we've used PA to emptiness. no need to modify buddy ++ * in this case, but we should care about concurrent init ++ * ++ */ ++ ++ /* ++ * Logic in few words: ++ * ++ * - allocation: ++ * load group ++ * find blocks ++ * mark bits in on-disk bitmap ++ * release group ++ * ++ * - use preallocation: ++ * find proper PA (per-inode or group) ++ * load group ++ * mark bits in on-disk bitmap ++ * release group ++ * release PA ++ * ++ * - free: ++ * load group ++ * mark bits in on-disk bitmap ++ * release group ++ * ++ * - discard preallocations in group: ++ * mark PAs deleted ++ * move them onto local list ++ * load on-disk bitmap ++ * load group ++ * remove PA from object (inode or locality group) ++ * mark free blocks in-core ++ * ++ * - discard inode's preallocations: ++ */ ++ ++/* ++ * Locking rules ++ * ++ * Locks: ++ * - bitlock on a group (group) ++ * - object (inode/locality) (object) ++ * - per-pa lock (pa) ++ * ++ * Paths: ++ * - new pa ++ * object ++ * group ++ * ++ * - find and use pa: ++ * pa ++ * ++ * - release consumed pa: ++ * pa ++ * group ++ * object ++ * ++ * - generate in-core bitmap: ++ * group ++ * pa ++ * ++ * - discard all for given object (inode, locality group): ++ * object ++ * pa ++ * group ++ * ++ * - discard all for given group: ++ * group ++ * pa ++ * group ++ * object ++ * ++ */ ++ ++/* ++ * with AGGRESSIVE_CHECK allocator runs consistency checks over ++ * structures. these checks slow things down a lot ++ */ ++#define AGGRESSIVE_CHECK__ ++ ++/* ++ * with DOUBLE_CHECK defined mballoc creates persistent in-core ++ * bitmaps, maintains and uses them to check for double allocations ++ */ ++#define DOUBLE_CHECK__ ++ ++/* ++ */ ++#define MB_DEBUG__ ++#ifdef MB_DEBUG ++#define mb_debug(fmt,a...) printk(fmt, ##a) ++#else ++#define mb_debug(fmt,a...) ++#endif ++ ++/* ++ * with EXT3_MB_HISTORY mballoc stores last N allocations in memory ++ * and you can monitor it in /proc/fs/ext3/<dev>/mb_history ++ */ ++#define EXT3_MB_HISTORY ++#define EXT3_MB_HISTORY_ALLOC 1 /* allocation */ ++#define EXT3_MB_HISTORY_PREALLOC 2 /* preallocated blocks used */ ++#define EXT3_MB_HISTORY_DISCARD 4 /* preallocation discarded */ ++#define EXT3_MB_HISTORY_FREE 8 /* free */ ++ ++#define EXT3_MB_HISTORY_DEFAULT (EXT3_MB_HISTORY_ALLOC | \ ++ EXT3_MB_HISTORY_PREALLOC | \ ++ EXT3_MB_HISTORY_DISCARD | \ ++ EXT3_MB_HISTORY_FREE) ++ ++/* ++ * How long mballoc can look for a best extent (in found extents) ++ */ ++#define MB_DEFAULT_MAX_TO_SCAN 200 ++ ++/* ++ * How long mballoc must look for a best extent ++ */ ++#define MB_DEFAULT_MIN_TO_SCAN 10 ++ ++/* ++ * How many groups mballoc will scan looking for the best chunk ++ */ ++#define MB_DEFAULT_MAX_GROUPS_TO_SCAN 5 ++ ++/* ++ * with 'ext3_mb_stats' allocator will collect stats that will be ++ * shown at umount. The collecting costs though! ++ */ ++#define MB_DEFAULT_STATS 1 ++ ++/* ++ * files smaller than MB_DEFAULT_STREAM_THRESHOLD are served ++ * by the stream allocator, which purpose is to pack requests ++ * as close each to other as possible to produce smooth I/O traffic ++ */ ++#define MB_DEFAULT_STREAM_THRESHOLD 16 /* 64K */ ++ ++/* ++ * for which requests use 2^N search using buddies ++ */ ++#define MB_DEFAULT_ORDER2_REQS 8 ++ ++/* ++ * default stripe size = 1MB ++ */ ++#define MB_DEFAULT_STRIPE 256 ++ ++static kmem_cache_t *ext3_pspace_cachep = NULL; ++ ++#ifdef EXT3_BB_MAX_BLOCKS ++#undef EXT3_BB_MAX_BLOCKS ++#endif ++#define EXT3_BB_MAX_BLOCKS 30 ++ ++struct ext3_free_metadata { ++ unsigned short group; ++ unsigned short num; ++ unsigned short blocks[EXT3_BB_MAX_BLOCKS]; ++ struct list_head list; ++}; ++ ++struct ext3_group_info { ++ unsigned long bb_state; ++ unsigned long bb_tid; ++ struct ext3_free_metadata *bb_md_cur; ++ unsigned short bb_first_free; ++ unsigned short bb_free; ++ unsigned short bb_fragments; ++ struct list_head bb_prealloc_list; ++#ifdef DOUBLE_CHECK ++ void *bb_bitmap; ++#endif ++ unsigned short bb_counters[]; ++}; ++ ++#define EXT3_GROUP_INFO_NEED_INIT_BIT 0 ++#define EXT3_GROUP_INFO_LOCKED_BIT 1 ++ ++#define EXT3_MB_GRP_NEED_INIT(grp) \ ++ (test_bit(EXT3_GROUP_INFO_NEED_INIT_BIT, &(grp)->bb_state)) ++ ++ ++struct ext3_prealloc_space { ++ struct list_head pa_inode_list; ++ struct list_head pa_group_list; ++ union { ++ struct list_head pa_tmp_list; ++ struct rcu_head pa_rcu; ++ } u; ++ spinlock_t pa_lock; ++ atomic_t pa_count; ++ unsigned pa_deleted; ++ unsigned long pa_pstart; /* phys. block */ ++ unsigned long pa_lstart; /* log. block */ ++ unsigned short pa_len; /* len of preallocated chunk */ ++ unsigned short pa_free; /* how many blocks are free */ ++ unsigned short pa_linear; /* consumed in one direction ++ * strictly, for group prealloc */ ++ spinlock_t *pa_obj_lock; ++ struct inode *pa_inode; /* hack, for history only */ ++}; ++ ++ ++struct ext3_free_extent { ++ unsigned long fe_logical; ++ unsigned long fe_start; ++ unsigned long fe_group; ++ unsigned long fe_len; ++}; ++ ++/* ++ * Locality group: ++ * we try to group all related changes together ++ * so that writeback can flush/allocate them together as well ++ */ ++struct ext3_locality_group { ++ /* for allocator */ ++ struct semaphore lg_sem; /* to serialize allocates */ ++ struct list_head lg_prealloc_list;/* list of preallocations */ ++ spinlock_t lg_prealloc_lock; ++}; ++ ++struct ext3_allocation_context { ++ struct inode *ac_inode; ++ struct super_block *ac_sb; ++ ++ /* original request */ ++ struct ext3_free_extent ac_o_ex; ++ ++ /* goal request (after normalization) */ ++ struct ext3_free_extent ac_g_ex; ++ ++ /* the best found extent */ ++ struct ext3_free_extent ac_b_ex; ++ ++ /* copy of the bext found extent taken before preallocation efforts */ ++ struct ext3_free_extent ac_f_ex; ++ ++ /* number of iterations done. we have to track to limit searching */ ++ unsigned long ac_ex_scanned; ++ __u16 ac_groups_scanned; ++ __u16 ac_found; ++ __u16 ac_tail; ++ __u16 ac_buddy; ++ __u16 ac_flags; /* allocation hints */ ++ __u8 ac_status; ++ __u8 ac_criteria; ++ __u8 ac_repeats; ++ __u8 ac_2order; /* if request is to allocate 2^N blocks and ++ * N > 0, the field stores N, otherwise 0 */ ++ __u8 ac_op; /* operation, for history only */ ++ struct page *ac_bitmap_page; ++ struct page *ac_buddy_page; ++ struct ext3_prealloc_space *ac_pa; ++ struct ext3_locality_group *ac_lg; ++}; ++ ++#define AC_STATUS_CONTINUE 1 ++#define AC_STATUS_FOUND 2 ++#define AC_STATUS_BREAK 3 ++ ++struct ext3_mb_history { ++ struct ext3_free_extent orig; /* orig allocation */ ++ struct ext3_free_extent goal; /* goal allocation */ ++ struct ext3_free_extent result; /* result allocation */ ++ unsigned pid; ++ unsigned ino; ++ __u16 found; /* how many extents have been found */ ++ __u16 groups; /* how many groups have been scanned */ ++ __u16 tail; /* what tail broke some buddy */ ++ __u16 buddy; /* buddy the tail ^^^ broke */ ++ __u16 flags; ++ __u8 cr:3; /* which phase the result extent was found at */ ++ __u8 op:4; ++ __u8 merged:1; ++}; ++ ++struct ext3_buddy { ++ struct page *bd_buddy_page; ++ void *bd_buddy; ++ struct page *bd_bitmap_page; ++ void *bd_bitmap; ++ struct ext3_group_info *bd_info; ++ struct super_block *bd_sb; ++ __u16 bd_blkbits; ++ __u16 bd_group; ++}; ++#define EXT3_MB_BITMAP(e3b) ((e3b)->bd_bitmap) ++#define EXT3_MB_BUDDY(e3b) ((e3b)->bd_buddy) ++ ++#ifndef EXT3_MB_HISTORY ++#define ext3_mb_store_history(ac) ++#else ++static void ext3_mb_store_history(struct ext3_allocation_context *ac); ++#endif ++ ++#define in_range(b, first, len) ((b) >= (first) && (b) <= (first) + (len) - 1) ++ ++static struct proc_dir_entry *proc_root_ext3; ++ ++int ext3_create (struct inode *, struct dentry *, int, struct nameidata *); ++struct buffer_head * read_block_bitmap(struct super_block *, unsigned int); ++unsigned long ext3_new_blocks_old(handle_t *handle, struct inode *inode, ++ unsigned long goal, unsigned long *count, int *errp); ++void ext3_mb_release_blocks(struct super_block *, int); ++void ext3_mb_poll_new_transaction(struct super_block *, handle_t *); ++void ext3_mb_free_committed_blocks(struct super_block *); ++void ext3_mb_generate_from_pa(struct super_block *sb, void *bitmap, int group); ++void ext3_mb_free_consumed_preallocations(struct ext3_allocation_context *ac); ++void ext3_mb_return_to_preallocation(struct inode *inode, struct ext3_buddy *e3b, ++ sector_t block, int count); ++void ext3_mb_show_ac(struct ext3_allocation_context *ac); ++void ext3_mb_check_with_pa(struct ext3_buddy *e3b, int first, int count); ++void ext3_mb_put_pa(struct ext3_allocation_context *, struct super_block *, struct ext3_prealloc_space *pa); ++int ext3_mb_init_per_dev_proc(struct super_block *sb); ++int ext3_mb_destroy_per_dev_proc(struct super_block *sb); ++ ++/* ++ * Calculate the block group number and offset, given a block number ++ */ ++static void ext3_get_group_no_and_offset(struct super_block *sb, ++ unsigned long blocknr, ++ unsigned long *blockgrpp, ++ unsigned long *offsetp) ++{ ++ struct ext3_super_block *es = EXT3_SB(sb)->s_es; ++ unsigned long offset; ++ ++ blocknr = blocknr - le32_to_cpu(es->s_first_data_block); ++ offset = blocknr % EXT3_BLOCKS_PER_GROUP(sb); ++ blocknr = blocknr / EXT3_BLOCKS_PER_GROUP(sb); ++ if (offsetp) ++ *offsetp = offset; ++ if (blockgrpp) ++ *blockgrpp = blocknr; ++ ++} ++ ++static inline void ++ext3_lock_group(struct super_block *sb, int group) ++{ ++ bit_spin_lock(EXT3_GROUP_INFO_LOCKED_BIT, ++ &EXT3_GROUP_INFO(sb, group)->bb_state); ++} ++ ++static inline void ++ext3_unlock_group(struct super_block *sb, int group) ++{ ++ bit_spin_unlock(EXT3_GROUP_INFO_LOCKED_BIT, ++ &EXT3_GROUP_INFO(sb, group)->bb_state); ++} ++ ++static inline int ++ext3_is_group_locked(struct super_block *sb, int group) ++{ ++ return bit_spin_is_locked(EXT3_GROUP_INFO_LOCKED_BIT, ++ &EXT3_GROUP_INFO(sb, group)->bb_state); ++} ++ ++unsigned long ext3_grp_offs_to_block(struct super_block *sb, ++ struct ext3_free_extent *fex) ++{ ++ unsigned long block; ++ ++ block = (unsigned long) fex->fe_group * EXT3_BLOCKS_PER_GROUP(sb) ++ + fex->fe_start ++ + le32_to_cpu(EXT3_SB(sb)->s_es->s_first_data_block); ++ return block; ++} ++ ++#if BITS_PER_LONG == 64 ++#define mb_correct_addr_and_bit(bit,addr) \ ++{ \ ++ bit += ((unsigned long) addr & 7UL) << 3; \ ++ addr = (void *) ((unsigned long) addr & ~7UL); \ ++} ++#elif BITS_PER_LONG == 32 ++#define mb_correct_addr_and_bit(bit,addr) \ ++{ \ ++ bit += ((unsigned long) addr & 3UL) << 3; \ ++ addr = (void *) ((unsigned long) addr & ~3UL); \ ++} ++#else ++#error "how many bits you are?!" ++#endif ++ ++static inline int mb_test_bit(int bit, void *addr) ++{ ++ mb_correct_addr_and_bit(bit,addr); ++ return ext2_test_bit(bit, addr); ++} ++ ++static inline void mb_set_bit(int bit, void *addr) ++{ ++ mb_correct_addr_and_bit(bit,addr); ++ ext2_set_bit(bit, addr); ++} ++ ++static inline void mb_set_bit_atomic(int bit, void *addr) ++{ ++ mb_correct_addr_and_bit(bit,addr); ++ ext2_set_bit_atomic(NULL, bit, addr); ++} ++ ++static inline void mb_clear_bit(int bit, void *addr) ++{ ++ mb_correct_addr_and_bit(bit,addr); ++ ext2_clear_bit(bit, addr); ++} ++ ++static inline void mb_clear_bit_atomic(int bit, void *addr) ++{ ++ mb_correct_addr_and_bit(bit,addr); ++ ext2_clear_bit_atomic(NULL, bit, addr); ++} ++ ++static inline int mb_find_next_zero_bit(void *addr, int max, int start) ++{ ++ int fix; ++#if BITS_PER_LONG == 64 ++ fix = ((unsigned long) addr & 7UL) << 3; ++ addr = (void *) ((unsigned long) addr & ~7UL); ++#elif BITS_PER_LONG == 32 ++ fix = ((unsigned long) addr & 3UL) << 3; ++ addr = (void *) ((unsigned long) addr & ~3UL); ++#else ++#error "how many bits you are?!" ++#endif ++ max += fix; ++ start += fix; ++ return ext2_find_next_zero_bit(addr, max, start) - fix; ++} ++ ++static inline int mb_find_next_bit(void *addr, int max, int start) ++{ ++ int fix; ++#if BITS_PER_LONG == 64 ++ fix = ((unsigned long) addr & 7UL) << 3; ++ addr = (void *) ((unsigned long) addr & ~7UL); ++#elif BITS_PER_LONG == 32 ++ fix = ((unsigned long) addr & 3UL) << 3; ++ addr = (void *) ((unsigned long) addr & ~3UL); ++#else ++#error "how many bits you are?!" ++#endif ++ max += fix; ++ start += fix; ++ ++#ifdef __BIG_ENDIAN ++#else ++ return find_next_bit(addr, max, start) - fix; ++#endif ++} ++ ++static inline void *mb_find_buddy(struct ext3_buddy *e3b, int order, int *max) ++{ ++ char *bb; ++ ++ BUG_ON(EXT3_MB_BITMAP(e3b) == EXT3_MB_BUDDY(e3b)); ++ BUG_ON(max == NULL); ++ ++ if (order > e3b->bd_blkbits + 1) { ++ *max = 0; ++ return NULL; ++ } ++ ++ /* at order 0 we see each particular block */ ++ *max = 1 << (e3b->bd_blkbits + 3); ++ if (order == 0) ++ return EXT3_MB_BITMAP(e3b); ++ ++ bb = EXT3_MB_BUDDY(e3b) + EXT3_SB(e3b->bd_sb)->s_mb_offsets[order]; ++ *max = EXT3_SB(e3b->bd_sb)->s_mb_maxs[order]; ++ ++ return bb; ++} ++ ++#ifdef DOUBLE_CHECK ++void mb_free_blocks_double(struct ext3_buddy *e3b, int first, int count) ++{ ++ int i; ++ if (unlikely(e3b->bd_info->bb_bitmap == NULL)) ++ return; ++ BUG_ON(!ext3_is_group_locked(e3b->bd_sb, e3b->bd_group)); ++ for (i = 0; i < count; i++) { ++ BUG_ON(!mb_test_bit(first + i, e3b->bd_info->bb_bitmap)); ++ mb_clear_bit(first + i, e3b->bd_info->bb_bitmap); ++ } ++} ++ ++void mb_mark_used_double(struct ext3_buddy *e3b, int first, int count) ++{ ++ int i; ++ if (unlikely(e3b->bd_info->bb_bitmap == NULL)) ++ return; ++ BUG_ON(!ext3_is_group_locked(e3b->bd_sb, e3b->bd_group)); ++ for (i = 0; i < count; i++) { ++ BUG_ON(mb_test_bit(first + i, e3b->bd_info->bb_bitmap)); ++ mb_set_bit(first + i, e3b->bd_info->bb_bitmap); ++ } ++} ++ ++void mb_cmp_bitmaps(struct ext3_buddy *e3b, void *bitmap) ++{ ++ if (memcmp(e3b->bd_info->bb_bitmap, bitmap, e3b->bd_sb->s_blocksize)) { ++ unsigned char *b1, *b2; ++ int i; ++ b1 = (unsigned char *) e3b->bd_info->bb_bitmap; ++ b2 = (unsigned char *) bitmap; ++ for (i = 0; i < e3b->bd_sb->s_blocksize; i++) { ++ if (b1[i] != b2[i]) { ++ printk("corruption in group %u at byte %u(%u): " ++ "%x in copy != %x on disk/prealloc\n", ++ e3b->bd_group, i, i * 8, b1[i], b2[i]); ++ BUG(); ++ } ++ } ++ } ++} ++ ++#else ++#define mb_free_blocks_double(a,b,c) ++#define mb_mark_used_double(a,b,c) ++#define mb_cmp_bitmaps(a,b) ++#endif ++ ++#ifdef AGGRESSIVE_CHECK ++ ++#define MB_CHECK_ASSERT(assert) \ ++do { \ ++ if (!(assert)) { \ ++ printk (KERN_EMERG \ ++ "Assertion failure in %s() at %s:%d: \"%s\"\n", \ ++ function, file, line, # assert); \ ++ BUG(); \ ++ } \ ++} while (0) ++ ++static int __mb_check_buddy(struct ext3_buddy *e3b, char *file, ++ const char *function, int line) ++{ ++ struct super_block *sb = e3b->bd_sb; ++ int order = e3b->bd_blkbits + 1; ++ int max, max2, i, j, k, count; ++ struct ext3_group_info *grp; ++ int fragments = 0, fstart; ++ struct list_head *cur; ++ void *buddy, *buddy2; ++ ++ if (!test_opt(sb, MBALLOC)) ++ return 0; ++ ++ { ++ static int mb_check_counter = 0; ++ if (mb_check_counter++ % 100 != 0) ++ return 0; ++ } ++ ++ while (order > 1) { ++ buddy = mb_find_buddy(e3b, order, &max); ++ MB_CHECK_ASSERT(buddy); ++ buddy2 = mb_find_buddy(e3b, order - 1, &max2); ++ MB_CHECK_ASSERT(buddy2); ++ MB_CHECK_ASSERT(buddy != buddy2); ++ MB_CHECK_ASSERT(max * 2 == max2); ++ ++ count = 0; ++ for (i = 0; i < max; i++) { ++ ++ if (mb_test_bit(i, buddy)) { ++ /* only single bit in buddy2 may be 1 */ ++ if (!mb_test_bit(i << 1, buddy2)) ++ MB_CHECK_ASSERT(mb_test_bit((i<<1)+1, buddy2)); ++ else if (!mb_test_bit((i << 1) + 1, buddy2)) ++ MB_CHECK_ASSERT(mb_test_bit(i << 1, buddy2)); ++ continue; ++ } ++ ++ /* both bits in buddy2 must be 0 */ ++ MB_CHECK_ASSERT(mb_test_bit(i << 1, buddy2)); ++ MB_CHECK_ASSERT(mb_test_bit((i << 1) + 1, buddy2)); ++ ++ for (j = 0; j < (1 << order); j++) { ++ k = (i * (1 << order)) + j; ++ MB_CHECK_ASSERT(!mb_test_bit(k, EXT3_MB_BITMAP(e3b))); ++ } ++ count++; ++ } ++ MB_CHECK_ASSERT(e3b->bd_info->bb_counters[order] == count); ++ order--; ++ } ++ ++ fstart = -1; ++ buddy = mb_find_buddy(e3b, 0, &max); ++ for (i = 0; i < max; i++) { ++ if (!mb_test_bit(i, buddy)) { ++ MB_CHECK_ASSERT(i >= e3b->bd_info->bb_first_free); ++ if (fstart == -1) { ++ fragments++; ++ fstart = i; ++ } ++ continue; ++ } ++ fstart = -1; ++ /* check used bits only */ ++ for (j = 0; j < e3b->bd_blkbits + 1; j++) { ++ buddy2 = mb_find_buddy(e3b, j, &max2); ++ k = i >> j; ++ MB_CHECK_ASSERT(k < max2); ++ MB_CHECK_ASSERT(mb_test_bit(k, buddy2)); ++ } ++ } ++ MB_CHECK_ASSERT(!EXT3_MB_GRP_NEED_INIT(e3b->bd_info)); ++ MB_CHECK_ASSERT(e3b->bd_info->bb_fragments == fragments); ++ ++ grp = EXT3_GROUP_INFO(sb, e3b->bd_group); ++ buddy = mb_find_buddy(e3b, 0, &max); ++ list_for_each(cur, &grp->bb_prealloc_list) { ++ unsigned long groupnr; ++ struct ext3_prealloc_space *pa; ++ pa = list_entry(cur, struct ext3_prealloc_space, group_list); ++ ext3_get_group_no_and_offset(sb, pa->pstart, &groupnr, &k); ++ MB_CHECK_ASSERT(groupnr == e3b->bd_group); ++ for (i = 0; i < pa->len; i++) ++ MB_CHECK_ASSERT(mb_test_bit(k + i, buddy)); ++ } ++ return 0; ++} ++#undef MB_CHECK_ASSERT ++#define mb_check_buddy(e3b) __mb_check_buddy(e3b,__FILE__,__FUNCTION__,__LINE__) ++#else ++#define mb_check_buddy(e3b) ++#endif ++ ++/* find most significant bit */ ++static int inline fmsb(unsigned short word) ++{ ++ int order; ++ ++ if (word > 255) { ++ order = 7; ++ word >>= 8; ++ } else { ++ order = -1; ++ } ++ ++ do { ++ order++; ++ word >>= 1; ++ } while (word != 0); ++ ++ return order; ++} ++ ++static void inline ++ext3_mb_mark_free_simple(struct super_block *sb, void *buddy, unsigned first, ++ int len, struct ext3_group_info *grp) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ unsigned short min, max, chunk, border; ++ ++ BUG_ON(len >= EXT3_BLOCKS_PER_GROUP(sb)); ++ ++ border = 2 << sb->s_blocksize_bits; ++ ++ while (len > 0) { ++ /* find how many blocks can be covered since this position */ ++ max = ffs(first | border) - 1; ++ ++ /* find how many blocks of power 2 we need to mark */ ++ min = fmsb(len); ++ ++ if (max < min) ++ min = max; ++ chunk = 1 << min; ++ ++ /* mark multiblock chunks only */ ++ grp->bb_counters[min]++; ++ if (min > 0) ++ mb_clear_bit(first >> min, buddy + sbi->s_mb_offsets[min]); ++ ++ len -= chunk; ++ first += chunk; ++ } ++} ++ ++static void ++ext3_mb_generate_buddy(struct super_block *sb, void *buddy, void *bitmap, ++ int group) ++{ ++ struct ext3_group_info *grp = EXT3_GROUP_INFO(sb, group); ++ unsigned short max = EXT3_BLOCKS_PER_GROUP(sb); ++ unsigned short i = 0, first, len; ++ unsigned free = 0, fragments = 0; ++ unsigned long long period = get_cycles(); ++ ++ /* initialize buddy from bitmap which is aggregation ++ * of on-disk bitmap and preallocations */ ++ i = mb_find_next_zero_bit(bitmap, max, 0); ++ grp->bb_first_free = i; ++ while (i < max) { ++ fragments++; ++ first = i; ++ i = ext2_find_next_le_bit(bitmap, max, i); ++ len = i - first; ++ free += len; ++ if (len > 1) ++ ext3_mb_mark_free_simple(sb, buddy, first, len, grp); ++ else ++ grp->bb_counters[0]++; ++ if (i < max) ++ i = mb_find_next_zero_bit(bitmap, max, i); ++ } ++ grp->bb_fragments = fragments; ++ ++ if (free != grp->bb_free) { ++ printk("EXT3-fs: group %u: %u blocks in bitmap, %u in gd\n", ++ group, free, grp->bb_free); ++ grp->bb_free = free; ++ } ++ ++ clear_bit(EXT3_GROUP_INFO_NEED_INIT_BIT, &grp->bb_state); ++ ++ period = get_cycles() - period; ++ spin_lock(&EXT3_SB(sb)->s_bal_lock); ++ EXT3_SB(sb)->s_mb_buddies_generated++; ++ EXT3_SB(sb)->s_mb_generation_time += period; ++ spin_unlock(&EXT3_SB(sb)->s_bal_lock); ++} ++ ++static int ext3_mb_init_cache(struct page *page, char *incore) ++{ ++ int blocksize, blocks_per_page, groups_per_page; ++ int err = 0, i, first_group, first_block; ++ struct super_block *sb; ++ struct buffer_head *bhs; ++ struct buffer_head **bh; ++ struct inode *inode; ++ char *data, *bitmap; ++ ++ mb_debug("init page %lu\n", page->index); ++ ++ inode = page->mapping->host; ++ sb = inode->i_sb; ++ blocksize = 1 << inode->i_blkbits; ++ blocks_per_page = PAGE_CACHE_SIZE / blocksize; ++ ++ groups_per_page = blocks_per_page >> 1; ++ if (groups_per_page == 0) ++ groups_per_page = 1; ++ ++ /* allocate buffer_heads to read bitmaps */ ++ if (groups_per_page > 1) { ++ err = -ENOMEM; ++ i = sizeof(struct buffer_head *) * groups_per_page; ++ bh = kmalloc(i, GFP_NOFS); ++ if (bh == NULL) ++ goto out; ++ memset(bh, 0, i); ++ } else ++ bh = &bhs; ++ ++ first_group = page->index * blocks_per_page / 2; ++ ++ /* read all groups the page covers into the cache */ ++ for (i = 0; i < groups_per_page; i++) { ++ struct ext3_group_desc * desc; ++ ++ if (first_group + i >= EXT3_SB(sb)->s_groups_count) ++ break; ++ ++ err = -EIO; ++ desc = ext3_get_group_desc(sb, first_group + i, NULL); ++ if (desc == NULL) ++ goto out; ++ ++ err = -ENOMEM; ++ bh[i] = sb_getblk(sb, le32_to_cpu(desc->bg_block_bitmap)); ++ if (bh[i] == NULL) ++ goto out; ++ ++ if (buffer_uptodate(bh[i])) ++ continue; ++ ++ lock_buffer(bh[i]); ++ if (buffer_uptodate(bh[i])) { ++ unlock_buffer(bh[i]); ++ continue; ++ } ++ ++ get_bh(bh[i]); ++ bh[i]->b_end_io = end_buffer_read_sync; ++ submit_bh(READ, bh[i]); ++ mb_debug("read bitmap for group %u\n", first_group + i); ++ } ++ ++ /* wait for I/O completion */ ++ for (i = 0; i < groups_per_page && bh[i]; i++) ++ wait_on_buffer(bh[i]); ++ ++ err = -EIO; ++ for (i = 0; i < groups_per_page && bh[i]; i++) ++ if (!buffer_uptodate(bh[i])) ++ goto out; ++ ++ first_block = page->index * blocks_per_page; ++ for (i = 0; i < blocks_per_page; i++) { ++ int group; ++ ++ group = (first_block + i) >> 1; ++ if (group >= EXT3_SB(sb)->s_groups_count) ++ break; ++ ++ data = page_address(page) + (i * blocksize); ++ bitmap = bh[group - first_group]->b_data; ++ ++ if ((first_block + i) & 1) { ++ /* this is block of buddy */ ++ BUG_ON(incore == NULL); ++ mb_debug("put buddy for group %u in page %lu/%x\n", ++ group, page->index, i * blocksize); ++ memset(data, 0xff, blocksize); ++ EXT3_GROUP_INFO(sb, group)->bb_fragments = 0; ++ memset(EXT3_GROUP_INFO(sb, group)->bb_counters, 0, ++ sizeof(unsigned short)*(sb->s_blocksize_bits+2)); ++ ext3_mb_generate_buddy(sb, data, incore, group); ++ incore = NULL; ++ } else { ++ /* this is block of bitmap */ ++ BUG_ON(incore != NULL); ++ mb_debug("put bitmap for group %u in page %lu/%x\n", ++ group, page->index, i * blocksize); ++ ++ /* see comments in ext3_mb_put_pa() */ ++ ext3_lock_group(sb, group); ++ memcpy(data, bitmap, blocksize); ++ ++ /* mark all preallocated blocks used in in-core bitmap */ ++ ext3_mb_generate_from_pa(sb, data, group); ++ ext3_unlock_group(sb, group); ++ ++ incore = data; ++ } ++ } ++ SetPageUptodate(page); ++ ++out: ++ if (bh) { ++ for (i = 0; i < groups_per_page && bh[i]; i++) ++ brelse(bh[i]); ++ if (bh != &bhs) ++ kfree(bh); ++ } ++ return err; ++} ++ ++static int ext3_mb_load_buddy(struct super_block *sb, int group, ++ struct ext3_buddy *e3b) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ struct inode *inode = sbi->s_buddy_cache; ++ int blocks_per_page, block, pnum, poff; ++ struct page *page; ++ ++ mb_debug("load group %u\n", group); ++ ++ blocks_per_page = PAGE_CACHE_SIZE / sb->s_blocksize; ++ ++ e3b->bd_blkbits = sb->s_blocksize_bits; ++ e3b->bd_info = EXT3_GROUP_INFO(sb, group); ++ e3b->bd_sb = sb; ++ e3b->bd_group = group; ++ e3b->bd_buddy_page = NULL; ++ e3b->bd_bitmap_page = NULL; ++ ++ block = group * 2; ++ pnum = block / blocks_per_page; ++ poff = block % blocks_per_page; ++ ++ /* we could use find_or_create_page(), but it locks page ++ * what we'd like to avoid in fast path ... */ ++ page = find_get_page(inode->i_mapping, pnum); ++ if (page == NULL || !PageUptodate(page)) { ++ if (page) ++ page_cache_release(page); ++ page = find_or_create_page(inode->i_mapping, pnum, GFP_NOFS); ++ if (page) { ++ BUG_ON(page->mapping != inode->i_mapping); ++ if (!PageUptodate(page)) { ++ ext3_mb_init_cache(page, NULL); ++ mb_cmp_bitmaps(e3b, page_address(page) + ++ (poff * sb->s_blocksize)); ++ } ++ unlock_page(page); ++ } ++ } ++ if (page == NULL || !PageUptodate(page)) ++ goto err; ++ e3b->bd_bitmap_page = page; ++ e3b->bd_bitmap = page_address(page) + (poff * sb->s_blocksize); ++ mark_page_accessed(page); ++ ++ block++; ++ pnum = block / blocks_per_page; ++ poff = block % blocks_per_page; ++ ++ page = find_get_page(inode->i_mapping, pnum); ++ if (page == NULL || !PageUptodate(page)) { ++ if (page) ++ page_cache_release(page); ++ page = find_or_create_page(inode->i_mapping, pnum, GFP_NOFS); ++ if (page) { ++ BUG_ON(page->mapping != inode->i_mapping); ++ if (!PageUptodate(page)) ++ ext3_mb_init_cache(page, e3b->bd_bitmap); ++ ++ unlock_page(page); ++ } ++ } ++ if (page == NULL || !PageUptodate(page)) ++ goto err; ++ e3b->bd_buddy_page = page; ++ e3b->bd_buddy = page_address(page) + (poff * sb->s_blocksize); ++ mark_page_accessed(page); ++ ++ BUG_ON(e3b->bd_bitmap_page == NULL); ++ BUG_ON(e3b->bd_buddy_page == NULL); ++ ++ return 0; ++ ++err: ++ if (e3b->bd_bitmap_page) ++ page_cache_release(e3b->bd_bitmap_page); ++ if (e3b->bd_buddy_page) ++ page_cache_release(e3b->bd_buddy_page); ++ e3b->bd_buddy = NULL; ++ e3b->bd_bitmap = NULL; ++ return -EIO; ++} ++ ++static void ext3_mb_release_desc(struct ext3_buddy *e3b) ++{ ++ if (e3b->bd_bitmap_page) ++ page_cache_release(e3b->bd_bitmap_page); ++ if (e3b->bd_buddy_page) ++ page_cache_release(e3b->bd_buddy_page); ++} ++ ++ ++static int mb_find_order_for_block(struct ext3_buddy *e3b, int block) ++{ ++ int order = 1; ++ void *bb; ++ ++ BUG_ON(EXT3_MB_BITMAP(e3b) == EXT3_MB_BUDDY(e3b)); ++ BUG_ON(block >= (1 << (e3b->bd_blkbits + 3))); ++ ++ bb = EXT3_MB_BUDDY(e3b); ++ while (order <= e3b->bd_blkbits + 1) { ++ block = block >> 1; ++ if (!mb_test_bit(block, bb)) { ++ /* this block is part of buddy of order 'order' */ ++ return order; ++ } ++ bb += 1 << (e3b->bd_blkbits - order); ++ order++; ++ } ++ return 0; ++} ++ ++static inline void mb_clear_bits(void *bm, int cur, int len) ++{ ++ __u32 *addr; ++ ++ len = cur + len; ++ while (cur < len) { ++ if ((cur & 31) == 0 && (len - cur) >= 32) { ++ /* fast path: clear whole word at once */ ++ addr = bm + (cur >> 3); ++ *addr = 0; ++ cur += 32; ++ continue; ++ } ++ mb_clear_bit_atomic(cur, bm); ++ cur++; ++ } ++} ++ ++static inline void mb_set_bits(void *bm, int cur, int len) ++{ ++ __u32 *addr; ++ ++ len = cur + len; ++ while (cur < len) { ++ if ((cur & 31) == 0 && (len - cur) >= 32) { ++ /* fast path: clear whole word at once */ ++ addr = bm + (cur >> 3); ++ *addr = 0xffffffff; ++ cur += 32; ++ continue; ++ } ++ mb_set_bit_atomic(cur, bm); ++ cur++; ++ } ++} ++ ++static int mb_free_blocks(struct ext3_buddy *e3b, int first, int count) ++{ ++ int block = 0, max = 0, order; ++ void *buddy, *buddy2; ++ ++ BUG_ON(first + count > (e3b->bd_sb->s_blocksize << 3)); ++ BUG_ON(!ext3_is_group_locked(e3b->bd_sb, e3b->bd_group)); ++ mb_check_buddy(e3b); ++ mb_free_blocks_double(e3b, first, count); ++ ++ e3b->bd_info->bb_free += count; ++ if (first < e3b->bd_info->bb_first_free) ++ e3b->bd_info->bb_first_free = first; ++ ++ /* let's maintain fragments counter */ ++ if (first != 0) ++ block = !mb_test_bit(first - 1, EXT3_MB_BITMAP(e3b)); ++ if (first + count < EXT3_SB(e3b->bd_sb)->s_mb_maxs[0]) ++ max = !mb_test_bit(first + count, EXT3_MB_BITMAP(e3b)); ++ if (block && max) ++ e3b->bd_info->bb_fragments--; ++ else if (!block && !max) ++ e3b->bd_info->bb_fragments++; ++ ++ /* let's maintain buddy itself */ ++ while (count-- > 0) { ++ block = first++; ++ order = 0; ++ ++ BUG_ON(!mb_test_bit(block, EXT3_MB_BITMAP(e3b))); ++ mb_clear_bit(block, EXT3_MB_BITMAP(e3b)); ++ e3b->bd_info->bb_counters[order]++; ++ ++ /* start of the buddy */ ++ buddy = mb_find_buddy(e3b, order, &max); ++ ++ do { ++ block &= ~1UL; ++ if (mb_test_bit(block, buddy) || ++ mb_test_bit(block + 1, buddy)) ++ break; ++ ++ /* both the buddies are free, try to coalesce them */ ++ buddy2 = mb_find_buddy(e3b, order + 1, &max); ++ ++ if (!buddy2) ++ break; ++ ++ if (order > 0) { ++ /* for special purposes, we don't set ++ * free bits in bitmap */ ++ mb_set_bit(block, buddy); ++ mb_set_bit(block + 1, buddy); ++ } ++ e3b->bd_info->bb_counters[order]--; ++ e3b->bd_info->bb_counters[order]--; ++ ++ block = block >> 1; ++ order++; ++ e3b->bd_info->bb_counters[order]++; ++ ++ mb_clear_bit(block, buddy2); ++ buddy = buddy2; ++ } while (1); ++ } ++ mb_check_buddy(e3b); ++ ++ return 0; ++} ++ ++static int mb_find_extent(struct ext3_buddy *e3b, int order, int block, ++ int needed, struct ext3_free_extent *ex) ++{ ++ int next = block, max, ord; ++ void *buddy; ++ ++ BUG_ON(!ext3_is_group_locked(e3b->bd_sb, e3b->bd_group)); ++ BUG_ON(ex == NULL); ++ ++ buddy = mb_find_buddy(e3b, order, &max); ++ BUG_ON(buddy == NULL); ++ BUG_ON(block >= max); ++ if (mb_test_bit(block, buddy)) { ++ ex->fe_len = 0; ++ ex->fe_start = 0; ++ ex->fe_group = 0; ++ return 0; ++ } ++ ++ if (likely(order == 0)) { ++ /* find actual order */ ++ order = mb_find_order_for_block(e3b, block); ++ block = block >> order; ++ } ++ ++ ex->fe_len = 1 << order; ++ ex->fe_start = block << order; ++ ex->fe_group = e3b->bd_group; ++ ++ /* calc difference from given start */ ++ next = next - ex->fe_start; ++ ex->fe_len -= next; ++ ex->fe_start += next; ++ ++ while (needed > ex->fe_len && (buddy = mb_find_buddy(e3b, order, &max))) { ++ ++ if (block + 1 >= max) ++ break; ++ ++ next = (block + 1) * (1 << order); ++ if (mb_test_bit(next, EXT3_MB_BITMAP(e3b))) ++ break; ++ ++ ord = mb_find_order_for_block(e3b, next); ++ ++ order = ord; ++ block = next >> order; ++ ex->fe_len += 1 << order; ++ } ++ ++ BUG_ON(ex->fe_start + ex->fe_len > (1 << (e3b->bd_blkbits + 3))); ++ return ex->fe_len; ++} ++ ++static int mb_mark_used(struct ext3_buddy *e3b, struct ext3_free_extent *ex) ++{ ++ int ord, mlen = 0, max = 0, cur; ++ int start = ex->fe_start; ++ int len = ex->fe_len; ++ unsigned ret = 0; ++ int len0 = len; ++ void *buddy; ++ ++ BUG_ON(start + len > (e3b->bd_sb->s_blocksize << 3)); ++ BUG_ON(e3b->bd_group != ex->fe_group); ++ BUG_ON(!ext3_is_group_locked(e3b->bd_sb, e3b->bd_group)); ++ mb_check_buddy(e3b); ++ mb_mark_used_double(e3b, start, len); ++ ++ e3b->bd_info->bb_free -= len; ++ if (e3b->bd_info->bb_first_free == start) ++ e3b->bd_info->bb_first_free += len; ++ ++ /* let's maintain fragments counter */ ++ if (start != 0) ++ mlen = !mb_test_bit(start - 1, EXT3_MB_BITMAP(e3b)); ++ if (start + len < EXT3_SB(e3b->bd_sb)->s_mb_maxs[0]) ++ max = !mb_test_bit(start + len, EXT3_MB_BITMAP(e3b)); ++ if (mlen && max) ++ e3b->bd_info->bb_fragments++; ++ else if (!mlen && !max) ++ e3b->bd_info->bb_fragments--; ++ ++ /* let's maintain buddy itself */ ++ while (len) { ++ ord = mb_find_order_for_block(e3b, start); ++ ++ if (((start >> ord) << ord) == start && len >= (1 << ord)) { ++ /* the whole chunk may be allocated at once! */ ++ mlen = 1 << ord; ++ buddy = mb_find_buddy(e3b, ord, &max); ++ BUG_ON((start >> ord) >= max); ++ mb_set_bit(start >> ord, buddy); ++ e3b->bd_info->bb_counters[ord]--; ++ start += mlen; ++ len -= mlen; ++ BUG_ON(len < 0); ++ continue; ++ } ++ ++ /* store for history */ ++ if (ret == 0) ++ ret = len | (ord << 16); ++ ++ /* we have to split large buddy */ ++ BUG_ON(ord <= 0); ++ buddy = mb_find_buddy(e3b, ord, &max); ++ mb_set_bit(start >> ord, buddy); ++ e3b->bd_info->bb_counters[ord]--; ++ ++ ord--; ++ cur = (start >> ord) & ~1U; ++ buddy = mb_find_buddy(e3b, ord, &max); ++ mb_clear_bit(cur, buddy); ++ mb_clear_bit(cur + 1, buddy); ++ e3b->bd_info->bb_counters[ord]++; ++ e3b->bd_info->bb_counters[ord]++; ++ } ++ ++ mb_set_bits(EXT3_MB_BITMAP(e3b), ex->fe_start, len0); ++ mb_check_buddy(e3b); ++ ++ return ret; ++} ++ ++/* ++ * Must be called under group lock! ++ */ ++static void ext3_mb_use_best_found(struct ext3_allocation_context *ac, ++ struct ext3_buddy *e3b) ++{ ++ unsigned long ret; ++ ++ BUG_ON(ac->ac_b_ex.fe_group != e3b->bd_group); ++ BUG_ON(ac->ac_status == AC_STATUS_FOUND); ++ ++ ac->ac_b_ex.fe_len = min(ac->ac_b_ex.fe_len, ac->ac_g_ex.fe_len); ++ ac->ac_b_ex.fe_logical = ac->ac_g_ex.fe_logical; ++ ret = mb_mark_used(e3b, &ac->ac_b_ex); ++ ++ /* preallocation can change ac_b_ex, thus we store actually ++ * allocated blocks for history */ ++ ac->ac_f_ex = ac->ac_b_ex; ++ ++ ac->ac_status = AC_STATUS_FOUND; ++ ac->ac_tail = ret & 0xffff; ++ ac->ac_buddy = ret >> 16; ++ ++ /* XXXXXXX: SUCH A HORRIBLE **CK */ ++ ac->ac_bitmap_page = e3b->bd_bitmap_page; ++ get_page(ac->ac_bitmap_page); ++ ac->ac_buddy_page = e3b->bd_buddy_page; ++ get_page(ac->ac_buddy_page); ++} ++ ++/* ++ * regular allocator, for general purposes allocation ++ */ ++ ++void ext3_mb_check_limits(struct ext3_allocation_context *ac, ++ struct ext3_buddy *e3b, ++ int finish_group) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(ac->ac_sb); ++ struct ext3_free_extent *bex = &ac->ac_b_ex; ++ struct ext3_free_extent *gex = &ac->ac_g_ex; ++ struct ext3_free_extent ex; ++ int max; ++ ++ /* ++ * We don't want to scan for a whole year ++ */ ++ if (ac->ac_found > sbi->s_mb_max_to_scan && ++ !(ac->ac_flags & EXT3_MB_HINT_FIRST)) { ++ ac->ac_status = AC_STATUS_BREAK; ++ return; ++ } ++ ++ /* ++ * Haven't found good chunk so far, let's continue ++ */ ++ if (bex->fe_len < gex->fe_len) ++ return; ++ ++ if ((finish_group || ac->ac_found > sbi->s_mb_min_to_scan) ++ && bex->fe_group == e3b->bd_group) { ++ /* recheck chunk's availability - we don't know ++ * when it was found (within this lock-unlock ++ * period or not) */ ++ max = mb_find_extent(e3b, 0, bex->fe_start, gex->fe_len, &ex); ++ if (max >= gex->fe_len) { ++ ext3_mb_use_best_found(ac, e3b); ++ return; ++ } ++ } ++} ++ ++/* ++ * The routine checks whether found extent is good enough. If it is, ++ * then the extent gets marked used and flag is set to the context ++ * to stop scanning. Otherwise, the extent is compared with the ++ * previous found extent and if new one is better, then it's stored ++ * in the context. Later, the best found extent will be used, if ++ * mballoc can't find good enough extent. ++ * ++ * FIXME: real allocation policy is to be designed yet! ++ */ ++static void ext3_mb_measure_extent(struct ext3_allocation_context *ac, ++ struct ext3_free_extent *ex, ++ struct ext3_buddy *e3b) ++{ ++ struct ext3_free_extent *bex = &ac->ac_b_ex; ++ struct ext3_free_extent *gex = &ac->ac_g_ex; ++ ++ BUG_ON(ex->fe_len <= 0); ++ BUG_ON(ex->fe_len >= (1 << ac->ac_sb->s_blocksize_bits) * 8); ++ BUG_ON(ex->fe_start >= (1 << ac->ac_sb->s_blocksize_bits) * 8); ++ BUG_ON(ac->ac_status != AC_STATUS_CONTINUE); ++ ++ ac->ac_found++; ++ ++ /* ++ * The special case - take what you catch first ++ */ ++ if (unlikely(ac->ac_flags & EXT3_MB_HINT_FIRST)) { ++ *bex = *ex; ++ ext3_mb_use_best_found(ac, e3b); ++ return; ++ } ++ ++ /* ++ * Let's check whether the chuck is good enough ++ */ ++ if (ex->fe_len == gex->fe_len) { ++ *bex = *ex; ++ ext3_mb_use_best_found(ac, e3b); ++ return; ++ } ++ ++ /* ++ * If this is first found extent, just store it in the context ++ */ ++ if (bex->fe_len == 0) { ++ *bex = *ex; ++ return; ++ } ++ ++ /* ++ * If new found extent is better, store it in the context ++ */ ++ if (bex->fe_len < gex->fe_len) { ++ /* if the request isn't satisfied, any found extent ++ * larger than previous best one is better */ ++ if (ex->fe_len > bex->fe_len) ++ *bex = *ex; ++ } else if (ex->fe_len > gex->fe_len) { ++ /* if the request is satisfied, then we try to find ++ * an extent that still satisfy the request, but is ++ * smaller than previous one */ ++ *bex = *ex; ++ } ++ ++ ext3_mb_check_limits(ac, e3b, 0); ++} ++ ++static int ext3_mb_try_best_found(struct ext3_allocation_context *ac, ++ struct ext3_buddy *e3b) ++{ ++ struct ext3_free_extent ex = ac->ac_b_ex; ++ int group = ex.fe_group, max, err; ++ ++ BUG_ON(ex.fe_len <= 0); ++ err = ext3_mb_load_buddy(ac->ac_sb, group, e3b); ++ if (err) ++ return err; ++ ++ ext3_lock_group(ac->ac_sb, group); ++ max = mb_find_extent(e3b, 0, ex.fe_start, ex.fe_len, &ex); ++ ++ if (max > 0) { ++ ac->ac_b_ex = ex; ++ ext3_mb_use_best_found(ac, e3b); ++ } ++ ++ ext3_unlock_group(ac->ac_sb, group); ++ ext3_mb_release_desc(e3b); ++ ++ return 0; ++} ++ ++static int ext3_mb_find_by_goal(struct ext3_allocation_context *ac, ++ struct ext3_buddy *e3b) ++{ ++ int group = ac->ac_g_ex.fe_group, max, err; ++ struct ext3_sb_info *sbi = EXT3_SB(ac->ac_sb); ++ struct ext3_super_block *es = sbi->s_es; ++ struct ext3_free_extent ex; ++ ++ err = ext3_mb_load_buddy(ac->ac_sb, group, e3b); ++ if (err) ++ return err; ++ ++ ext3_lock_group(ac->ac_sb, group); ++ max = mb_find_extent(e3b, 0, ac->ac_g_ex.fe_start, ++ ac->ac_g_ex.fe_len, &ex); ++ ++ if (max >= ac->ac_g_ex.fe_len && ac->ac_g_ex.fe_len == sbi->s_stripe) { ++ unsigned long start; ++ start = (e3b->bd_group * EXT3_BLOCKS_PER_GROUP(ac->ac_sb) + ++ ex.fe_start + le32_to_cpu(es->s_first_data_block)); ++ if (start % sbi->s_stripe == 0) { ++ ac->ac_found++; ++ ac->ac_b_ex = ex; ++ ext3_mb_use_best_found(ac, e3b); ++ } ++ } else if (max >= ac->ac_g_ex.fe_len) { ++ BUG_ON(ex.fe_len <= 0); ++ BUG_ON(ex.fe_group != ac->ac_g_ex.fe_group); ++ BUG_ON(ex.fe_start != ac->ac_g_ex.fe_start); ++ ac->ac_found++; ++ ac->ac_b_ex = ex; ++ ext3_mb_use_best_found(ac, e3b); ++ } else if (max > 0 && (ac->ac_flags & EXT3_MB_HINT_MERGE)) { ++ /* Sometimes, caller may want to merge even small ++ * number of blocks to an existing extent */ ++ BUG_ON(ex.fe_len <= 0); ++ BUG_ON(ex.fe_group != ac->ac_g_ex.fe_group); ++ BUG_ON(ex.fe_start != ac->ac_g_ex.fe_start); ++ ac->ac_found++; ++ ac->ac_b_ex = ex; ++ ext3_mb_use_best_found(ac, e3b); ++ } ++ ext3_unlock_group(ac->ac_sb, group); ++ ext3_mb_release_desc(e3b); ++ ++ return 0; ++} ++ ++/* ++ * The routine scans buddy structures (not bitmap!) from given order ++ * to max order and tries to find big enough chunk to satisfy the req ++ */ ++static void ext3_mb_simple_scan_group(struct ext3_allocation_context *ac, ++ struct ext3_buddy *e3b) ++{ ++ struct super_block *sb = ac->ac_sb; ++ struct ext3_group_info *grp = e3b->bd_info; ++ void *buddy; ++ int i, k, max; ++ ++ BUG_ON(ac->ac_2order <= 0); ++ for (i = ac->ac_2order; i <= sb->s_blocksize_bits + 1; i++) { ++ if (grp->bb_counters[i] == 0) ++ continue; ++ ++ buddy = mb_find_buddy(e3b, i, &max); ++ BUG_ON(buddy == NULL); ++ ++ k = mb_find_next_zero_bit(buddy, max, 0); ++ BUG_ON(k >= max); ++ ++ ac->ac_found++; ++ ++ ac->ac_b_ex.fe_len = 1 << i; ++ ac->ac_b_ex.fe_start = k << i; ++ ac->ac_b_ex.fe_group = e3b->bd_group; ++ ++ ext3_mb_use_best_found(ac, e3b); ++ ++ BUG_ON(ac->ac_b_ex.fe_len != ac->ac_g_ex.fe_len); ++ ++ if (EXT3_SB(sb)->s_mb_stats) ++ atomic_inc(&EXT3_SB(sb)->s_bal_2orders); ++ ++ break; ++ } ++} ++ ++/* ++ * The routine scans the group and measures all found extents. ++ * In order to optimize scanning, caller must pass number of ++ * free blocks in the group, so the routine can know upper limit. ++ */ ++static void ext3_mb_complex_scan_group(struct ext3_allocation_context *ac, ++ struct ext3_buddy *e3b) ++{ ++ struct super_block *sb = ac->ac_sb; ++ void *bitmap = EXT3_MB_BITMAP(e3b); ++ struct ext3_free_extent ex; ++ int i, free; ++ ++ free = e3b->bd_info->bb_free; ++ BUG_ON(free <= 0); ++ ++ i = e3b->bd_info->bb_first_free; ++ ++ while (free && ac->ac_status == AC_STATUS_CONTINUE) { ++ i = mb_find_next_zero_bit(bitmap, sb->s_blocksize * 8, i); ++ if (i >= sb->s_blocksize * 8) { ++ BUG_ON(free != 0); ++ break; ++ } ++ ++ mb_find_extent(e3b, 0, i, ac->ac_g_ex.fe_len, &ex); ++ BUG_ON(ex.fe_len <= 0); ++ BUG_ON(free < ex.fe_len); ++ ++ ext3_mb_measure_extent(ac, &ex, e3b); ++ ++ i += ex.fe_len; ++ free -= ex.fe_len; ++ } ++ ++ ext3_mb_check_limits(ac, e3b, 1); ++} ++ ++/* ++ * This is a special case for storages like raid5 ++ * we try to find stripe-aligned chunks for stripe-size requests ++ */ ++static void ext3_mb_scan_aligned(struct ext3_allocation_context *ac, ++ struct ext3_buddy *e3b) ++{ ++ struct super_block *sb = ac->ac_sb; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ void *bitmap = EXT3_MB_BITMAP(e3b); ++ struct ext3_free_extent ex; ++ unsigned long i, max; ++ ++ BUG_ON(sbi->s_stripe == 0); ++ ++ /* find first stripe-aligned block */ ++ i = e3b->bd_group * EXT3_BLOCKS_PER_GROUP(sb) ++ + le32_to_cpu(sbi->s_es->s_first_data_block); ++ i = ((i + sbi->s_stripe - 1) / sbi->s_stripe) * sbi->s_stripe; ++ i = (i - le32_to_cpu(sbi->s_es->s_first_data_block)) ++ % EXT3_BLOCKS_PER_GROUP(sb); ++ ++ while (i < sb->s_blocksize * 8) { ++ if (!mb_test_bit(i, bitmap)) { ++ max = mb_find_extent(e3b, 0, i, sbi->s_stripe, &ex); ++ if (max >= sbi->s_stripe) { ++ ac->ac_found++; ++ ac->ac_b_ex = ex; ++ ext3_mb_use_best_found(ac, e3b); ++ break; ++ } ++ } ++ i += sbi->s_stripe; ++ } ++} ++ ++static int ext3_mb_good_group(struct ext3_allocation_context *ac, ++ int group, int cr) ++{ ++ struct ext3_group_info *grp = EXT3_GROUP_INFO(ac->ac_sb, group); ++ unsigned free, fragments, i, bits; ++ ++ BUG_ON(cr < 0 || cr >= 4); ++ BUG_ON(EXT3_MB_GRP_NEED_INIT(grp)); ++ ++ free = grp->bb_free; ++ fragments = grp->bb_fragments; ++ if (free == 0) ++ return 0; ++ if (fragments == 0) ++ return 0; ++ ++ switch (cr) { ++ case 0: ++ BUG_ON(ac->ac_2order == 0); ++ bits = ac->ac_sb->s_blocksize_bits + 1; ++ for (i = ac->ac_2order; i <= bits; i++) ++ if (grp->bb_counters[i] > 0) ++ return 1; ++ break; ++ case 1: ++ if ((free / fragments) >= ac->ac_g_ex.fe_len) ++ return 1; ++ break; ++ case 2: ++ if (free >= ac->ac_g_ex.fe_len) ++ return 1; ++ break; ++ case 3: ++ return 1; ++ default: ++ BUG(); ++ } ++ ++ return 0; ++} ++ ++int ext3_mb_regular_allocator(struct ext3_allocation_context *ac) ++{ ++ int group, i, cr, err = 0; ++ struct ext3_sb_info *sbi; ++ struct super_block *sb; ++ struct ext3_buddy e3b; ++ ++ sb = ac->ac_sb; ++ sbi = EXT3_SB(sb); ++ BUG_ON(ac->ac_status == AC_STATUS_FOUND); ++ ++ /* first, try the goal */ ++ err = ext3_mb_find_by_goal(ac, &e3b); ++ if (err || ac->ac_status == AC_STATUS_FOUND) ++ goto out; ++ ++ if (unlikely(ac->ac_flags & EXT3_MB_HINT_GOAL_ONLY)) ++ goto out; ++ ++ i = ffs(ac->ac_g_ex.fe_len); ++ ac->ac_2order = 0; ++ if (i >= sbi->s_mb_order2_reqs) { ++ i--; ++ if ((ac->ac_g_ex.fe_len & (~(1 << i))) == 0) ++ ac->ac_2order = i; ++ } ++ ++ group = ac->ac_g_ex.fe_group; ++ ++ /* Let's just scan groups to find more-less suitable blocks */ ++ cr = ac->ac_2order ? 0 : 1; ++repeat: ++ for (; cr < 4 && ac->ac_status == AC_STATUS_CONTINUE; cr++) { ++ ac->ac_criteria = cr; ++ for (i = 0; i < EXT3_SB(sb)->s_groups_count; group++, i++) { ++ struct ext3_group_info *grp; ++ ++ if (group == EXT3_SB(sb)->s_groups_count) ++ group = 0; ++ ++ /* quick check to skip empty groups */ ++ grp = EXT3_GROUP_INFO(ac->ac_sb, group); ++ if (grp->bb_free == 0) ++ continue; ++ ++ if (EXT3_MB_GRP_NEED_INIT(EXT3_GROUP_INFO(sb, group))) { ++ /* we need full data about the group ++ * to make a good selection */ ++ err = ext3_mb_load_buddy(sb, group, &e3b); ++ if (err) ++ goto out; ++ ext3_mb_release_desc(&e3b); ++ } ++ ++ /* check is group good for our criteries */ ++ if (!ext3_mb_good_group(ac, group, cr)) ++ continue; ++ ++ err = ext3_mb_load_buddy(sb, group, &e3b); ++ if (err) ++ goto out; ++ ++ ext3_lock_group(sb, group); ++ if (!ext3_mb_good_group(ac, group, cr)) { ++ /* someone did allocation from this group */ ++ ext3_unlock_group(sb, group); ++ ext3_mb_release_desc(&e3b); ++ continue; ++ } ++ ++ ac->ac_groups_scanned++; ++ if (cr == 0) ++ ext3_mb_simple_scan_group(ac, &e3b); ++ else if (cr == 1 && ac->ac_g_ex.fe_len == sbi->s_stripe) ++ ext3_mb_scan_aligned(ac, &e3b); ++ else ++ ext3_mb_complex_scan_group(ac, &e3b); ++ ++ ext3_unlock_group(sb, group); ++ ext3_mb_release_desc(&e3b); ++ ++ if (ac->ac_status != AC_STATUS_CONTINUE) ++ break; ++ } ++ } ++ ++ if (ac->ac_b_ex.fe_len > 0 && ac->ac_status != AC_STATUS_FOUND && ++ !(ac->ac_flags & EXT3_MB_HINT_FIRST)) { ++ /* ++ * We've been searching too long. Let's try to allocate ++ * the best chunk we've found so far ++ */ ++ ++ ext3_mb_try_best_found(ac, &e3b); ++ if (ac->ac_status != AC_STATUS_FOUND) { ++ /* ++ * Someone more lucky has already allocated it. ++ * The only thing we can do is just take first ++ * found block(s) ++ printk(KERN_DEBUG "EXT3-fs: someone won our chunk\n"); ++ */ ++ ac->ac_b_ex.fe_group = 0; ++ ac->ac_b_ex.fe_start = 0; ++ ac->ac_b_ex.fe_len = 0; ++ ac->ac_status = AC_STATUS_CONTINUE; ++ ac->ac_flags |= EXT3_MB_HINT_FIRST; ++ cr = 3; ++ atomic_inc(&sbi->s_mb_lost_chunks); ++ goto repeat; ++ } ++ } ++out: ++ return err; ++} ++ ++#ifdef EXT3_MB_HISTORY ++struct ext3_mb_proc_session { ++ struct ext3_mb_history *history; ++ struct super_block *sb; ++ int start; ++ int max; ++}; ++ ++static void *ext3_mb_history_skip_empty(struct ext3_mb_proc_session *s, ++ struct ext3_mb_history *hs, ++ int first) ++{ ++ if (hs == s->history + s->max) ++ hs = s->history; ++ if (!first && hs == s->history + s->start) ++ return NULL; ++ while (hs->orig.fe_len == 0) { ++ hs++; ++ if (hs == s->history + s->max) ++ hs = s->history; ++ if (hs == s->history + s->start) ++ return NULL; ++ } ++ return hs; ++} ++ ++static void *ext3_mb_seq_history_start(struct seq_file *seq, loff_t *pos) ++{ ++ struct ext3_mb_proc_session *s = seq->private; ++ struct ext3_mb_history *hs; ++ int l = *pos; ++ ++ if (l == 0) ++ return SEQ_START_TOKEN; ++ hs = ext3_mb_history_skip_empty(s, s->history + s->start, 1); ++ if (!hs) ++ return NULL; ++ while (--l && (hs = ext3_mb_history_skip_empty(s, ++hs, 0)) != NULL); ++ return hs; ++} ++ ++static void *ext3_mb_seq_history_next(struct seq_file *seq, void *v, loff_t *pos) ++{ ++ struct ext3_mb_proc_session *s = seq->private; ++ struct ext3_mb_history *hs = v; ++ ++ ++*pos; ++ if (v == SEQ_START_TOKEN) ++ return ext3_mb_history_skip_empty(s, s->history + s->start, 1); ++ else ++ return ext3_mb_history_skip_empty(s, ++hs, 0); ++} ++ ++static int ext3_mb_seq_history_show(struct seq_file *seq, void *v) ++{ ++ char buf[25], buf2[25], buf3[25], *fmt; ++ struct ext3_mb_history *hs = v; ++ ++ if (v == SEQ_START_TOKEN) { ++ seq_printf(seq, "%-5s %-8s %-23s %-23s %-23s %-5s " ++ "%-5s %-2s %-5s %-5s %-5s %-6s\n", ++ "pid", "inode", "original", "goal", "result","found", ++ "grps", "cr", "flags", "merge", "tail", "broken"); ++ return 0; ++ } ++ ++ if (hs->op == EXT3_MB_HISTORY_ALLOC) { ++ fmt = "%-5u %-8u %-23s %-23s %-23s %-5u %-5u %-2u " ++ "%-5u %-5s %-5u %-6u\n"; ++ sprintf(buf2, "%lu/%lu/%lu@%lu", hs->result.fe_group, ++ hs->result.fe_start, hs->result.fe_len, ++ hs->result.fe_logical); ++ sprintf(buf, "%lu/%lu/%lu@%lu", hs->orig.fe_group, ++ hs->orig.fe_start, hs->orig.fe_len, ++ hs->orig.fe_logical); ++ sprintf(buf3, "%lu/%lu/%lu@%lu", hs->goal.fe_group, ++ hs->goal.fe_start, hs->goal.fe_len, ++ hs->goal.fe_logical); ++ seq_printf(seq, fmt, hs->pid, hs->ino, buf, buf3, buf2, ++ hs->found, hs->groups, hs->cr, hs->flags, ++ hs->merged ? "M" : "", hs->tail, ++ hs->buddy ? 1 << hs->buddy : 0); ++ } else if (hs->op == EXT3_MB_HISTORY_PREALLOC) { ++ fmt = "%-5u %-8u %-23s %-23s %-23s\n"; ++ sprintf(buf2, "%lu/%lu/%lu@%lu", hs->result.fe_group, ++ hs->result.fe_start, hs->result.fe_len, ++ hs->result.fe_logical); ++ sprintf(buf, "%lu/%lu/%lu@%lu", hs->orig.fe_group, ++ hs->orig.fe_start, hs->orig.fe_len, ++ hs->orig.fe_logical); ++ seq_printf(seq, fmt, hs->pid, hs->ino, buf, "", buf2); ++ } else if (hs->op == EXT3_MB_HISTORY_DISCARD) { ++ sprintf(buf2, "%lu/%lu/%lu", hs->result.fe_group, ++ hs->result.fe_start, hs->result.fe_len); ++ seq_printf(seq, "%-5u %-8u %-23s discard\n", ++ hs->pid, hs->ino, buf2); ++ } else if (hs->op == EXT3_MB_HISTORY_FREE) { ++ sprintf(buf2, "%lu/%lu/%lu", hs->result.fe_group, ++ hs->result.fe_start, hs->result.fe_len); ++ seq_printf(seq, "%-5u %-8u %-23s free\n", ++ hs->pid, hs->ino, buf2); ++ } ++ return 0; ++} ++ ++static void ext3_mb_seq_history_stop(struct seq_file *seq, void *v) ++{ ++} ++ ++static struct seq_operations ext3_mb_seq_history_ops = { ++ .start = ext3_mb_seq_history_start, ++ .next = ext3_mb_seq_history_next, ++ .stop = ext3_mb_seq_history_stop, ++ .show = ext3_mb_seq_history_show, ++}; ++ ++static int ext3_mb_seq_history_open(struct inode *inode, struct file *file) ++{ ++ struct super_block *sb = PDE(inode)->data; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ struct ext3_mb_proc_session *s; ++ int rc, size; ++ ++ s = kmalloc(sizeof(*s), GFP_KERNEL); ++ if (s == NULL) ++ return -ENOMEM; ++ s->sb = sb; ++ size = sizeof(struct ext3_mb_history) * sbi->s_mb_history_max; ++ s->history = kmalloc(size, GFP_KERNEL); ++ if (s->history == NULL) { ++ kfree(s); ++ return -ENOMEM; ++ } ++ ++ spin_lock(&sbi->s_mb_history_lock); ++ memcpy(s->history, sbi->s_mb_history, size); ++ s->max = sbi->s_mb_history_max; ++ s->start = sbi->s_mb_history_cur % s->max; ++ spin_unlock(&sbi->s_mb_history_lock); ++ ++ rc = seq_open(file, &ext3_mb_seq_history_ops); ++ if (rc == 0) { ++ struct seq_file *m = (struct seq_file *)file->private_data; ++ m->private = s; ++ } else { ++ kfree(s->history); ++ kfree(s); ++ } ++ return rc; ++ ++} ++ ++static int ext3_mb_seq_history_release(struct inode *inode, struct file *file) ++{ ++ struct seq_file *seq = (struct seq_file *)file->private_data; ++ struct ext3_mb_proc_session *s = seq->private; ++ kfree(s->history); ++ kfree(s); ++ return seq_release(inode, file); ++} ++ ++static ssize_t ext3_mb_seq_history_write(struct file *file, ++ const char __user *buffer, ++ size_t count, loff_t *ppos) ++{ ++ struct seq_file *seq = (struct seq_file *)file->private_data; ++ struct ext3_mb_proc_session *s = seq->private; ++ struct super_block *sb = s->sb; ++ char str[32]; ++ int value; ++ ++ if (count >= sizeof(str)) { ++ printk(KERN_ERR "EXT3-fs: %s string too long, max %u bytes\n", ++ "mb_history", (int)sizeof(str)); ++ return -EOVERFLOW; ++ } ++ ++ if (copy_from_user(str, buffer, count)) ++ return -EFAULT; ++ ++ value = simple_strtol(str, NULL, 0); ++ if (value < 0) ++ return -ERANGE; ++ EXT3_SB(sb)->s_mb_history_filter = value; ++ ++ return count; ++} ++ ++static struct file_operations ext3_mb_seq_history_fops = { ++ .owner = THIS_MODULE, ++ .open = ext3_mb_seq_history_open, ++ .read = seq_read, ++ .write = ext3_mb_seq_history_write, ++ .llseek = seq_lseek, ++ .release = ext3_mb_seq_history_release, ++}; ++ ++static void *ext3_mb_seq_groups_start(struct seq_file *seq, loff_t *pos) ++{ ++ struct super_block *sb = seq->private; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ long group; ++ ++ if (*pos < 0 || *pos >= sbi->s_groups_count) ++ return NULL; ++ ++ group = *pos + 1; ++ return (void *) group; ++} ++ ++static void *ext3_mb_seq_groups_next(struct seq_file *seq, void *v, loff_t *pos) ++{ ++ struct super_block *sb = seq->private; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ long group; ++ ++ ++*pos; ++ if (*pos < 0 || *pos >= sbi->s_groups_count) ++ return NULL; ++ group = *pos + 1; ++ return (void *) group;; ++} ++ ++static int ext3_mb_seq_groups_show(struct seq_file *seq, void *v) ++{ ++ struct super_block *sb = seq->private; ++ long group = (long) v; ++ int i, err; ++ struct ext3_buddy e3b; ++ struct sg { ++ struct ext3_group_info info; ++ unsigned short counters[16]; ++ } sg; ++ ++ group--; ++ if (group == 0) ++ seq_printf(seq, "#%-5s: %-5s %-5s %-5s " ++ "[ %-5s %-5s %-5s %-5s %-5s %-5s %-5s " ++ "%-5s %-5s %-5s %-5s %-5s %-5s %-5s ]\n", ++ "group", "free", "frags", "first", ++ "2^0", "2^1", "2^2", "2^3", "2^4", "2^5","2^6", ++ "2^7", "2^8", "2^9", "2^10", "2^11", "2^12", "2^13"); ++ ++ i = (sb->s_blocksize_bits + 2) * sizeof(sg.info.bb_counters[0]) + ++ sizeof(struct ext3_group_info); ++ err = ext3_mb_load_buddy(sb, group, &e3b); ++ if (err) { ++ seq_printf(seq, "#%-5lu: I/O error\n", group); ++ return 0; ++ } ++ ext3_lock_group(sb, group); ++ memcpy(&sg, EXT3_GROUP_INFO(sb, group), i); ++ ext3_unlock_group(sb, group); ++ ext3_mb_release_desc(&e3b); ++ ++ seq_printf(seq, "#%-5lu: %-5u %-5u %-5u [", group, sg.info.bb_free, ++ sg.info.bb_fragments, sg.info.bb_first_free); ++ for (i = 0; i <= 13; i++) ++ seq_printf(seq, " %-5u", i <= sb->s_blocksize_bits + 1 ? ++ sg.info.bb_counters[i] : 0); ++ seq_printf(seq, " ]\n"); ++ ++ return 0; ++} ++ ++static void ext3_mb_seq_groups_stop(struct seq_file *seq, void *v) ++{ ++} ++ ++static struct seq_operations ext3_mb_seq_groups_ops = { ++ .start = ext3_mb_seq_groups_start, ++ .next = ext3_mb_seq_groups_next, ++ .stop = ext3_mb_seq_groups_stop, ++ .show = ext3_mb_seq_groups_show, ++}; ++ ++static int ext3_mb_seq_groups_open(struct inode *inode, struct file *file) ++{ ++ struct super_block *sb = PDE(inode)->data; ++ int rc; ++ ++ rc = seq_open(file, &ext3_mb_seq_groups_ops); ++ if (rc == 0) { ++ struct seq_file *m = (struct seq_file *)file->private_data; ++ m->private = sb; ++ } ++ return rc; ++ ++} ++ ++static struct file_operations ext3_mb_seq_groups_fops = { ++ .owner = THIS_MODULE, ++ .open = ext3_mb_seq_groups_open, ++ .read = seq_read, ++ .llseek = seq_lseek, ++ .release = seq_release, ++}; ++ ++static void ext3_mb_history_release(struct super_block *sb) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ ++ remove_proc_entry("mb_groups", sbi->s_mb_proc); ++ remove_proc_entry("mb_history", sbi->s_mb_proc); ++ ++ if (sbi->s_mb_history) ++ kfree(sbi->s_mb_history); ++} ++ ++static void ext3_mb_history_init(struct super_block *sb) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ int i; ++ ++ if (sbi->s_mb_proc != NULL) { ++ struct proc_dir_entry *p; ++ p = create_proc_entry("mb_history", S_IRUGO, sbi->s_mb_proc); ++ if (p) { ++ p->proc_fops = &ext3_mb_seq_history_fops; ++ p->data = sb; ++ } ++ p = create_proc_entry("mb_groups", S_IRUGO, sbi->s_mb_proc); ++ if (p) { ++ p->proc_fops = &ext3_mb_seq_groups_fops; ++ p->data = sb; ++ } ++ } ++ ++ sbi->s_mb_history_max = 1000; ++ sbi->s_mb_history_cur = 0; ++ spin_lock_init(&sbi->s_mb_history_lock); ++ i = sbi->s_mb_history_max * sizeof(struct ext3_mb_history); ++ sbi->s_mb_history = kmalloc(i, GFP_KERNEL); ++ memset(sbi->s_mb_history, 0, i); ++ /* if we can't allocate history, then we simple won't use it */ ++} ++ ++static void ++ext3_mb_store_history(struct ext3_allocation_context *ac) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(ac->ac_sb); ++ struct ext3_mb_history h; ++ ++ if (likely(sbi->s_mb_history == NULL)) ++ return; ++ ++ if (!(ac->ac_op & sbi->s_mb_history_filter)) ++ return; ++ ++ h.op = ac->ac_op; ++ h.pid = current->pid; ++ h.ino = ac->ac_inode ? ac->ac_inode->i_ino : 0; ++ h.orig = ac->ac_o_ex; ++ h.result = ac->ac_b_ex; ++ h.flags = ac->ac_flags; ++ h.merged = 0; ++ if (ac->ac_op == EXT3_MB_HISTORY_ALLOC) { ++ if (ac->ac_g_ex.fe_start == ac->ac_b_ex.fe_start && ++ ac->ac_g_ex.fe_group == ac->ac_b_ex.fe_group) ++ h.merged = 1; ++ h.goal = ac->ac_g_ex; ++ h.result = ac->ac_f_ex; ++ } ++ ++ spin_lock(&sbi->s_mb_history_lock); ++ memcpy(sbi->s_mb_history + sbi->s_mb_history_cur, &h, sizeof(h)); ++ if (++sbi->s_mb_history_cur >= sbi->s_mb_history_max) ++ sbi->s_mb_history_cur = 0; ++ spin_unlock(&sbi->s_mb_history_lock); ++} ++ ++#else ++#define ext3_mb_history_release(sb) ++#define ext3_mb_history_init(sb) ++#endif ++ ++int ext3_mb_init_backend(struct super_block *sb) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ int i, j, len, metalen; ++ int num_meta_group_infos = ++ (sbi->s_groups_count + EXT3_DESC_PER_BLOCK(sb) - 1) >> ++ EXT3_DESC_PER_BLOCK_BITS(sb); ++ struct ext3_group_info **meta_group_info; ++ ++ /* An 8TB filesystem with 64-bit pointers requires a 4096 byte ++ * kmalloc. A 128kb malloc should suffice for a 256TB filesystem. ++ * So a two level scheme suffices for now. */ ++ sbi->s_group_info = kmalloc(sizeof(*sbi->s_group_info) * ++ num_meta_group_infos, GFP_KERNEL); ++ if (sbi->s_group_info == NULL) { ++ printk(KERN_ERR "EXT3-fs: can't allocate buddy meta group\n"); ++ return -ENOMEM; ++ } ++ sbi->s_buddy_cache = new_inode(sb); ++ if (sbi->s_buddy_cache == NULL) { ++ printk(KERN_ERR "EXT3-fs: can't get new inode\n"); ++ goto err_freesgi; ++ } ++ EXT3_I(sbi->s_buddy_cache)->i_disksize = 0; ++ ++ metalen = sizeof(*meta_group_info) << EXT3_DESC_PER_BLOCK_BITS(sb); ++ for (i = 0; i < num_meta_group_infos; i++) { ++ if ((i + 1) == num_meta_group_infos) ++ metalen = sizeof(*meta_group_info) * ++ (sbi->s_groups_count - ++ (i << EXT3_DESC_PER_BLOCK_BITS(sb))); ++ meta_group_info = kmalloc(metalen, GFP_KERNEL); ++ if (meta_group_info == NULL) { ++ printk(KERN_ERR "EXT3-fs: can't allocate mem for a " ++ "buddy group\n"); ++ goto err_freemeta; ++ } ++ sbi->s_group_info[i] = meta_group_info; ++ } ++ ++ /* ++ * calculate needed size. if change bb_counters size, ++ * don't forget about ext3_mb_generate_buddy() ++ */ ++ len = sizeof(struct ext3_group_info); ++ len += sizeof(unsigned short) * (sb->s_blocksize_bits + 2); ++ for (i = 0; i < sbi->s_groups_count; i++) { ++ struct ext3_group_desc * desc; ++ ++ meta_group_info = ++ sbi->s_group_info[i >> EXT3_DESC_PER_BLOCK_BITS(sb)]; ++ j = i & (EXT3_DESC_PER_BLOCK(sb) - 1); ++ ++ meta_group_info[j] = kmalloc(len, GFP_KERNEL); ++ if (meta_group_info[j] == NULL) { ++ printk(KERN_ERR "EXT3-fs: can't allocate buddy mem\n"); ++ i--; ++ goto err_freebuddy; ++ } ++ desc = ext3_get_group_desc(sb, i, NULL); ++ if (desc == NULL) { ++ printk(KERN_ERR"EXT3-fs: can't read descriptor %u\n",i); ++ goto err_freebuddy; ++ } ++ memset(meta_group_info[j], 0, len); ++ set_bit(EXT3_GROUP_INFO_NEED_INIT_BIT, ++ &meta_group_info[j]->bb_state); ++ ++ /* initialize bb_free to be able to skip ++ * empty groups without initialization */ ++ meta_group_info[j]->bb_free = ++ le16_to_cpu(desc->bg_free_blocks_count); ++ ++ INIT_LIST_HEAD(&meta_group_info[j]->bb_prealloc_list); ++ ++#ifdef DOUBLE_CHECK ++ { ++ struct buffer_head *bh; ++ meta_group_info[j]->bb_bitmap = ++ kmalloc(sb->s_blocksize, GFP_KERNEL); ++ BUG_ON(meta_group_info[j]->bb_bitmap == NULL); ++ bh = read_block_bitmap(sb, i); ++ BUG_ON(bh == NULL); ++ memcpy(meta_group_info[j]->bb_bitmap, bh->b_data, ++ sb->s_blocksize); ++ brelse(bh); ++ } ++#endif ++ ++ } ++ ++ return 0; ++ ++err_freebuddy: ++ while (i >= 0) { ++ kfree(EXT3_GROUP_INFO(sb, i)); ++ i--; ++ } ++ i = num_meta_group_infos; ++err_freemeta: ++ while (--i >= 0) ++ kfree(sbi->s_group_info[i]); ++ iput(sbi->s_buddy_cache); ++err_freesgi: ++ kfree(sbi->s_group_info); ++ return -ENOMEM; ++} ++ ++int ext3_mb_init(struct super_block *sb, int needs_recovery) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ unsigned i, offset, max; ++ ++ if (!test_opt(sb, MBALLOC)) ++ return 0; ++ ++ i = (sb->s_blocksize_bits + 2) * sizeof(unsigned short); ++ ++ sbi->s_mb_offsets = kmalloc(i, GFP_KERNEL); ++ if (sbi->s_mb_offsets == NULL) { ++ clear_opt(sbi->s_mount_opt, MBALLOC); ++ return -ENOMEM; ++ } ++ sbi->s_mb_maxs = kmalloc(i, GFP_KERNEL); ++ if (sbi->s_mb_maxs == NULL) { ++ clear_opt(sbi->s_mount_opt, MBALLOC); ++ kfree(sbi->s_mb_maxs); ++ return -ENOMEM; ++ } ++ ++ /* order 0 is regular bitmap */ ++ sbi->s_mb_maxs[0] = sb->s_blocksize << 3; ++ sbi->s_mb_offsets[0] = 0; ++ ++ i = 1; ++ offset = 0; ++ max = sb->s_blocksize << 2; ++ do { ++ sbi->s_mb_offsets[i] = offset; ++ sbi->s_mb_maxs[i] = max; ++ offset += 1 << (sb->s_blocksize_bits - i); ++ max = max >> 1; ++ i++; ++ } while (i <= sb->s_blocksize_bits + 1); ++ ++ /* init file for buddy data */ ++ if ((i = ext3_mb_init_backend(sb))) { ++ clear_opt(sbi->s_mount_opt, MBALLOC); ++ kfree(sbi->s_mb_offsets); ++ kfree(sbi->s_mb_maxs); ++ return i; ++ } ++ ++ spin_lock_init(&sbi->s_md_lock); ++ INIT_LIST_HEAD(&sbi->s_active_transaction); ++ INIT_LIST_HEAD(&sbi->s_closed_transaction); ++ INIT_LIST_HEAD(&sbi->s_committed_transaction); ++ spin_lock_init(&sbi->s_bal_lock); ++ ++ sbi->s_mb_max_to_scan = MB_DEFAULT_MAX_TO_SCAN; ++ sbi->s_mb_min_to_scan = MB_DEFAULT_MIN_TO_SCAN; ++ sbi->s_mb_max_groups_to_scan = MB_DEFAULT_MAX_GROUPS_TO_SCAN; ++ sbi->s_mb_stats = MB_DEFAULT_STATS; ++ sbi->s_mb_stream_request = MB_DEFAULT_STREAM_THRESHOLD; ++ sbi->s_mb_order2_reqs = MB_DEFAULT_ORDER2_REQS; ++ sbi->s_mb_history_filter = EXT3_MB_HISTORY_DEFAULT; ++ ++ i = sizeof(struct ext3_locality_group) * NR_CPUS; ++ sbi->s_locality_groups = kmalloc(i, GFP_NOFS); ++ if (sbi->s_locality_groups == NULL) { ++ clear_opt(sbi->s_mount_opt, MBALLOC); ++ kfree(sbi->s_mb_offsets); ++ kfree(sbi->s_mb_maxs); ++ return -ENOMEM; ++ } ++ for (i = 0; i < NR_CPUS; i++) { ++ struct ext3_locality_group *lg; ++ lg = &sbi->s_locality_groups[i]; ++ sema_init(&lg->lg_sem, 1); ++ INIT_LIST_HEAD(&lg->lg_prealloc_list); ++ spin_lock_init(&lg->lg_prealloc_lock); ++ } ++ ++ ext3_mb_init_per_dev_proc(sb); ++ ext3_mb_history_init(sb); ++ ++ printk("EXT3-fs: mballoc enabled\n"); ++ return 0; ++} ++ ++void ext3_mb_cleanup_pa(struct ext3_group_info *grp) ++{ ++ struct ext3_prealloc_space *pa; ++ struct list_head *cur, *tmp; ++ int count = 0; ++ ++ list_for_each_safe(cur, tmp, &grp->bb_prealloc_list) { ++ pa = list_entry(cur, struct ext3_prealloc_space, pa_group_list); ++ list_del_rcu(&pa->pa_group_list); ++ count++; ++ kfree(pa); ++ } ++ if (count) ++ mb_debug("mballoc: %u PAs left\n", count); ++ ++} ++ ++int ext3_mb_release(struct super_block *sb) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ int i, num_meta_group_infos; ++ ++ if (!test_opt(sb, MBALLOC)) ++ return 0; ++ ++ /* release freed, non-committed blocks */ ++ spin_lock(&sbi->s_md_lock); ++ list_splice_init(&sbi->s_closed_transaction, ++ &sbi->s_committed_transaction); ++ list_splice_init(&sbi->s_active_transaction, ++ &sbi->s_committed_transaction); ++ spin_unlock(&sbi->s_md_lock); ++ ext3_mb_free_committed_blocks(sb); ++ ++ if (sbi->s_group_info) { ++ for (i = 0; i < sbi->s_groups_count; i++) { ++#ifdef DOUBLE_CHECK ++ if (EXT3_GROUP_INFO(sb, i)->bb_bitmap) ++ kfree(EXT3_GROUP_INFO(sb, i)->bb_bitmap); ++#endif ++ ext3_mb_cleanup_pa(EXT3_GROUP_INFO(sb, i)); ++ kfree(EXT3_GROUP_INFO(sb, i)); ++ } ++ num_meta_group_infos = (sbi->s_groups_count + ++ EXT3_DESC_PER_BLOCK(sb) - 1) >> ++ EXT3_DESC_PER_BLOCK_BITS(sb); ++ for (i = 0; i < num_meta_group_infos; i++) ++ kfree(sbi->s_group_info[i]); ++ kfree(sbi->s_group_info); ++ } ++ if (sbi->s_mb_offsets) ++ kfree(sbi->s_mb_offsets); ++ if (sbi->s_mb_maxs) ++ kfree(sbi->s_mb_maxs); ++ if (sbi->s_buddy_cache) ++ iput(sbi->s_buddy_cache); ++ if (sbi->s_mb_stats) { ++ printk("EXT3-fs: mballoc: %u blocks %u reqs (%u success)\n", ++ atomic_read(&sbi->s_bal_allocated), ++ atomic_read(&sbi->s_bal_reqs), ++ atomic_read(&sbi->s_bal_success)); ++ printk("EXT3-fs: mballoc: %u extents scanned, %u goal hits, " ++ "%u 2^N hits, %u breaks, %u lost\n", ++ atomic_read(&sbi->s_bal_ex_scanned), ++ atomic_read(&sbi->s_bal_goals), ++ atomic_read(&sbi->s_bal_2orders), ++ atomic_read(&sbi->s_bal_breaks), ++ atomic_read(&sbi->s_mb_lost_chunks)); ++ printk("EXT3-fs: mballoc: %lu generated and it took %Lu\n", ++ sbi->s_mb_buddies_generated++, ++ sbi->s_mb_generation_time); ++ printk("EXT3-fs: mballoc: %u preallocated, %u discarded\n", ++ atomic_read(&sbi->s_mb_preallocated), ++ atomic_read(&sbi->s_mb_discarded)); ++ } ++ ++ if (sbi->s_locality_groups) ++ kfree(sbi->s_locality_groups); ++ ++ ext3_mb_history_release(sb); ++ ext3_mb_destroy_per_dev_proc(sb); ++ ++ return 0; ++} ++ ++void ext3_mb_free_committed_blocks(struct super_block *sb) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ int err, i, count = 0, count2 = 0; ++ struct ext3_free_metadata *md; ++ struct ext3_buddy e3b; ++ ++ if (list_empty(&sbi->s_committed_transaction)) ++ return; ++ ++ /* there is committed blocks to be freed yet */ ++ do { ++ /* get next array of blocks */ ++ md = NULL; ++ spin_lock(&sbi->s_md_lock); ++ if (!list_empty(&sbi->s_committed_transaction)) { ++ md = list_entry(sbi->s_committed_transaction.next, ++ struct ext3_free_metadata, list); ++ list_del(&md->list); ++ } ++ spin_unlock(&sbi->s_md_lock); ++ ++ if (md == NULL) ++ break; ++ ++ mb_debug("gonna free %u blocks in group %u (0x%p):", ++ md->num, md->group, md); ++ ++ err = ext3_mb_load_buddy(sb, md->group, &e3b); ++ /* we expect to find existing buddy because it's pinned */ ++ BUG_ON(err != 0); ++ ++ /* there are blocks to put in buddy to make them really free */ ++ count += md->num; ++ count2++; ++ ext3_lock_group(sb, md->group); ++ for (i = 0; i < md->num; i++) { ++ mb_debug(" %u", md->blocks[i]); ++ err = mb_free_blocks(&e3b, md->blocks[i], 1); ++ BUG_ON(err != 0); ++ } ++ mb_debug("\n"); ++ ext3_unlock_group(sb, md->group); ++ ++ /* balance refcounts from ext3_mb_free_metadata() */ ++ page_cache_release(e3b.bd_buddy_page); ++ page_cache_release(e3b.bd_bitmap_page); ++ ++ kfree(md); ++ ext3_mb_release_desc(&e3b); ++ ++ } while (md); ++ ++ mb_debug("freed %u blocks in %u structures\n", count, count2); ++} ++ ++#define EXT3_ROOT "ext3" ++#define EXT3_MB_STATS_NAME "stats" ++#define EXT3_MB_MAX_TO_SCAN_NAME "max_to_scan" ++#define EXT3_MB_MIN_TO_SCAN_NAME "min_to_scan" ++#define EXT3_MB_ORDER2_REQ "order2_req" ++#define EXT3_MB_STREAM_REQ "stream_req" ++ ++static int ext3_mb_stats_read(char *page, char **start, off_t off, ++ int count, int *eof, void *data) ++{ ++ struct ext3_sb_info *sbi = data; ++ int len; ++ ++ *eof = 1; ++ if (off != 0) ++ return 0; ++ ++ len = sprintf(page, "%ld\n", sbi->s_mb_stats); ++ *start = page; ++ return len; ++} ++ ++static int ext3_mb_stats_write(struct file *file, const char *buffer, ++ unsigned long count, void *data) ++{ ++ struct ext3_sb_info *sbi = data; ++ char str[32]; ++ ++ if (count >= sizeof(str)) { ++ printk(KERN_ERR "EXT3-fs: %s string too long, max %u bytes\n", ++ EXT3_MB_STATS_NAME, (int)sizeof(str)); ++ return -EOVERFLOW; ++ } ++ ++ if (copy_from_user(str, buffer, count)) ++ return -EFAULT; ++ ++ /* Only set to 0 or 1 respectively; zero->0; non-zero->1 */ ++ sbi->s_mb_stats = (simple_strtol(str, NULL, 0) != 0); ++ return count; ++} ++ ++static int ext3_mb_max_to_scan_read(char *page, char **start, off_t off, ++ int count, int *eof, void *data) ++{ ++ struct ext3_sb_info *sbi = data; ++ int len; ++ ++ *eof = 1; ++ if (off != 0) ++ return 0; ++ ++ len = sprintf(page, "%ld\n", sbi->s_mb_max_to_scan); ++ *start = page; ++ return len; ++} ++ ++static int ext3_mb_max_to_scan_write(struct file *file, const char *buffer, ++ unsigned long count, void *data) ++{ ++ struct ext3_sb_info *sbi = data; ++ char str[32]; ++ long value; ++ ++ if (count >= sizeof(str)) { ++ printk(KERN_ERR "EXT3-fs: %s string too long, max %u bytes\n", ++ EXT3_MB_MAX_TO_SCAN_NAME, (int)sizeof(str)); ++ return -EOVERFLOW; ++ } ++ ++ if (copy_from_user(str, buffer, count)) ++ return -EFAULT; ++ ++ /* Only set to 0 or 1 respectively; zero->0; non-zero->1 */ ++ value = simple_strtol(str, NULL, 0); ++ if (value <= 0) ++ return -ERANGE; ++ ++ sbi->s_mb_max_to_scan = value; ++ ++ return count; ++} ++ ++static int ext3_mb_min_to_scan_read(char *page, char **start, off_t off, ++ int count, int *eof, void *data) ++{ ++ struct ext3_sb_info *sbi = data; ++ int len; ++ ++ *eof = 1; ++ if (off != 0) ++ return 0; ++ ++ len = sprintf(page, "%ld\n", sbi->s_mb_min_to_scan); ++ *start = page; ++ return len; ++} ++ ++static int ext3_mb_order2_req_write(struct file *file, const char *buffer, ++ unsigned long count, void *data) ++{ ++ struct ext3_sb_info *sbi = data; ++ char str[32]; ++ long value; ++ ++ if (count >= sizeof(str)) { ++ printk(KERN_ERR "EXT3-fs: %s string too long, max %u bytes\n", ++ EXT3_MB_MIN_TO_SCAN_NAME, (int)sizeof(str)); ++ return -EOVERFLOW; ++ } ++ ++ if (copy_from_user(str, buffer, count)) ++ return -EFAULT; ++ ++ /* Only set to 0 or 1 respectively; zero->0; non-zero->1 */ ++ value = simple_strtol(str, NULL, 0); ++ if (value <= 0) ++ return -ERANGE; ++ ++ sbi->s_mb_order2_reqs = value; ++ ++ return count; ++} ++ ++static int ext3_mb_order2_req_read(char *page, char **start, off_t off, ++ int count, int *eof, void *data) ++{ ++ struct ext3_sb_info *sbi = data; ++ int len; ++ ++ *eof = 1; ++ if (off != 0) ++ return 0; ++ ++ len = sprintf(page, "%ld\n", sbi->s_mb_order2_reqs); ++ *start = page; ++ return len; ++} ++ ++static int ext3_mb_min_to_scan_write(struct file *file, const char *buffer, ++ unsigned long count, void *data) ++{ ++ struct ext3_sb_info *sbi = data; ++ char str[32]; ++ long value; ++ ++ if (count >= sizeof(str)) { ++ printk(KERN_ERR "EXT3-fs: %s string too long, max %u bytes\n", ++ EXT3_MB_MIN_TO_SCAN_NAME, (int)sizeof(str)); ++ return -EOVERFLOW; ++ } ++ ++ if (copy_from_user(str, buffer, count)) ++ return -EFAULT; ++ ++ /* Only set to 0 or 1 respectively; zero->0; non-zero->1 */ ++ value = simple_strtol(str, NULL, 0); ++ if (value <= 0) ++ return -ERANGE; ++ ++ sbi->s_mb_min_to_scan = value; ++ ++ return count; ++} ++ ++static int ext3_mb_stream_req_read(char *page, char **start, off_t off, ++ int count, int *eof, void *data) ++{ ++ struct ext3_sb_info *sbi = data; ++ int len; ++ ++ *eof = 1; ++ if (off != 0) ++ return 0; ++ ++ len = sprintf(page, "%ld\n", sbi->s_mb_stream_request); ++ *start = page; ++ return len; ++} ++ ++static int ext3_mb_stream_req_write(struct file *file, const char *buffer, ++ unsigned long count, void *data) ++{ ++ struct ext3_sb_info *sbi = data; ++ char str[32]; ++ long value; ++ ++ if (count >= sizeof(str)) { ++ printk(KERN_ERR "EXT3-fs: %s string too long, max %u bytes\n", ++ EXT3_MB_STREAM_REQ, (int)sizeof(str)); ++ return -EOVERFLOW; ++ } ++ ++ if (copy_from_user(str, buffer, count)) ++ return -EFAULT; ++ ++ /* Only set to 0 or 1 respectively; zero->0; non-zero->1 */ ++ value = simple_strtol(str, NULL, 0); ++ if (value <= 0) ++ return -ERANGE; ++ ++ sbi->s_mb_stream_request = value; ++ ++ return count; ++} ++ ++int ext3_mb_init_per_dev_proc(struct super_block *sb) ++{ ++ mode_t mode = S_IFREG | S_IRUGO | S_IWUSR; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ struct proc_dir_entry *proc; ++ char devname[64], *name; ++ ++ snprintf(devname, sizeof(devname) - 1, "%s", ++ bdevname(sb->s_bdev, devname)); ++ sbi->s_mb_proc = proc_mkdir(devname, proc_root_ext3); ++ ++ name = EXT3_MB_STATS_NAME; ++ proc = create_proc_entry(name, mode, sbi->s_mb_proc); ++ if (proc == NULL) ++ goto err_out; ++ proc->data = sbi; ++ proc->read_proc = ext3_mb_stats_read; ++ proc->write_proc = ext3_mb_stats_write; ++ ++ name = EXT3_MB_MAX_TO_SCAN_NAME; ++ proc = create_proc_entry(name, mode, sbi->s_mb_proc); ++ if (proc == NULL) ++ goto err_out; ++ proc->data = sbi; ++ proc->read_proc = ext3_mb_max_to_scan_read; ++ proc->write_proc = ext3_mb_max_to_scan_write; ++ ++ name = EXT3_MB_MIN_TO_SCAN_NAME; ++ proc = create_proc_entry(name, mode, sbi->s_mb_proc); ++ if (proc == NULL) ++ goto err_out; ++ proc->data = sbi; ++ proc->read_proc = ext3_mb_min_to_scan_read; ++ proc->write_proc = ext3_mb_min_to_scan_write; ++ ++ name = EXT3_MB_ORDER2_REQ; ++ proc = create_proc_entry(name, mode, sbi->s_mb_proc); ++ if (proc == NULL) ++ goto err_out; ++ proc->data = sbi; ++ proc->read_proc = ext3_mb_order2_req_read; ++ proc->write_proc = ext3_mb_order2_req_write; ++ ++ name = EXT3_MB_STREAM_REQ; ++ proc = create_proc_entry(name, mode, sbi->s_mb_proc); ++ if (proc == NULL) ++ goto err_out; ++ proc->data = sbi; ++ proc->read_proc = ext3_mb_stream_req_read; ++ proc->write_proc = ext3_mb_stream_req_write; ++ ++ return 0; ++ ++err_out: ++ printk(KERN_ERR "EXT3-fs: Unable to create %s\n", name); ++ remove_proc_entry(EXT3_MB_STREAM_REQ, sbi->s_mb_proc); ++ remove_proc_entry(EXT3_MB_ORDER2_REQ, sbi->s_mb_proc); ++ remove_proc_entry(EXT3_MB_MIN_TO_SCAN_NAME, sbi->s_mb_proc); ++ remove_proc_entry(EXT3_MB_MAX_TO_SCAN_NAME, sbi->s_mb_proc); ++ remove_proc_entry(EXT3_MB_STATS_NAME, sbi->s_mb_proc); ++ remove_proc_entry(devname, proc_root_ext3); ++ sbi->s_mb_proc = NULL; ++ ++ return -ENOMEM; ++} ++ ++int ext3_mb_destroy_per_dev_proc(struct super_block *sb) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ char devname[64]; ++ ++ if (sbi->s_mb_proc == NULL) ++ return -EINVAL; ++ ++ snprintf(devname, sizeof(devname) - 1, "%s", ++ bdevname(sb->s_bdev, devname)); ++ remove_proc_entry(EXT3_MB_STREAM_REQ, sbi->s_mb_proc); ++ remove_proc_entry(EXT3_MB_ORDER2_REQ, sbi->s_mb_proc); ++ remove_proc_entry(EXT3_MB_MIN_TO_SCAN_NAME, sbi->s_mb_proc); ++ remove_proc_entry(EXT3_MB_MAX_TO_SCAN_NAME, sbi->s_mb_proc); ++ remove_proc_entry(EXT3_MB_STATS_NAME, sbi->s_mb_proc); ++ remove_proc_entry(devname, proc_root_ext3); ++ ++ return 0; ++} ++ ++int __init init_ext3_proc(void) ++{ ++ ext3_pspace_cachep = ++ kmem_cache_create("ext3_prealloc_space", ++ sizeof(struct ext3_prealloc_space), ++ 0, SLAB_RECLAIM_ACCOUNT, NULL, NULL); ++ if (ext3_pspace_cachep == NULL) ++ return -ENOMEM; ++ ++ proc_root_ext3 = proc_mkdir(EXT3_ROOT, proc_root_fs); ++ if (proc_root_ext3 == NULL) ++ printk(KERN_ERR "EXT3-fs: Unable to create %s\n", EXT3_ROOT); ++ ++ return 0; ++} ++ ++void exit_ext3_proc(void) ++{ ++ /* XXX: synchronize_rcu(); */ ++ kmem_cache_destroy(ext3_pspace_cachep); ++ remove_proc_entry(EXT3_ROOT, proc_root_fs); ++} ++ ++ ++/* ++ * Check quota and mark choosed space (ac->ac_b_ex) non-free in bitmaps ++ * Returns 0 if success or error code ++ */ ++int ext3_mb_mark_diskspace_used(struct ext3_allocation_context *ac, handle_t *handle) ++{ ++ struct buffer_head *bitmap_bh = NULL; ++ struct ext3_super_block *es; ++ struct ext3_group_desc *gdp; ++ struct buffer_head *gdp_bh; ++ struct ext3_sb_info *sbi; ++ struct super_block *sb; ++ sector_t block; ++ int len, err; ++ ++ BUG_ON(ac->ac_status != AC_STATUS_FOUND); ++ BUG_ON(ac->ac_b_ex.fe_len <= 0); ++ ++ sb = ac->ac_sb; ++ sbi = EXT3_SB(sb); ++ es = sbi->s_es; ++ ++ ext3_debug("using block group %d(%d)\n", ac->ac_b_group.group, ++ gdp->bg_free_blocks_count); ++ ++ /* time to check quota, we can't do this before because ++ * having quota spent on preallocated-unused-yet blocks ++ * would be wrong */ ++ len = ac->ac_b_ex.fe_len; ++ while (len && DQUOT_ALLOC_BLOCK(ac->ac_inode, len)) len--; ++ if (ac->ac_b_ex.fe_len != len) { ++ /* some blocks can't be allocated due to quota ++ * we have to return them back */ ++ BUG(); ++ } ++ err = -EDQUOT; ++ if (len == 0) ++ goto out_err; ++ ++ err = -EIO; ++ bitmap_bh = read_block_bitmap(sb, ac->ac_b_ex.fe_group); ++ if (!bitmap_bh) ++ goto out_err; ++ ++ err = ext3_journal_get_write_access(handle, bitmap_bh); ++ if (err) ++ goto out_err; ++ ++ err = -EIO; ++ gdp = ext3_get_group_desc(sb, ac->ac_b_ex.fe_group, &gdp_bh); ++ if (!gdp) ++ goto out_err; ++ ++ err = ext3_journal_get_write_access(handle, gdp_bh); ++ if (err) ++ goto out_err; ++ ++ block = ac->ac_b_ex.fe_group * EXT3_BLOCKS_PER_GROUP(sb) ++ + ac->ac_b_ex.fe_start ++ + le32_to_cpu(es->s_first_data_block); ++ ++ if (block == le32_to_cpu(gdp->bg_block_bitmap) || ++ block == le32_to_cpu(gdp->bg_inode_bitmap) || ++ in_range(block, le32_to_cpu(gdp->bg_inode_table), ++ EXT3_SB(sb)->s_itb_per_group)) ++ ext3_error(sb, __FUNCTION__, ++ "Allocating block in system zone - block = %lu", ++ (unsigned long) block); ++#ifdef AGGRESSIVE_CHECK ++ { ++ int i; ++ for (i = 0; i < ac->ac_b_ex.fe_len; i++) { ++ BUG_ON(mb_test_bit(ac->ac_b_ex.fe_start + i, ++ bitmap_bh->b_data)); ++ } ++ } ++#endif ++ mb_set_bits(bitmap_bh->b_data, ac->ac_b_ex.fe_start, ac->ac_b_ex.fe_len); ++ ++ spin_lock(sb_bgl_lock(sbi, ac->ac_b_ex.fe_group)); ++ gdp->bg_free_blocks_count = ++ cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) ++ - ac->ac_b_ex.fe_len); ++ spin_unlock(sb_bgl_lock(sbi, ac->ac_b_ex.fe_group)); ++ percpu_counter_mod(&sbi->s_freeblocks_counter, - ac->ac_b_ex.fe_len); ++ ++ err = ext3_journal_dirty_metadata(handle, bitmap_bh); ++ if (err) ++ goto out_err; ++ err = ext3_journal_dirty_metadata(handle, gdp_bh); ++ ++out_err: ++ sb->s_dirt = 1; ++ brelse(bitmap_bh); ++ return err; ++} ++ ++/* ++ * here we normalize request for locality group ++ * XXX: should we try to preallocate more than the group has now? ++ */ ++void ext3_mb_normalize_group_request(struct ext3_allocation_context *ac) ++{ ++ struct super_block *sb = ac->ac_sb; ++ struct ext3_locality_group *lg = ac->ac_lg; ++ ++ BUG_ON(lg == NULL); ++ if (EXT3_SB(sb)->s_stripe) ++ ac->ac_g_ex.fe_len = EXT3_SB(sb)->s_stripe; ++ else ++ ac->ac_g_ex.fe_len = (1024 * 1024) >> sb->s_blocksize_bits; ++ ++ mb_debug("#%u: goal %u blocks for locality group\n", ++ current->pid, ac->ac_g_ex.fe_len); ++} ++ ++/* ++ * Normalization means making request better in terms of ++ * size and alignment ++ */ ++void ext3_mb_normalize_request(struct ext3_allocation_context *ac, ++ struct ext3_allocation_request *ar) ++{ ++ struct ext3_inode_info *ei = EXT3_I(ac->ac_inode); ++ loff_t start, end, size, orig_size, orig_start; ++ struct list_head *cur; ++ int bsbits; ++ ++ /* do normalize only data requests, metadata requests ++ do not need preallocation */ ++ if (!(ac->ac_flags & EXT3_MB_HINT_DATA)) ++ return; ++ ++ /* sometime caller may want exact blocks */ ++ if (unlikely(ac->ac_flags & EXT3_MB_HINT_GOAL_ONLY)) ++ return; ++ ++ /* caller may indicate that preallocation isn't ++ * required (it's a tail, for example) */ ++ if (ac->ac_flags & EXT3_MB_HINT_NOPREALLOC) ++ return; ++ ++ if (ac->ac_flags & EXT3_MB_HINT_GROUP_ALLOC) ++ return ext3_mb_normalize_group_request(ac); ++ ++ bsbits = ac->ac_sb->s_blocksize_bits; ++ ++ /* first, let's learn actual file size ++ * given current request is allocated */ ++ size = ac->ac_o_ex.fe_logical + ac->ac_o_ex.fe_len; ++ size = size << bsbits; ++ if (size < i_size_read(ac->ac_inode)) ++ size = i_size_read(ac->ac_inode); ++ ++ /* first, try to predict filesize */ ++ /* XXX: should this table be tunable? */ ++ start = 0; ++ if (size <= 16 * 1024) { ++ size = 16 * 1024; ++ } else if (size <= 32 * 1024) { ++ size = 32 * 1024; ++ } else if (size <= 64 * 1024) { ++ size = 64 * 1024; ++ } else if (size <= 128 * 1024) { ++ size = 128 * 1024; ++ } else if (size <= 256 * 1024) { ++ size = 256 * 1024; ++ } else if (size <= 512 * 1024) { ++ size = 512 * 1024; ++ } else if (size <= 1024 * 1024) { ++ size = 1024 * 1024; ++ } else if (size < 4 * 1024 * 1024) { ++ start = ac->ac_o_ex.fe_logical << bsbits; ++ start = (start / (1024 * 1024)) * (1024 * 1024); ++ size = 1024 * 1024; ++ } else if (size < 8 * 1024 * 1024) { ++ start = ac->ac_o_ex.fe_logical << bsbits; ++ start = (start / (4 * (1024 * 1024))) * 4 * (1024 * 1024); ++ size = 4 * 1024 * 1024; ++ } else if (ac->ac_o_ex.fe_len < ((8 << 20) >> bsbits)) { ++ start = ac->ac_o_ex.fe_logical; ++ start = start << bsbits; ++ start = (start / (8 * (1024 * 1024))) * 8 * (1024 * 1024); ++ size = 8 * 1024 * 1024; ++ } else { ++ start = ac->ac_o_ex.fe_logical; ++ start = start << bsbits; ++ size = ac->ac_o_ex.fe_len << bsbits; ++ } ++ orig_size = size = size >> bsbits; ++ orig_start = start = start >> bsbits; ++ ++ /* don't cover already allocated blocks in selected range */ ++ if (ar->pleft && start <= ar->lleft) { ++ size -= ar->lleft + 1 - start; ++ start = ar->lleft + 1; ++ } ++ if (ar->pright && start + size - 1 >= ar->lright) ++ size -= start + size - ar->lright; ++ ++ end = start + size; ++ ++ /* check we don't cross already preallocated blocks */ ++ rcu_read_lock(); ++ list_for_each_rcu(cur, &ei->i_prealloc_list) { ++ struct ext3_prealloc_space *pa; ++ unsigned long pa_end; ++ ++ pa = list_entry(cur, struct ext3_prealloc_space, pa_inode_list); ++ pa_end = pa->pa_lstart + pa->pa_len; ++ ++ /* PA must not overlap original request */ ++ BUG_ON(!(ac->ac_o_ex.fe_logical >= pa_end || ++ ac->ac_o_ex.fe_logical < pa->pa_lstart)); ++ ++ /* skip PA normalized request doesn't overlap with */ ++ if (pa->pa_lstart >= end) ++ continue; ++ if (pa_end <= start) ++ continue; ++ BUG_ON(pa->pa_lstart <= start && pa_end >= end); ++ ++ if (pa_end <= ac->ac_o_ex.fe_logical) { ++ BUG_ON(pa_end < start); ++ start = pa_end; ++ } ++ ++ if (pa->pa_lstart > ac->ac_o_ex.fe_logical) { ++ BUG_ON(pa->pa_lstart > end); ++ end = pa->pa_lstart; ++ } ++ } ++ rcu_read_unlock(); ++ size = end - start; ++ ++ /* XXX: extra loop to check we really don't overlap preallocations */ ++ rcu_read_lock(); ++ list_for_each_rcu(cur, &ei->i_prealloc_list) { ++ struct ext3_prealloc_space *pa; ++ unsigned long pa_end; ++ pa = list_entry(cur, struct ext3_prealloc_space, pa_inode_list); ++ pa_end = pa->pa_lstart + pa->pa_len; ++ BUG_ON(!(start >= pa_end || end <= pa->pa_lstart)); ++ } ++ rcu_read_unlock(); ++ ++ if (start + size <= ac->ac_o_ex.fe_logical && ++ start > ac->ac_o_ex.fe_logical) { ++ printk("start %lu, size %lu, fe_logical %lu\n", ++ (unsigned long) start, (unsigned long) size, ++ (unsigned long) ac->ac_o_ex.fe_logical); ++ } ++ BUG_ON(start + size <= ac->ac_o_ex.fe_logical && ++ start > ac->ac_o_ex.fe_logical); ++ ++ /* now prepare goal request */ ++ BUG_ON(size <= 0 || size >= EXT3_BLOCKS_PER_GROUP(ac->ac_sb)); ++ if (size < ac->ac_o_ex.fe_len) { ++ /* XXX: don't normalize tails? */ ++ } ++ ++ /* XXX: is it better to align blocks WRT to logical placement ++ * or satisfy big request as is */ ++ ac->ac_g_ex.fe_logical = start; ++ ac->ac_g_ex.fe_len = size; ++ ++ mb_debug("goal: %u(was %u) blocks at %u\n", (unsigned) size, ++ (unsigned) orig_size, (unsigned) start); ++} ++ ++void ext3_mb_collect_stats(struct ext3_allocation_context *ac) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(ac->ac_sb); ++ ++ if (sbi->s_mb_stats && ac->ac_g_ex.fe_len > 1) { ++ atomic_inc(&sbi->s_bal_reqs); ++ atomic_add(ac->ac_b_ex.fe_len, &sbi->s_bal_allocated); ++ if (ac->ac_o_ex.fe_len >= ac->ac_g_ex.fe_len) ++ atomic_inc(&sbi->s_bal_success); ++ atomic_add(ac->ac_found, &sbi->s_bal_ex_scanned); ++ if (ac->ac_g_ex.fe_start == ac->ac_b_ex.fe_start && ++ ac->ac_g_ex.fe_group == ac->ac_b_ex.fe_group) ++ atomic_inc(&sbi->s_bal_goals); ++ if (ac->ac_found > sbi->s_mb_max_to_scan) ++ atomic_inc(&sbi->s_bal_breaks); ++ } ++ ++ ext3_mb_store_history(ac); ++} ++ ++/* ++ * use blocks preallocated to inode ++ */ ++void ext3_mb_use_inode_pa(struct ext3_allocation_context *ac, ++ struct ext3_prealloc_space *pa) ++{ ++ unsigned long start, len; ++ ++ /* found preallocated blocks, use them */ ++ start = pa->pa_pstart + (ac->ac_o_ex.fe_logical - pa->pa_lstart); ++ len = min(pa->pa_pstart + pa->pa_len, start + ac->ac_o_ex.fe_len); ++ len = len - start; ++ ext3_get_group_no_and_offset(ac->ac_sb, start, &ac->ac_b_ex.fe_group, ++ &ac->ac_b_ex.fe_start); ++ ac->ac_b_ex.fe_len = len; ++ ac->ac_status = AC_STATUS_FOUND; ++ ac->ac_pa = pa; ++ ++ BUG_ON(start < pa->pa_pstart); ++ BUG_ON(start + len > pa->pa_pstart + pa->pa_len); ++ BUG_ON(pa->pa_free < len); ++ pa->pa_free -= len; ++ ++ mb_debug("use %lu/%lu from inode pa %p\n", start, len, pa); ++} ++ ++/* ++ * use blocks preallocated to locality group ++ */ ++void ext3_mb_use_group_pa(struct ext3_allocation_context *ac, ++ struct ext3_prealloc_space *pa) ++{ ++ unsigned len = ac->ac_o_ex.fe_len; ++ ++ ext3_get_group_no_and_offset(ac->ac_sb, pa->pa_pstart, ++ &ac->ac_b_ex.fe_group, ++ &ac->ac_b_ex.fe_start); ++ ac->ac_b_ex.fe_len = len; ++ ac->ac_status = AC_STATUS_FOUND; ++ ac->ac_pa = pa; ++ ++ /* we don't correct pa_pstart or pa_plen here to avoid ++ * possible race when tte group is being loaded concurrently ++ * instead we correct pa later, after blocks are marked ++ * in on-disk bitmap -- see ext3_mb_release_context() */ ++ mb_debug("use %lu/%lu from group pa %p\n", pa->pa_lstart-len, len, pa); ++} ++ ++/* ++ * search goal blocks in preallocated space ++ */ ++int ext3_mb_use_preallocated(struct ext3_allocation_context *ac) ++{ ++ struct ext3_inode_info *ei = EXT3_I(ac->ac_inode); ++ struct ext3_locality_group *lg; ++ struct ext3_prealloc_space *pa; ++ struct list_head *cur; ++ ++ /* only data can be preallocated */ ++ if (!(ac->ac_flags & EXT3_MB_HINT_DATA)) ++ return 0; ++ ++ /* first, try per-file preallocation */ ++ rcu_read_lock(); ++ list_for_each_rcu(cur, &ei->i_prealloc_list) { ++ pa = list_entry(cur, struct ext3_prealloc_space, pa_inode_list); ++ ++ /* all fields in this condition don't change, ++ * so we can skip locking for them */ ++ if (ac->ac_o_ex.fe_logical < pa->pa_lstart || ++ ac->ac_o_ex.fe_logical >= pa->pa_lstart + pa->pa_len) ++ continue; ++ ++ /* found preallocated blocks, use them */ ++ spin_lock(&pa->pa_lock); ++ if (pa->pa_deleted == 0 && pa->pa_free) { ++ atomic_inc(&pa->pa_count); ++ ext3_mb_use_inode_pa(ac, pa); ++ spin_unlock(&pa->pa_lock); ++ ac->ac_criteria = 10; ++ rcu_read_unlock(); ++ return 1; ++ } ++ spin_unlock(&pa->pa_lock); ++ } ++ rcu_read_unlock(); ++ ++ /* can we use group allocation? */ ++ if (!(ac->ac_flags & EXT3_MB_HINT_GROUP_ALLOC)) ++ return 0; ++ ++ /* inode may have no locality group for some reason */ ++ lg = ac->ac_lg; ++ if (lg == NULL) ++ return 0; ++ ++ rcu_read_lock(); ++ list_for_each_rcu(cur, &lg->lg_prealloc_list) { ++ pa = list_entry(cur, struct ext3_prealloc_space, pa_inode_list); ++ spin_lock(&pa->pa_lock); ++ if (pa->pa_deleted == 0 && pa->pa_free >= ac->ac_o_ex.fe_len) { ++ atomic_inc(&pa->pa_count); ++ ext3_mb_use_group_pa(ac, pa); ++ spin_unlock(&pa->pa_lock); ++ ac->ac_criteria = 20; ++ rcu_read_unlock(); ++ return 1; ++ } ++ spin_unlock(&pa->pa_lock); ++ } ++ rcu_read_unlock(); ++ ++ return 0; ++} ++ ++/* ++ * the function goes through all preallocation in this group and marks them ++ * used in in-core bitmap. buddy must be generated from this bitmap ++ */ ++void ext3_mb_generate_from_pa(struct super_block *sb, void *bitmap, int group) ++{ ++ struct ext3_group_info *grp = EXT3_GROUP_INFO(sb, group); ++ struct ext3_prealloc_space *pa; ++ struct list_head *cur; ++ unsigned long groupnr; ++ unsigned long start; ++ int preallocated = 0, count = 0, len; ++ ++ /* all form of preallocation discards first load group, ++ * so the only competing code is preallocation use. ++ * we don't need any locking here ++ * notice we do NOT ignore preallocations with pa_deleted ++ * otherwise we could leave used blocks available for ++ * allocation in buddy when concurrent ext3_mb_put_pa() ++ * is dropping preallocation ++ */ ++ list_for_each_rcu(cur, &grp->bb_prealloc_list) { ++ pa = list_entry(cur, struct ext3_prealloc_space, pa_group_list); ++ spin_lock(&pa->pa_lock); ++ ext3_get_group_no_and_offset(sb, pa->pa_pstart, &groupnr, &start); ++ len = pa->pa_len; ++ spin_unlock(&pa->pa_lock); ++ BUG_ON(groupnr != group); ++ mb_set_bits(bitmap, start, len); ++ preallocated += len; ++ count++; ++ } ++ mb_debug("prellocated %u for group %u\n", preallocated, group); ++} ++ ++#if LINUX_VERSION_CODE > KERNEL_VERSION(2,6,5) ++static void ext3_mb_pa_callback(struct rcu_head *head) ++{ ++ struct ext3_prealloc_space *pa; ++ pa = container_of(head, struct ext3_prealloc_space, u.pa_rcu); ++ kmem_cache_free(ext3_pspace_cachep, pa); ++} ++#define mb_call_rcu(__pa) call_rcu(&(__pa)->u.pa_rcu, ext3_mb_pa_callback) ++#else ++static void ext3_mb_pa_callback(void *pa) ++{ ++ kmem_cache_free(ext3_pspace_cachep, pa); ++} ++#define mb_call_rcu(__pa) call_rcu(&(__pa)->u.pa_rcu, ext3_mb_pa_callback, pa) ++#endif ++ ++/* ++ * drops a reference to preallocated space descriptor ++ * if this was the last reference and the space is consumed ++ */ ++void ext3_mb_put_pa(struct ext3_allocation_context *ac, ++ struct super_block *sb, struct ext3_prealloc_space *pa) ++{ ++ unsigned long grp; ++ ++ if (!atomic_dec_and_test(&pa->pa_count) || pa->pa_free != 0) ++ return; ++ ++ /* in this short window concurrent discard can set pa_deleted */ ++ spin_lock(&pa->pa_lock); ++ if (pa->pa_deleted == 0) { ++ spin_unlock(&pa->pa_lock); ++ return; ++ } ++ ++ pa->pa_deleted = 1; ++ spin_unlock(&pa->pa_lock); ++ ++ /* -1 is to protect from crossing allocation group */ ++ ext3_get_group_no_and_offset(sb, pa->pa_pstart - 1, &grp, NULL); ++ ++ /* ++ * possible race: ++ * ++ * P1 (buddy init) P2 (regular allocation) ++ * find block B in PA ++ * copy on-disk bitmap to buddy ++ * mark B in on-disk bitmap ++ * drop PA from group ++ * mark all PAs in buddy ++ * ++ * thus, P1 initializes buddy with B available. to prevent this ++ * we make "copy" and "mark all PAs" atomic and serialize "drop PA" ++ * against that pair ++ */ ++ ext3_lock_group(sb, grp); ++ list_del_rcu(&pa->pa_group_list); ++ ext3_unlock_group(sb, grp); ++ ++ spin_lock(pa->pa_obj_lock); ++ list_del_rcu(&pa->pa_inode_list); ++ spin_unlock(pa->pa_obj_lock); ++ ++ mb_call_rcu(pa); ++} ++ ++/* ++ * creates new preallocated space for given inode ++ */ ++int ext3_mb_new_inode_pa(struct ext3_allocation_context *ac) ++{ ++ struct super_block *sb = ac->ac_sb; ++ struct ext3_prealloc_space *pa; ++ struct ext3_group_info *grp; ++ struct ext3_inode_info *ei; ++ ++ /* preallocate only when found space is larger then requested */ ++ BUG_ON(ac->ac_o_ex.fe_len >= ac->ac_b_ex.fe_len); ++ BUG_ON(ac->ac_status != AC_STATUS_FOUND); ++ BUG_ON(!S_ISREG(ac->ac_inode->i_mode)); ++ ++ pa = kmem_cache_alloc(ext3_pspace_cachep, SLAB_NOFS); ++ if (pa == NULL) ++ return -ENOMEM; ++ ++ if (ac->ac_b_ex.fe_len < ac->ac_g_ex.fe_len) { ++ int winl, wins, win, offs; ++ ++ /* we can't allocate as much as normalizer wants. ++ * so, found space must get proper lstart ++ * to cover original request */ ++ BUG_ON(ac->ac_g_ex.fe_logical > ac->ac_o_ex.fe_logical); ++ BUG_ON(ac->ac_g_ex.fe_len < ac->ac_o_ex.fe_len); ++ ++ /* we're limited by original request in that ++ * logical block must be covered any way ++ * winl is window we can move our chunk within */ ++ winl = ac->ac_o_ex.fe_logical - ac->ac_g_ex.fe_logical; ++ ++ /* also, we should cover whole original request */ ++ wins = ac->ac_b_ex.fe_len - ac->ac_o_ex.fe_len; ++ ++ /* the smallest one defines real window */ ++ win = min(winl, wins); ++ ++ offs = ac->ac_o_ex.fe_logical % ac->ac_b_ex.fe_len; ++ if (offs && offs < win) ++ win = offs; ++ ++ ac->ac_b_ex.fe_logical = ac->ac_o_ex.fe_logical - win; ++ BUG_ON(ac->ac_o_ex.fe_logical < ac->ac_b_ex.fe_logical); ++ BUG_ON(ac->ac_o_ex.fe_len > ac->ac_b_ex.fe_len); ++ } ++ ++ /* preallocation can change ac_b_ex, thus we store actually ++ * allocated blocks for history */ ++ ac->ac_f_ex = ac->ac_b_ex; ++ ++ pa->pa_lstart = ac->ac_b_ex.fe_logical; ++ pa->pa_pstart = ext3_grp_offs_to_block(sb, &ac->ac_b_ex); ++ pa->pa_len = ac->ac_b_ex.fe_len; ++ pa->pa_free = pa->pa_len; ++ atomic_set(&pa->pa_count, 1); ++ spin_lock_init(&pa->pa_lock); ++ pa->pa_deleted = 0; ++ pa->pa_linear = 0; ++ ++ mb_debug("new inode pa %p: %lu/%lu for %lu\n", pa, ++ pa->pa_pstart, pa->pa_len, pa->pa_lstart); ++ ++ ext3_mb_use_inode_pa(ac, pa); ++ atomic_add(pa->pa_free, &EXT3_SB(sb)->s_mb_preallocated); ++ ++ ei = EXT3_I(ac->ac_inode); ++ grp = EXT3_GROUP_INFO(sb, ac->ac_b_ex.fe_group); ++ ++ pa->pa_obj_lock = &ei->i_prealloc_lock; ++ pa->pa_inode = ac->ac_inode; ++ ++ ext3_lock_group(sb, ac->ac_b_ex.fe_group); ++ list_add_rcu(&pa->pa_group_list, &grp->bb_prealloc_list); ++ ext3_unlock_group(sb, ac->ac_b_ex.fe_group); ++ ++ spin_lock(pa->pa_obj_lock); ++ list_add_rcu(&pa->pa_inode_list, &ei->i_prealloc_list); ++ spin_unlock(pa->pa_obj_lock); ++ ++ return 0; ++} ++ ++/* ++ * creates new preallocated space for locality group inodes belongs to ++ */ ++int ext3_mb_new_group_pa(struct ext3_allocation_context *ac) ++{ ++ struct super_block *sb = ac->ac_sb; ++ struct ext3_locality_group *lg; ++ struct ext3_prealloc_space *pa; ++ struct ext3_group_info *grp; ++ ++ /* preallocate only when found space is larger then requested */ ++ BUG_ON(ac->ac_o_ex.fe_len >= ac->ac_b_ex.fe_len); ++ BUG_ON(ac->ac_status != AC_STATUS_FOUND); ++ BUG_ON(!S_ISREG(ac->ac_inode->i_mode)); ++ ++ BUG_ON(ext3_pspace_cachep == NULL); ++ pa = kmem_cache_alloc(ext3_pspace_cachep, SLAB_NOFS); ++ if (pa == NULL) ++ return -ENOMEM; ++ ++ /* preallocation can change ac_b_ex, thus we store actually ++ * allocated blocks for history */ ++ ac->ac_f_ex = ac->ac_b_ex; ++ ++ pa->pa_pstart = ext3_grp_offs_to_block(sb, &ac->ac_b_ex); ++ pa->pa_lstart = pa->pa_pstart; ++ pa->pa_len = ac->ac_b_ex.fe_len; ++ pa->pa_free = pa->pa_len; ++ atomic_set(&pa->pa_count, 1); ++ spin_lock_init(&pa->pa_lock); ++ pa->pa_deleted = 0; ++ pa->pa_linear = 1; ++ ++ mb_debug("new group pa %p: %lu/%lu for %lu\n", pa, ++ pa->pa_pstart, pa->pa_len, pa->pa_lstart); ++ ++ ext3_mb_use_group_pa(ac, pa); ++ atomic_add(pa->pa_free, &EXT3_SB(sb)->s_mb_preallocated); ++ ++ grp = EXT3_GROUP_INFO(sb, ac->ac_b_ex.fe_group); ++ lg = ac->ac_lg; ++ BUG_ON(lg == NULL); ++ ++ pa->pa_obj_lock = &lg->lg_prealloc_lock; ++ pa->pa_inode = NULL; ++ ++ ext3_lock_group(sb, ac->ac_b_ex.fe_group); ++ list_add_rcu(&pa->pa_group_list, &grp->bb_prealloc_list); ++ ext3_unlock_group(sb, ac->ac_b_ex.fe_group); ++ ++ spin_lock(pa->pa_obj_lock); ++ list_add_tail_rcu(&pa->pa_inode_list, &lg->lg_prealloc_list); ++ spin_unlock(pa->pa_obj_lock); ++ ++ return 0; ++} ++ ++int ext3_mb_new_preallocation(struct ext3_allocation_context *ac) ++{ ++ int err; ++ ++ if (ac->ac_flags & EXT3_MB_HINT_GROUP_ALLOC) ++ err = ext3_mb_new_group_pa(ac); ++ else ++ err = ext3_mb_new_inode_pa(ac); ++ return err; ++} ++ ++/* ++ * finds all unused blocks in on-disk bitmap, frees them in ++ * in-core bitmap and buddy. ++ * @pa must be unlinked from inode and group lists, so that ++ * nobody else can find/use it. ++ * the caller MUST hold group/inode locks. ++ * TODO: optimize the case when there are no in-core structures yet ++ */ ++int ext3_mb_release_inode_pa(struct ext3_buddy *e3b, ++ struct buffer_head *bitmap_bh, ++ struct ext3_prealloc_space *pa) ++{ ++ struct ext3_allocation_context ac; ++ struct super_block *sb = e3b->bd_sb; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ unsigned long bit, end, next, group; ++ sector_t start; ++ int err = 0, free = 0; ++ ++ BUG_ON(pa->pa_deleted == 0); ++ ext3_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit); ++ BUG_ON(group != e3b->bd_group); ++ end = bit + pa->pa_len; ++ ++ ac.ac_sb = sb; ++ ac.ac_inode = pa->pa_inode; ++ ac.ac_op = EXT3_MB_HISTORY_DISCARD; ++ ++ while (bit < end) { ++ bit = mb_find_next_zero_bit(bitmap_bh->b_data, end, bit); ++ if (bit >= end) ++ break; ++ next = mb_find_next_bit(bitmap_bh->b_data, end, bit); ++ if (next > end) ++ next = end; ++ start = group * EXT3_BLOCKS_PER_GROUP(sb) + bit + ++ le32_to_cpu(sbi->s_es->s_first_data_block); ++ mb_debug(" free preallocated %u/%u in group %u\n", ++ (unsigned) start, (unsigned) next - bit, ++ (unsigned) group); ++ free += next - bit; ++ ++ ac.ac_b_ex.fe_group = group; ++ ac.ac_b_ex.fe_start = bit; ++ ac.ac_b_ex.fe_len = next - bit; ++ ac.ac_b_ex.fe_logical = 0; ++ ext3_mb_store_history(&ac); ++ ++ mb_free_blocks(e3b, bit, next - bit); ++ bit = next + 1; ++ } ++ if (free != pa->pa_free) { ++ printk("pa %p: logic %lu, phys. %lu, len %lu\n", ++ pa, (unsigned long) pa->pa_lstart, ++ (unsigned long) pa->pa_pstart, ++ (unsigned long) pa->pa_len); ++ printk("free %u, pa_free %u\n", free, pa->pa_free); ++ } ++ BUG_ON(free != pa->pa_free); ++ atomic_add(free, &sbi->s_mb_discarded); ++ ++ return err; ++} ++ ++int ext3_mb_release_group_pa(struct ext3_buddy *e3b, ++ struct ext3_prealloc_space *pa) ++{ ++ struct ext3_allocation_context ac; ++ struct super_block *sb = e3b->bd_sb; ++ unsigned long bit, group; ++ ++ ac.ac_op = EXT3_MB_HISTORY_DISCARD; ++ ++ BUG_ON(pa->pa_deleted == 0); ++ ext3_get_group_no_and_offset(sb, pa->pa_pstart, &group, &bit); ++ BUG_ON(group != e3b->bd_group); ++ mb_free_blocks(e3b, bit, pa->pa_len); ++ atomic_add(pa->pa_len, &EXT3_SB(sb)->s_mb_discarded); ++ ++ ac.ac_sb = sb; ++ ac.ac_inode = NULL; ++ ac.ac_b_ex.fe_group = group; ++ ac.ac_b_ex.fe_start = bit; ++ ac.ac_b_ex.fe_len = pa->pa_len; ++ ac.ac_b_ex.fe_logical = 0; ++ ext3_mb_store_history(&ac); ++ ++ return 0; ++} ++ ++/* ++ * releases all preallocations in given group ++ * ++ * first, we need to decide discard policy: ++ * - when do we discard ++ * 1) ENOSPC ++ * - how many do we discard ++ * 1) how many requested ++ */ ++int ext3_mb_discard_group_preallocations(struct super_block *sb, ++ int group, int needed) ++{ ++ struct ext3_group_info *grp = EXT3_GROUP_INFO(sb, group); ++ struct buffer_head *bitmap_bh = NULL; ++ struct ext3_prealloc_space *pa, *tmp; ++ struct list_head list; ++ struct ext3_buddy e3b; ++ int err, busy, free = 0; ++ ++ mb_debug("discard preallocation for group %lu\n", group); ++ ++ if (list_empty(&grp->bb_prealloc_list)) ++ return 0; ++ ++ bitmap_bh = read_block_bitmap(sb, group); ++ if (bitmap_bh == NULL) { ++ /* error handling here */ ++ ext3_mb_release_desc(&e3b); ++ BUG_ON(bitmap_bh == NULL); ++ } ++ ++ err = ext3_mb_load_buddy(sb, group, &e3b); ++ BUG_ON(err != 0); /* error handling here */ ++ ++ if (needed == 0) ++ needed = EXT3_BLOCKS_PER_GROUP(sb) + 1; ++ ++ grp = EXT3_GROUP_INFO(sb, group); ++ INIT_LIST_HEAD(&list); ++ ++repeat: ++ busy = 0; ++ ext3_lock_group(sb, group); ++ list_for_each_entry_safe (pa, tmp, &grp->bb_prealloc_list, pa_group_list) { ++ spin_lock(&pa->pa_lock); ++ if (atomic_read(&pa->pa_count)) { ++ spin_unlock(&pa->pa_lock); ++ printk("uh! busy PA\n"); ++ dump_stack(); ++ busy = 1; ++ continue; ++ } ++ if (pa->pa_deleted) { ++ spin_unlock(&pa->pa_lock); ++ continue; ++ } ++ ++ /* seems this one can be freed ... */ ++ pa->pa_deleted = 1; ++ ++ /* we can trust pa_free ... */ ++ free += pa->pa_free; ++ ++ spin_unlock(&pa->pa_lock); ++ ++ list_del_rcu(&pa->pa_group_list); ++ list_add(&pa->u.pa_tmp_list, &list); ++ } ++ ++ /* if we still need more blocks and some PAs were used, try again */ ++ if (free < needed && busy) ++ goto repeat; ++ ++ /* found anything to free? */ ++ if (list_empty(&list)) { ++ BUG_ON(free != 0); ++ goto out; ++ } ++ ++ /* now free all selected PAs */ ++ list_for_each_entry_safe(pa, tmp, &list, u.pa_tmp_list) { ++ ++ /* remove from object (inode or locality group) */ ++ spin_lock(pa->pa_obj_lock); ++ list_del_rcu(&pa->pa_inode_list); ++ spin_unlock(pa->pa_obj_lock); ++ ++ if (pa->pa_linear) ++ ext3_mb_release_group_pa(&e3b, pa); ++ else ++ ext3_mb_release_inode_pa(&e3b, bitmap_bh, pa); ++ ++ list_del(&pa->u.pa_tmp_list); ++ mb_call_rcu(pa); ++ } ++ ++out: ++ ext3_unlock_group(sb, group); ++ ext3_mb_release_desc(&e3b); ++ brelse(bitmap_bh); ++ return free; ++} ++ ++/* ++ * releases all non-used preallocated blocks for given inode ++ */ ++void ext3_mb_discard_inode_preallocations(struct inode *inode) ++{ ++ struct ext3_inode_info *ei = EXT3_I(inode); ++ struct super_block *sb = inode->i_sb; ++ struct buffer_head *bitmap_bh = NULL; ++ struct ext3_prealloc_space *pa, *tmp; ++ unsigned long group = 0; ++ struct list_head list; ++ struct ext3_buddy e3b; ++ int err; ++ ++ if (!test_opt(sb, MBALLOC) || !S_ISREG(inode->i_mode)) { ++ /*BUG_ON(!list_empty(&ei->i_prealloc_list));*/ ++ return; ++ } ++ ++ mb_debug("discard preallocation for inode %lu\n", inode->i_ino); ++ ++ INIT_LIST_HEAD(&list); ++ ++repeat: ++ /* first, collect all pa's in the inode */ ++ spin_lock(&ei->i_prealloc_lock); ++ while (!list_empty(&ei->i_prealloc_list)) { ++ pa = list_entry(ei->i_prealloc_list.next, ++ struct ext3_prealloc_space, pa_inode_list); ++ BUG_ON(pa->pa_obj_lock != &ei->i_prealloc_lock); ++ spin_lock(&pa->pa_lock); ++ if (atomic_read(&pa->pa_count)) { ++ /* this shouldn't happen often - nobody should ++ * use preallocation while we're discarding it */ ++ spin_unlock(&pa->pa_lock); ++ spin_unlock(&ei->i_prealloc_lock); ++ printk("uh-oh! used pa while discarding\n"); ++ dump_stack(); ++ current->state = TASK_UNINTERRUPTIBLE; ++ schedule_timeout(HZ); ++ goto repeat; ++ ++ } ++ if (pa->pa_deleted == 0) { ++ pa->pa_deleted = 1; ++ spin_unlock(&pa->pa_lock); ++ list_del_rcu(&pa->pa_inode_list); ++ list_add(&pa->u.pa_tmp_list, &list); ++ continue; ++ } ++ ++ /* someone is deleting pa right now */ ++ spin_unlock(&pa->pa_lock); ++ spin_unlock(&ei->i_prealloc_lock); ++ ++ /* we have to wait here because pa_deleted ++ * doesn't mean pa is already unlinked from ++ * the list. as we might be called from ++ * ->clear_inode() the inode will get freed ++ * and concurrent thread which is unlinking ++ * pa from inode's list may access already ++ * freed memory, bad-bad-bad */ ++ ++ /* XXX: if this happens too often, we can ++ * add a flag to force wait only in case ++ * of ->clear_inode(), but not in case of ++ * regular truncate */ ++ printk("uh-oh! some one just deleted it\n"); ++ dump_stack(); ++ current->state = TASK_UNINTERRUPTIBLE; ++ schedule_timeout(HZ); ++ goto repeat; ++ } ++ spin_unlock(&ei->i_prealloc_lock); ++ ++ list_for_each_entry_safe(pa, tmp, &list, u.pa_tmp_list) { ++ BUG_ON(pa->pa_linear != 0); ++ ext3_get_group_no_and_offset(sb, pa->pa_pstart, &group, NULL); ++ ++ err = ext3_mb_load_buddy(sb, group, &e3b); ++ BUG_ON(err != 0); /* error handling here */ ++ ++ bitmap_bh = read_block_bitmap(sb, group); ++ if (bitmap_bh == NULL) { ++ /* error handling here */ ++ ext3_mb_release_desc(&e3b); ++ BUG_ON(bitmap_bh == NULL); ++ } ++ ++ ext3_lock_group(sb, group); ++ list_del_rcu(&pa->pa_group_list); ++ ext3_mb_release_inode_pa(&e3b, bitmap_bh, pa); ++ ext3_unlock_group(sb, group); ++ ++ ext3_mb_release_desc(&e3b); ++ brelse(bitmap_bh); ++ ++ list_del(&pa->u.pa_tmp_list); ++ mb_call_rcu(pa); ++ } ++} ++ ++/* ++ * finds all preallocated spaces and return blocks being freed to them ++ * if preallocated space becomes full (no block is used from the space) ++ * then the function frees space in buddy ++ * XXX: at the moment, truncate (which is the only way to free blocks) ++ * discards all preallocations ++ */ ++void ext3_mb_return_to_preallocation(struct inode *inode, struct ext3_buddy *e3b, ++ sector_t block, int count) ++{ ++ BUG_ON(!list_empty(&EXT3_I(inode)->i_prealloc_list)); ++} ++ ++void ext3_mb_show_ac(struct ext3_allocation_context *ac) ++{ ++#if 0 ++ struct super_block *sb = ac->ac_sb; ++ int i; ++ ++ printk(KERN_ERR "EXT3-fs: can't allocate: status %d flags %d\n", ++ ac->ac_status, ac->ac_flags); ++ printk(KERN_ERR "EXT3-fs: orig %lu/%lu/%lu@%lu, goal %lu/%lu/%lu@%lu, " ++ "best %lu/%lu/%lu@%lu cr %d\n", ++ ac->ac_o_ex.fe_group, ac->ac_o_ex.fe_start, ++ ac->ac_o_ex.fe_len, ac->ac_o_ex.fe_logical, ++ ac->ac_g_ex.fe_group, ac->ac_g_ex.fe_start, ++ ac->ac_g_ex.fe_len, ac->ac_g_ex.fe_logical, ++ ac->ac_b_ex.fe_group, ac->ac_b_ex.fe_start, ++ ac->ac_b_ex.fe_len, ac->ac_b_ex.fe_logical, ++ ac->ac_criteria); ++ printk(KERN_ERR "EXT3-fs: %lu scanned, %d found\n", ac->ac_ex_scanned, ++ ac->ac_found); ++ printk("EXT3-fs: groups: "); ++ for (i = 0; i < EXT3_SB(sb)->s_groups_count; i++) { ++ struct ext3_group_info *grp = EXT3_GROUP_INFO(sb, i); ++ struct ext3_prealloc_space *pa; ++ unsigned long start; ++ struct list_head *cur; ++ list_for_each_rcu(cur, &grp->bb_prealloc_list) { ++ pa = list_entry(cur, struct ext3_prealloc_space, ++ pa_group_list); ++ spin_lock(&pa->pa_lock); ++ ext3_get_group_no_and_offset(sb, pa->pa_pstart, NULL, &start); ++ spin_unlock(&pa->pa_lock); ++ printk("PA:%u:%lu:%u ", i, start, pa->pa_len); ++ } ++ ++ if (grp->bb_free == 0) ++ continue; ++ printk("%d: %d/%d ", i, grp->bb_free, grp->bb_fragments); ++ } ++ printk("\n"); ++ //dump_stack(); ++#endif ++} ++ ++void ext3_mb_group_or_file(struct ext3_allocation_context *ac) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(ac->ac_sb); ++ int bsbits = ac->ac_sb->s_blocksize_bits; ++ loff_t size, isize; ++ ++ if (!(ac->ac_flags & EXT3_MB_HINT_DATA)) ++ return; ++ ++ size = ac->ac_o_ex.fe_logical + ac->ac_o_ex.fe_len; ++ isize = i_size_read(ac->ac_inode) >> bsbits; ++ if (size < isize) ++ size = isize; ++ ++ /* don't use group allocation for large files */ ++ if (size >= sbi->s_mb_stream_request) ++ return; ++ ++ if (unlikely(ac->ac_flags & EXT3_MB_HINT_GOAL_ONLY)) ++ return; ++ ++ BUG_ON(ac->ac_lg != NULL); ++ ac->ac_lg = &sbi->s_locality_groups[smp_processor_id()]; ++ ++ /* we're going to use group allocation */ ++ ac->ac_flags |= EXT3_MB_HINT_GROUP_ALLOC; ++ ++ /* serialize all allocations in the group */ ++ down(&ac->ac_lg->lg_sem); ++} ++ ++int ext3_mb_initialize_context(struct ext3_allocation_context *ac, ++ struct ext3_allocation_request *ar) ++{ ++ struct super_block *sb = ar->inode->i_sb; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ struct ext3_super_block *es = sbi->s_es; ++ unsigned long group, len, goal; ++ unsigned long block; ++ ++ /* we can't allocate > group size */ ++ len = ar->len; ++ if (len >= EXT3_BLOCKS_PER_GROUP(sb) - 10) ++ len = EXT3_BLOCKS_PER_GROUP(sb) - 10; ++ ++ /* start searching from the goal */ ++ goal = ar->goal; ++ if (goal < le32_to_cpu(es->s_first_data_block) || ++ goal >= le32_to_cpu(es->s_blocks_count)) ++ goal = le32_to_cpu(es->s_first_data_block); ++ ext3_get_group_no_and_offset(sb, goal, &group, &block); ++ ++ /* set up allocation goals */ ++ ac->ac_b_ex.fe_logical = ar->logical; ++ ac->ac_b_ex.fe_group = 0; ++ ac->ac_b_ex.fe_start = 0; ++ ac->ac_b_ex.fe_len = 0; ++ ac->ac_status = AC_STATUS_CONTINUE; ++ ac->ac_groups_scanned = 0; ++ ac->ac_ex_scanned = 0; ++ ac->ac_found = 0; ++ ac->ac_sb = sb; ++ ac->ac_inode = ar->inode; ++ ac->ac_o_ex.fe_logical = ar->logical; ++ ac->ac_o_ex.fe_group = group; ++ ac->ac_o_ex.fe_start = block; ++ ac->ac_o_ex.fe_len = len; ++ ac->ac_g_ex.fe_logical = ar->logical; ++ ac->ac_g_ex.fe_group = group; ++ ac->ac_g_ex.fe_start = block; ++ ac->ac_g_ex.fe_len = len; ++ ac->ac_f_ex.fe_len = 0; ++ ac->ac_flags = ar->flags; ++ ac->ac_2order = 0; ++ ac->ac_criteria = 0; ++ ac->ac_pa = NULL; ++ ac->ac_bitmap_page = NULL; ++ ac->ac_buddy_page = NULL; ++ ac->ac_lg = NULL; ++ ++ /* we have to define context: we'll we work with a file or ++ * locality group. this is a policy, actually */ ++ ext3_mb_group_or_file(ac); ++ ++ mb_debug("init ac: %u blocks @ %llu, goal %llu, flags %x, 2^%d, " ++ "left: %llu/%llu, right %llu/%llu to %swritable\n", ++ (unsigned) ar->len, (unsigned) ar->logical, ++ (unsigned) ar->goal, ac->ac_flags, ac->ac_2order, ++ (unsigned) ar->lleft, (unsigned) ar->pleft, ++ (unsigned) ar->lright, (unsigned) ar->pright, ++ atomic_read(&ar->inode->i_writecount) ? "" : "non-"); ++ return 0; ++ ++} ++ ++/* ++ * release all resource we used in allocation ++ */ ++int ext3_mb_release_context(struct ext3_allocation_context *ac) ++{ ++ if (ac->ac_pa) { ++ if (ac->ac_pa->pa_linear) { ++ /* see comment in ext3_mb_use_group_pa() */ ++ spin_lock(&ac->ac_pa->pa_lock); ++ ac->ac_pa->pa_pstart += ac->ac_b_ex.fe_len; ++ ac->ac_pa->pa_lstart += ac->ac_b_ex.fe_len; ++ ac->ac_pa->pa_free -= ac->ac_b_ex.fe_len; ++ ac->ac_pa->pa_len -= ac->ac_b_ex.fe_len; ++ spin_unlock(&ac->ac_pa->pa_lock); ++ } ++ ext3_mb_put_pa(ac, ac->ac_sb, ac->ac_pa); ++ } ++ if (ac->ac_bitmap_page) ++ page_cache_release(ac->ac_bitmap_page); ++ if (ac->ac_buddy_page) ++ page_cache_release(ac->ac_buddy_page); ++ if (ac->ac_flags & EXT3_MB_HINT_GROUP_ALLOC) ++ up(&ac->ac_lg->lg_sem); ++ ext3_mb_collect_stats(ac); ++ return 0; ++} ++ ++int ext3_mb_discard_preallocations(struct super_block *sb, int needed) ++{ ++ int i, ret, freed = 0; ++ ++ for (i = 0; i < EXT3_SB(sb)->s_groups_count && needed > 0; i++) { ++ ret = ext3_mb_discard_group_preallocations(sb, i, needed); ++ freed += ret; ++ needed -= ret; ++ } ++ ++ return freed; ++} ++ ++/* ++ * Main entry point into mballoc to allocate blocks ++ * it tries to use preallocation first, then falls back ++ * to usual allocation ++ */ ++unsigned long ext3_mb_new_blocks(handle_t *handle, ++ struct ext3_allocation_request *ar, int *errp) ++{ ++ struct ext3_allocation_context ac; ++ struct ext3_sb_info *sbi; ++ struct super_block *sb; ++ unsigned long block; ++ int err, freed; ++ ++ sb = ar->inode->i_sb; ++ sbi = EXT3_SB(sb); ++ ++ if (!test_opt(sb, MBALLOC)) { ++ static int ext3_mballoc_warning = 0; ++ if (ext3_mballoc_warning++ == 0) ++ printk(KERN_ERR "EXT3-fs: multiblock request with " ++ "mballoc disabled!\n"); ++ ar->len = 1; ++ err = ext3_new_block_old(handle, ar->inode, ar->goal, errp); ++ return err; ++ } ++ ++ ext3_mb_poll_new_transaction(sb, handle); ++ ++ if ((err = ext3_mb_initialize_context(&ac, ar))) ++ return err; ++ ++ ac.ac_op = EXT3_MB_HISTORY_PREALLOC; ++ if (!ext3_mb_use_preallocated(&ac)) { ++ ++ ac.ac_op = EXT3_MB_HISTORY_ALLOC; ++ ext3_mb_normalize_request(&ac, ar); ++ ++repeat: ++ /* allocate space in core */ ++ ext3_mb_regular_allocator(&ac); ++ ++ /* as we've just preallocated more space than ++ * user requested orinally, we store allocated ++ * space in a special descriptor */ ++ if (ac.ac_status == AC_STATUS_FOUND && ++ ac.ac_o_ex.fe_len < ac.ac_b_ex.fe_len) ++ ext3_mb_new_preallocation(&ac); ++ } ++ ++ if (likely(ac.ac_status == AC_STATUS_FOUND)) { ++ ext3_mb_mark_diskspace_used(&ac, handle); ++ *errp = 0; ++ block = ext3_grp_offs_to_block(sb, &ac.ac_b_ex); ++ ar->len = ac.ac_b_ex.fe_len; ++ } else { ++ freed = ext3_mb_discard_preallocations(sb, ac.ac_o_ex.fe_len); ++ if (freed) ++ goto repeat; ++ *errp = -ENOSPC; ++ ac.ac_b_ex.fe_len = 0; ++ block = 0; ++ ext3_mb_show_ac(&ac); ++ } ++ ++ ext3_mb_release_context(&ac); ++ ++ return block; ++} ++EXPORT_SYMBOL(ext3_mb_new_blocks); ++ ++int ext3_new_block(handle_t *handle, struct inode *inode, ++ unsigned long goal, int *errp) ++{ ++ struct ext3_allocation_request ar; ++ unsigned long ret; ++ ++ if (!test_opt(inode->i_sb, MBALLOC)) { ++ ret = ext3_new_block_old(handle, inode, goal, errp); ++ return ret; ++ } ++ ++ ar.inode = inode; ++ ar.goal = goal; ++ ar.len = 1; ++ ar.logical = 0; ++ ar.lleft = 0; ++ ar.pleft = 0; ++ ar.lright = 0; ++ ar.pright = 0; ++ ar.flags = 0; ++ ret = ext3_mb_new_blocks(handle, &ar, errp); ++ return ret; ++} ++ ++void ext3_mb_poll_new_transaction(struct super_block *sb, handle_t *handle) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ ++ if (sbi->s_last_transaction == handle->h_transaction->t_tid) ++ return; ++ ++ /* new transaction! time to close last one and free blocks for ++ * committed transaction. we know that only transaction can be ++ * active, so previos transaction can be being logged and we ++ * know that transaction before previous is known to be already ++ * logged. this means that now we may free blocks freed in all ++ * transactions before previous one. hope I'm clear enough ... */ ++ ++ spin_lock(&sbi->s_md_lock); ++ if (sbi->s_last_transaction != handle->h_transaction->t_tid) { ++ mb_debug("new transaction %lu, old %lu\n", ++ (unsigned long) handle->h_transaction->t_tid, ++ (unsigned long) sbi->s_last_transaction); ++ list_splice_init(&sbi->s_closed_transaction, ++ &sbi->s_committed_transaction); ++ list_splice_init(&sbi->s_active_transaction, ++ &sbi->s_closed_transaction); ++ sbi->s_last_transaction = handle->h_transaction->t_tid; ++ } ++ spin_unlock(&sbi->s_md_lock); ++ ++ ext3_mb_free_committed_blocks(sb); ++} ++ ++int ext3_mb_free_metadata(handle_t *handle, struct ext3_buddy *e3b, ++ int group, int block, int count) ++{ ++ struct ext3_group_info *db = e3b->bd_info; ++ struct super_block *sb = e3b->bd_sb; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ struct ext3_free_metadata *md; ++ int i; ++ ++ BUG_ON(e3b->bd_bitmap_page == NULL); ++ BUG_ON(e3b->bd_buddy_page == NULL); ++ ++ ext3_lock_group(sb, group); ++ for (i = 0; i < count; i++) { ++ md = db->bb_md_cur; ++ if (md && db->bb_tid != handle->h_transaction->t_tid) { ++ db->bb_md_cur = NULL; ++ md = NULL; ++ } ++ ++ if (md == NULL) { ++ ext3_unlock_group(sb, group); ++ md = kmalloc(sizeof(*md), GFP_KERNEL); ++ if (md == NULL) ++ return -ENOMEM; ++ md->num = 0; ++ md->group = group; ++ ++ ext3_lock_group(sb, group); ++ if (db->bb_md_cur == NULL) { ++ spin_lock(&sbi->s_md_lock); ++ list_add(&md->list, &sbi->s_active_transaction); ++ spin_unlock(&sbi->s_md_lock); ++ /* protect buddy cache from being freed, ++ * otherwise we'll refresh it from ++ * on-disk bitmap and lose not-yet-available ++ * blocks */ ++ page_cache_get(e3b->bd_buddy_page); ++ page_cache_get(e3b->bd_bitmap_page); ++ db->bb_md_cur = md; ++ db->bb_tid = handle->h_transaction->t_tid; ++ mb_debug("new md 0x%p for group %u\n", ++ md, md->group); ++ } else { ++ kfree(md); ++ md = db->bb_md_cur; ++ } ++ } ++ ++ BUG_ON(md->num >= EXT3_BB_MAX_BLOCKS); ++ md->blocks[md->num] = block + i; ++ md->num++; ++ if (md->num == EXT3_BB_MAX_BLOCKS) { ++ /* no more space, put full container on a sb's list */ ++ db->bb_md_cur = NULL; ++ } ++ } ++ ext3_unlock_group(sb, group); ++ return 0; ++} ++ ++/* ++ * Main entry point into mballoc to free blocks ++ */ ++void ext3_mb_free_blocks(handle_t *handle, struct inode *inode, ++ unsigned long block, unsigned long count, ++ int metadata, int *freed) ++{ ++ struct buffer_head *bitmap_bh = NULL; ++ struct super_block *sb = inode->i_sb; ++ struct ext3_allocation_context ac; ++ struct ext3_group_desc *gdp; ++ struct ext3_super_block *es; ++ unsigned long bit, overflow; ++ struct buffer_head *gd_bh; ++ unsigned long block_group; ++ struct ext3_sb_info *sbi; ++ struct ext3_buddy e3b; ++ int err = 0, ret; ++ ++ *freed = 0; ++ ++ ext3_mb_poll_new_transaction(sb, handle); ++ ++ sbi = EXT3_SB(sb); ++ es = EXT3_SB(sb)->s_es; ++ if (block < le32_to_cpu(es->s_first_data_block) || ++ block + count < block || ++ block + count > le32_to_cpu(es->s_blocks_count)) { ++ ext3_error (sb, __FUNCTION__, ++ "Freeing blocks not in datazone - " ++ "block = %lu, count = %lu", block, count); ++ goto error_return; ++ } ++ ++ ext3_debug("freeing block %lu\n", block); ++ ++ ac.ac_op = EXT3_MB_HISTORY_FREE; ++ ac.ac_inode = inode; ++ ac.ac_sb = sb; ++ ++do_more: ++ overflow = 0; ++ ext3_get_group_no_and_offset(sb, block, &block_group, &bit); ++ ++ /* ++ * Check to see if we are freeing blocks across a group ++ * boundary. ++ */ ++ if (bit + count > EXT3_BLOCKS_PER_GROUP(sb)) { ++ overflow = bit + count - EXT3_BLOCKS_PER_GROUP(sb); ++ count -= overflow; ++ } ++ brelse(bitmap_bh); ++ bitmap_bh = read_block_bitmap(sb, block_group); ++ if (!bitmap_bh) ++ goto error_return; ++ gdp = ext3_get_group_desc (sb, block_group, &gd_bh); ++ if (!gdp) ++ goto error_return; ++ ++ if (in_range (le32_to_cpu(gdp->bg_block_bitmap), block, count) || ++ in_range (le32_to_cpu(gdp->bg_inode_bitmap), block, count) || ++ in_range (block, le32_to_cpu(gdp->bg_inode_table), ++ EXT3_SB(sb)->s_itb_per_group) || ++ in_range (block + count - 1, le32_to_cpu(gdp->bg_inode_table), ++ EXT3_SB(sb)->s_itb_per_group)) ++ ext3_error(sb, __FUNCTION__, ++ "Freeing blocks in system zone - " ++ "Block = %lu, count = %lu", block, count); ++ ++ BUFFER_TRACE(bitmap_bh, "getting write access"); ++ err = ext3_journal_get_write_access(handle, bitmap_bh); ++ if (err) ++ goto error_return; ++ ++ /* ++ * We are about to modify some metadata. Call the journal APIs ++ * to unshare ->b_data if a currently-committing transaction is ++ * using it ++ */ ++ BUFFER_TRACE(gd_bh, "get_write_access"); ++ err = ext3_journal_get_write_access(handle, gd_bh); ++ if (err) ++ goto error_return; ++ ++ err = ext3_mb_load_buddy(sb, block_group, &e3b); ++ if (err) ++ goto error_return; ++ ++#ifdef AGGRESSIVE_CHECK ++ { ++ int i; ++ for (i = 0; i < count; i++) ++ BUG_ON(!mb_test_bit(bit + i, bitmap_bh->b_data)); ++ } ++#endif ++ mb_clear_bits(bitmap_bh->b_data, bit, count); ++ ++ /* We dirtied the bitmap block */ ++ BUFFER_TRACE(bitmap_bh, "dirtied bitmap block"); ++ err = ext3_journal_dirty_metadata(handle, bitmap_bh); ++ ++ ac.ac_b_ex.fe_group = block_group; ++ ac.ac_b_ex.fe_start = bit; ++ ac.ac_b_ex.fe_len = count; ++ ext3_mb_store_history(&ac); ++ ++ if (metadata) { ++ /* blocks being freed are metadata. these blocks shouldn't ++ * be used until this transaction is committed */ ++ ext3_mb_free_metadata(handle, &e3b, block_group, bit, count); ++ } else { ++ ext3_lock_group(sb, block_group); ++ err = mb_free_blocks(&e3b, bit, count); ++ ext3_mb_return_to_preallocation(inode, &e3b, block, count); ++ ext3_unlock_group(sb, block_group); ++ BUG_ON(err != 0); ++ } ++ ++ spin_lock(sb_bgl_lock(sbi, block_group)); ++ gdp->bg_free_blocks_count = ++ cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) + count); ++ spin_unlock(sb_bgl_lock(sbi, block_group)); ++ percpu_counter_mod(&sbi->s_freeblocks_counter, count); ++ ++ ext3_mb_release_desc(&e3b); ++ ++ *freed += count; ++ ++ /* And the group descriptor block */ ++ BUFFER_TRACE(gd_bh, "dirtied group descriptor block"); ++ ret = ext3_journal_dirty_metadata(handle, gd_bh); ++ if (!err) err = ret; ++ ++ if (overflow && !err) { ++ block += count; ++ count = overflow; ++ goto do_more; ++ } ++ sb->s_dirt = 1; ++error_return: ++ brelse(bitmap_bh); ++ ext3_std_error(sb, err); ++ return; ++} diff --git a/ldiskfs/kernel_patches/patches/ext3-mballoc3-rhel4.patch b/ldiskfs/kernel_patches/patches/ext3-mballoc3-rhel4.patch new file mode 100644 index 0000000000000000000000000000000000000000..910df7c0d0e71789ec9b1e2237a0ed41b018bf5c --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-mballoc3-rhel4.patch @@ -0,0 +1,396 @@ +Index: linux-2.6.9-full/include/linux/ext3_fs_i.h +=================================================================== +--- linux-2.6.9-full.orig/include/linux/ext3_fs_i.h 2007-03-28 01:29:38.000000000 +0400 ++++ linux-2.6.9-full/include/linux/ext3_fs_i.h 2007-03-28 15:45:41.000000000 +0400 +@@ -130,6 +130,10 @@ struct ext3_inode_info { + struct inode vfs_inode; + + __u32 i_cached_extent[4]; ++ ++ /* mballoc */ ++ struct list_head i_prealloc_list; ++ spinlock_t i_prealloc_lock; + }; + + #endif /* _LINUX_EXT3_FS_I */ +Index: linux-2.6.9-full/include/linux/ext3_fs_sb.h +=================================================================== +--- linux-2.6.9-full.orig/include/linux/ext3_fs_sb.h 2007-03-28 15:42:16.000000000 +0400 ++++ linux-2.6.9-full/include/linux/ext3_fs_sb.h 2007-03-28 15:45:41.000000000 +0400 +@@ -23,9 +23,16 @@ + #define EXT_INCLUDE + #include <linux/blockgroup_lock.h> + #include <linux/percpu_counter.h> ++#include <linux/list.h> + #endif + #endif + #include <linux/rbtree.h> ++#include <linux/proc_fs.h> ++ ++struct ext3_buddy_group_blocks; ++struct ext3_locality_group; ++struct ext3_mb_history; ++#define EXT3_BB_MAX_BLOCKS + + /* + * third extended-fs super-block data in memory +Index: linux-2.6.9-full/include/linux/ext3_fs.h +=================================================================== +--- linux-2.6.9-full.orig/include/linux/ext3_fs.h 2007-03-28 15:45:07.000000000 +0400 ++++ linux-2.6.9-full/include/linux/ext3_fs.h 2007-03-28 15:45:41.000000000 +0400 +@@ -389,6 +389,7 @@ struct ext3_inode { + #define EXT3_MOUNT_IOPEN_NOPRIV 0x100000/* Make iopen world-readable */ + #define EXT3_MOUNT_EXTENTS 0x200000/* Extents support */ + #define EXT3_MOUNT_EXTDEBUG 0x400000/* Extents debug */ ++#define EXT3_MOUNT_MBALLOC 0x800000/* Buddy allocation support */ + + /* Compatibility, for having both ext2_fs.h and ext3_fs.h included at once */ + #ifndef clear_opt +@@ -749,8 +750,9 @@ struct dir_private_info { + extern int ext3_bg_has_super(struct super_block *sb, int group); + extern unsigned long ext3_bg_num_gdb(struct super_block *sb, int group); + extern int ext3_new_block (handle_t *, struct inode *, unsigned long, int *); ++extern int ext3_new_block_old (handle_t *, struct inode *, unsigned long, int *); + extern void ext3_free_blocks (handle_t *, struct inode *, unsigned long, +- unsigned long); ++ unsigned long, int); + extern void ext3_free_blocks_sb (handle_t *, struct super_block *, + unsigned long, unsigned long, int *); + extern unsigned long ext3_count_free_blocks (struct super_block *); +Index: linux-2.6.9-full/fs/ext3/super.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/super.c 2007-03-28 15:42:16.000000000 +0400 ++++ linux-2.6.9-full/fs/ext3/super.c 2007-03-28 15:45:41.000000000 +0400 +@@ -600,6 +600,7 @@ enum { + Opt_ignore, Opt_barrier, Opt_err, Opt_resize, + Opt_iopen, Opt_noiopen, Opt_iopen_nopriv, + Opt_extents, Opt_noextents, Opt_extdebug, ++ Opt_mballoc, Opt_nomballoc, Opt_stripe, + }; + + static match_table_t tokens = { +@@ -653,6 +654,9 @@ static match_table_t tokens = { + {Opt_noextents, "noextents"}, + {Opt_extdebug, "extdebug"}, + {Opt_barrier, "barrier=%u"}, ++ {Opt_mballoc, "mballoc"}, ++ {Opt_nomballoc, "nomballoc"}, ++ {Opt_stripe, "stripe=%u"}, + {Opt_err, NULL}, + {Opt_resize, "resize"}, + }; +@@ -965,6 +969,19 @@ clear_qf_name: + case Opt_extdebug: + set_opt (sbi->s_mount_opt, EXTDEBUG); + break; ++ case Opt_mballoc: ++ set_opt(sbi->s_mount_opt, MBALLOC); ++ break; ++ case Opt_nomballoc: ++ clear_opt(sbi->s_mount_opt, MBALLOC); ++ break; ++ case Opt_stripe: ++ if (match_int(&args[0], &option)) ++ return 0; ++ if (option < 0) ++ return 0; ++ sbi->s_stripe = option; ++ break; + default: + printk (KERN_ERR + "EXT3-fs: Unrecognized mount option \"%s\" " +@@ -1654,6 +1671,7 @@ static int ext3_fill_super (struct super + ext3_count_dirs(sb)); + + ext3_ext_init(sb); ++ ext3_mb_init(sb, needs_recovery); + + return 0; + +Index: linux-2.6.9-full/fs/ext3/extents.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/extents.c 2007-03-28 01:29:41.000000000 +0400 ++++ linux-2.6.9-full/fs/ext3/extents.c 2007-03-28 15:45:41.000000000 +0400 +@@ -779,7 +779,7 @@ cleanup: + for (i = 0; i < depth; i++) { + if (!ablocks[i]) + continue; +- ext3_free_blocks(handle, tree->inode, ablocks[i], 1); ++ ext3_free_blocks(handle, tree->inode, ablocks[i], 1, 1); + } + } + kfree(ablocks); +@@ -1586,7 +1586,7 @@ int ext3_ext_rm_idx(handle_t *handle, st + path->p_idx->ei_leaf); + bh = sb_find_get_block(tree->inode->i_sb, path->p_idx->ei_leaf); + ext3_forget(handle, 1, tree->inode, bh, path->p_idx->ei_leaf); +- ext3_free_blocks(handle, tree->inode, path->p_idx->ei_leaf, 1); ++ ext3_free_blocks(handle, tree->inode, path->p_idx->ei_leaf, 1, 1); + return err; + } + +@@ -2071,10 +2071,12 @@ ext3_remove_blocks(struct ext3_extents_t + int needed = ext3_remove_blocks_credits(tree, ex, from, to); + handle_t *handle = ext3_journal_start(tree->inode, needed); + struct buffer_head *bh; +- int i; ++ int i, metadata = 0; + + if (IS_ERR(handle)) + return PTR_ERR(handle); ++ if (S_ISDIR(tree->inode->i_mode) || S_ISLNK(tree->inode->i_mode)) ++ metadata = 1; + if (from >= ex->ee_block && to == ex->ee_block + ex->ee_len - 1) { + /* tail removal */ + unsigned long num, start; +@@ -2086,7 +2088,7 @@ ext3_remove_blocks(struct ext3_extents_t + bh = sb_find_get_block(tree->inode->i_sb, start + i); + ext3_forget(handle, 0, tree->inode, bh, start + i); + } +- ext3_free_blocks(handle, tree->inode, start, num); ++ ext3_free_blocks(handle, tree->inode, start, num, metadata); + } else if (from == ex->ee_block && to <= ex->ee_block + ex->ee_len - 1) { + printk("strange request: removal %lu-%lu from %u:%u\n", + from, to, ex->ee_block, ex->ee_len); +@@ -2177,11 +2179,8 @@ int ext3_ext_get_block(handle_t *handle, + struct ext3_extent *ex; + int goal, newblock, err = 0, depth; + struct ext3_extents_tree tree; +- unsigned long next; +- int allocated = 0; +- +- /* until we have multiblock allocation */ +- max_blocks = 1; ++ unsigned long allocated = 0; ++ struct ext3_allocation_request ar; + + clear_buffer_new(bh_result); + ext3_init_tree_desc(&tree, inode); +@@ -2253,18 +2252,33 @@ int ext3_ext_get_block(handle_t *handle, + goto out2; + } + ++ /* find neighbour allocated blocks */ ++ ar.lleft = iblock; ++ err = ext3_ext_search_left(&tree, path, &ar.lleft, &ar.pleft); ++ if (err) ++ goto out2; ++ ar.lright = iblock; ++ err = ext3_ext_search_right(&tree, path, &ar.lright, &ar.pright); ++ if (err) ++ goto out2; ++ + /* find next allocated block so that we know how many + * blocks we can allocate without ovelapping next extent */ +- EXT_ASSERT(iblock >= ex->ee_block + ex->ee_len); +- next = ext3_ext_next_allocated_block(path); +- EXT_ASSERT(next > iblock); +- allocated = next - iblock; ++ EXT_ASSERT(ar.pright == 0 || ar.lright > iblock); ++ if (ar.pright == 0) ++ allocated = EXT_MAX_BLOCK - iblock; ++ else ++ allocated = ar.lright - iblock; + if (allocated > max_blocks) + allocated = max_blocks; + + /* allocate new block */ +- goal = ext3_ext_find_goal(inode, path, iblock); +- newblock = ext3_new_block(handle, inode, goal, &err); ++ ar.inode = inode; ++ ar.goal = ext3_ext_find_goal(inode, path, iblock); ++ ar.logical = iblock; ++ ar.len = allocated; ++ ar.flags = EXT3_MB_HINT_DATA; ++ newblock = ext3_mb_new_blocks(handle, &ar, &err); + if (!newblock) + goto out2; + ext_debug(&tree, "allocate new block: goal %d, found %d\n", +@@ -2274,11 +2288,14 @@ int ext3_ext_get_block(handle_t *handle, + newex.ee_block = iblock; + newex.ee_start = newblock; + newex.ee_start_hi = 0; +- newex.ee_len = 1; ++ newex.ee_len = ar.len; + err = ext3_ext_insert_extent(handle, &tree, path, &newex); + if (err) { + /* free data blocks we just allocated */ +- ext3_free_blocks(handle, inode, newex.ee_start, newex.ee_len); ++ /* not a good idea to call discard here directly, ++ * but otherwise we'd need to call it every free() */ ++ ext3_mb_discard_inode_preallocations(inode); ++ ext3_free_blocks(handle, inode, newex.ee_start, newex.ee_len, 0); + goto out2; + } + +@@ -2287,6 +2304,7 @@ int ext3_ext_get_block(handle_t *handle, + + /* previous routine could use block we allocated */ + newblock = newex.ee_start; ++ allocated = newex.ee_len; + set_buffer_new(bh_result); + + ext3_ext_put_in_cache(&tree, newex.ee_block, newex.ee_len, +@@ -2339,6 +2357,9 @@ void ext3_ext_truncate(struct inode * in + down(&EXT3_I(inode)->truncate_sem); + ext3_ext_invalidate_cache(&tree); + ++ /* it's important to discard preallocations under truncate_sem */ ++ ext3_mb_discard_inode_preallocations(inode); ++ + /* + * TODO: optimization is possible here + * probably we need not scaning at all, +Index: linux-2.6.9-full/fs/ext3/Makefile +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/Makefile 2007-03-28 01:29:38.000000000 +0400 ++++ linux-2.6.9-full/fs/ext3/Makefile 2007-03-28 15:45:41.000000000 +0400 +@@ -6,7 +6,7 @@ obj-$(CONFIG_EXT3_FS) += ext3.o + + ext3-y := balloc.o bitmap.o dir.o file.o fsync.o ialloc.o inode.o iopen.o \ + ioctl.o namei.o super.o symlink.o hash.o resize.o \ +- extents.o ++ extents.o mballoc.o + + ext3-$(CONFIG_EXT3_FS_XATTR) += xattr.o xattr_user.o xattr_trusted.o + ext3-$(CONFIG_EXT3_FS_POSIX_ACL) += acl.o +Index: linux-2.6.9-full/fs/ext3/xattr.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/xattr.c 2006-05-18 23:57:04.000000000 +0400 ++++ linux-2.6.9-full/fs/ext3/xattr.c 2007-03-28 15:45:41.000000000 +0400 +@@ -1281,7 +1281,7 @@ ext3_xattr_set_handle2(handle_t *handle, + new_bh = sb_getblk(sb, block); + if (!new_bh) { + getblk_failed: +- ext3_free_blocks(handle, inode, block, 1); ++ ext3_free_blocks(handle, inode, block, 1, 1); + error = -EIO; + goto cleanup; + } +@@ -1328,7 +1328,7 @@ getblk_failed: + if (ce) + mb_cache_entry_free(ce); + ea_bdebug(old_bh, "freeing"); +- ext3_free_blocks(handle, inode, old_bh->b_blocknr, 1); ++ ext3_free_blocks(handle, inode, old_bh->b_blocknr, 1, 1); + + /* ext3_forget() calls bforget() for us, but we + let our caller release old_bh, so we need to +@@ -1427,7 +1427,7 @@ ext3_xattr_delete_inode(handle_t *handle + if (HDR(bh)->h_refcount == cpu_to_le32(1)) { + if (ce) + mb_cache_entry_free(ce); +- ext3_free_blocks(handle, inode, EXT3_I(inode)->i_file_acl, 1); ++ ext3_free_blocks(handle, inode, EXT3_I(inode)->i_file_acl, 1, 1); + get_bh(bh); + ext3_forget(handle, 1, inode, bh, EXT3_I(inode)->i_file_acl); + } else { +Index: linux-2.6.9-full/fs/ext3/balloc.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/balloc.c 2006-03-10 18:20:03.000000000 +0300 ++++ linux-2.6.9-full/fs/ext3/balloc.c 2007-03-28 15:45:41.000000000 +0400 +@@ -79,7 +79,7 @@ struct ext3_group_desc * ext3_get_group_ + * + * Return buffer_head on success or NULL in case of failure. + */ +-static struct buffer_head * ++struct buffer_head * + read_block_bitmap(struct super_block *sb, unsigned int block_group) + { + struct ext3_group_desc * desc; +@@ -267,6 +267,8 @@ void ext3_discard_reservation(struct ino + struct reserve_window_node *rsv = &ei->i_rsv_window; + spinlock_t *rsv_lock = &EXT3_SB(inode->i_sb)->s_rsv_window_lock; + ++ ext3_mb_discard_inode_preallocations(inode); ++ + if (!rsv_is_empty(&rsv->rsv_window)) { + spin_lock(rsv_lock); + if (!rsv_is_empty(&rsv->rsv_window)) +@@ -451,21 +453,25 @@ error_return: + return; + } + +-/* Free given blocks, update quota and i_blocks field */ +-void ext3_free_blocks(handle_t *handle, struct inode *inode, +- unsigned long block, unsigned long count) ++void ext3_free_blocks(handle_t *handle, struct inode * inode, ++ unsigned long block, unsigned long count, int metadata) + { +- struct super_block * sb; +- int dquot_freed_blocks; ++ struct super_block *sb; ++ int freed; ++ ++ /* this isn't the right place to decide whether block is metadata ++ * inode.c/extents.c knows better, but for safety ... */ ++ if (S_ISDIR(inode->i_mode) || S_ISLNK(inode->i_mode) || ++ ext3_should_journal_data(inode)) ++ metadata = 1; + + sb = inode->i_sb; +- if (!sb) { +- printk ("ext3_free_blocks: nonexistent device"); +- return; +- } +- ext3_free_blocks_sb(handle, sb, block, count, &dquot_freed_blocks); +- if (dquot_freed_blocks) +- DQUOT_FREE_BLOCK(inode, dquot_freed_blocks); ++ if (!test_opt(sb, MBALLOC) || !EXT3_SB(sb)->s_group_info) ++ ext3_free_blocks_sb(handle, sb, block, count, &freed); ++ else ++ ext3_mb_free_blocks(handle, inode, block, count, metadata, &freed); ++ if (freed) ++ DQUOT_FREE_BLOCK(inode, freed); + return; + } + +@@ -1131,7 +1137,7 @@ int ext3_should_retry_alloc(struct super + * bitmap, and then for any free bit if that fails. + * This function also updates quota and i_blocks field. + */ +-int ext3_new_block(handle_t *handle, struct inode *inode, ++int ext3_new_block_old(handle_t *handle, struct inode *inode, + unsigned long goal, int *errp) + { + struct buffer_head *bitmap_bh = NULL; +Index: linux-2.6.9-full/fs/ext3/inode.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/inode.c 2007-03-28 01:29:39.000000000 +0400 ++++ linux-2.6.9-full/fs/ext3/inode.c 2007-03-28 15:45:41.000000000 +0400 +@@ -572,7 +572,7 @@ static int ext3_alloc_branch(handle_t *h + ext3_journal_forget(handle, branch[i].bh); + } + for (i = 0; i < keys; i++) +- ext3_free_blocks(handle, inode, le32_to_cpu(branch[i].key), 1); ++ ext3_free_blocks(handle, inode, le32_to_cpu(branch[i].key), 1, 0); + return err; + } + +@@ -673,7 +673,7 @@ err_out: + if (err == -EAGAIN) + for (i = 0; i < num; i++) + ext3_free_blocks(handle, inode, +- le32_to_cpu(where[i].key), 1); ++ le32_to_cpu(where[i].key), 1, 0); + return err; + } + +@@ -1834,7 +1834,7 @@ ext3_clear_blocks(handle_t *handle, stru + } + } + +- ext3_free_blocks(handle, inode, block_to_free, count); ++ ext3_free_blocks(handle, inode, block_to_free, count, 0); + } + + /** +@@ -2007,7 +2007,7 @@ static void ext3_free_branches(handle_t + ext3_journal_test_restart(handle, inode); + } + +- ext3_free_blocks(handle, inode, nr, 1); ++ ext3_free_blocks(handle, inode, nr, 1, 1); + + if (parent_bh) { + /* diff --git a/ldiskfs/kernel_patches/patches/ext3-mballoc3-sles10.patch b/ldiskfs/kernel_patches/patches/ext3-mballoc3-sles10.patch new file mode 100644 index 0000000000000000000000000000000000000000..373f0c6680ec510a3f3035549707549d1e640140 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-mballoc3-sles10.patch @@ -0,0 +1,377 @@ +Index: linux-2.6.16.27-0.9-full/include/linux/ext3_fs_i.h +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/include/linux/ext3_fs_i.h 2007-03-28 05:12:50.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/include/linux/ext3_fs_i.h 2007-03-28 16:03:20.000000000 +0400 +@@ -135,6 +135,10 @@ struct ext3_inode_info { + struct inode vfs_inode; + + __u32 i_cached_extent[4]; ++ ++ /* mballoc */ ++ struct list_head i_prealloc_list; ++ spinlock_t i_prealloc_lock; + }; + + #endif /* _LINUX_EXT3_FS_I */ +Index: linux-2.6.16.27-0.9-full/include/linux/ext3_fs_sb.h +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/include/linux/ext3_fs_sb.h 2007-03-28 16:03:19.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/include/linux/ext3_fs_sb.h 2007-03-28 16:03:20.000000000 +0400 +@@ -21,8 +21,15 @@ + #include <linux/wait.h> + #include <linux/blockgroup_lock.h> + #include <linux/percpu_counter.h> ++#include <linux/list.h> + #endif + #include <linux/rbtree.h> ++#include <linux/proc_fs.h> ++ ++struct ext3_buddy_group_blocks; ++struct ext3_locality_group; ++struct ext3_mb_history; ++#define EXT3_BB_MAX_BLOCKS + + /* + * third extended-fs super-block data in memory +Index: linux-2.6.16.27-0.9-full/include/linux/ext3_fs.h +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/include/linux/ext3_fs.h 2007-03-28 16:03:19.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/include/linux/ext3_fs.h 2007-03-28 16:03:20.000000000 +0400 +@@ -407,6 +407,7 @@ struct ext3_inode { + #define EXT3_MOUNT_IOPEN_NOPRIV 0x800000/* Make iopen world-readable */ + #define EXT3_MOUNT_EXTENTS 0x1000000/* Extents support */ + #define EXT3_MOUNT_EXTDEBUG 0x2000000/* Extents debug */ ++#define EXT3_MOUNT_MBALLOC 0x4000000/* Buddy allocation support */ + + /* Compatibility, for having both ext2_fs.h and ext3_fs.h included at once */ + #ifndef clear_opt +@@ -767,8 +768,9 @@ struct dir_private_info { + extern int ext3_bg_has_super(struct super_block *sb, int group); + extern unsigned long ext3_bg_num_gdb(struct super_block *sb, int group); + extern int ext3_new_block (handle_t *, struct inode *, unsigned long, int *); ++extern int ext3_new_block_old (handle_t *, struct inode *, unsigned long, int *); + extern void ext3_free_blocks (handle_t *, struct inode *, unsigned long, +- unsigned long); ++ unsigned long, int); + extern void ext3_free_blocks_sb (handle_t *, struct super_block *, + unsigned long, unsigned long, int *); + extern unsigned long ext3_count_free_blocks (struct super_block *); +Index: linux-2.6.16.27-0.9-full/fs/ext3/super.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/super.c 2007-03-28 16:03:19.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/fs/ext3/super.c 2007-03-28 16:03:20.000000000 +0400 +@@ -688,6 +688,7 @@ enum { + Opt_ignore, Opt_barrier, Opt_err, Opt_resize, Opt_usrquota, + Opt_iopen, Opt_noiopen, Opt_iopen_nopriv, + Opt_extents, Opt_noextents, Opt_extdebug, ++ Opt_mballoc, Opt_nomballoc, Opt_stripe, + Opt_grpquota + }; + +@@ -743,6 +744,9 @@ static match_table_t tokens = { + {Opt_noextents, "noextents"}, + {Opt_extdebug, "extdebug"}, + {Opt_barrier, "barrier=%u"}, ++ {Opt_mballoc, "mballoc"}, ++ {Opt_nomballoc, "nomballoc"}, ++ {Opt_stripe, "stripe=%u"}, + {Opt_err, NULL}, + {Opt_resize, "resize"}, + }; +@@ -1092,6 +1096,19 @@ clear_qf_name: + case Opt_extdebug: + set_opt (sbi->s_mount_opt, EXTDEBUG); + break; ++ case Opt_mballoc: ++ set_opt(sbi->s_mount_opt, MBALLOC); ++ break; ++ case Opt_nomballoc: ++ clear_opt(sbi->s_mount_opt, MBALLOC); ++ break; ++ case Opt_stripe: ++ if (match_int(&args[0], &option)) ++ return 0; ++ if (option < 0) ++ return 0; ++ sbi->s_stripe = option; ++ break; + default: + printk (KERN_ERR + "EXT3-fs: Unrecognized mount option \"%s\" " +@@ -1819,6 +1836,7 @@ static int ext3_fill_super (struct super + ext3_count_dirs(sb)); + + ext3_ext_init(sb); ++ ext3_mb_init(sb, needs_recovery); + lock_kernel(); + return 0; + +Index: linux-2.6.16.27-0.9-full/fs/ext3/extents.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/extents.c 2007-03-28 05:13:39.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/fs/ext3/extents.c 2007-03-28 16:03:20.000000000 +0400 +@@ -779,7 +779,7 @@ cleanup: + for (i = 0; i < depth; i++) { + if (!ablocks[i]) + continue; +- ext3_free_blocks(handle, tree->inode, ablocks[i], 1); ++ ext3_free_blocks(handle, tree->inode, ablocks[i], 1, 1); + } + } + kfree(ablocks); +@@ -1586,7 +1586,7 @@ int ext3_ext_rm_idx(handle_t *handle, st + path->p_idx->ei_leaf); + bh = sb_find_get_block(tree->inode->i_sb, path->p_idx->ei_leaf); + ext3_forget(handle, 1, tree->inode, bh, path->p_idx->ei_leaf); +- ext3_free_blocks(handle, tree->inode, path->p_idx->ei_leaf, 1); ++ ext3_free_blocks(handle, tree->inode, path->p_idx->ei_leaf, 1, 1); + return err; + } + +@@ -2071,10 +2071,12 @@ ext3_remove_blocks(struct ext3_extents_t + int needed = ext3_remove_blocks_credits(tree, ex, from, to); + handle_t *handle = ext3_journal_start(tree->inode, needed); + struct buffer_head *bh; +- int i; ++ int i, metadata = 0; + + if (IS_ERR(handle)) + return PTR_ERR(handle); ++ if (S_ISDIR(tree->inode->i_mode) || S_ISLNK(tree->inode->i_mode)) ++ metadata = 1; + if (from >= ex->ee_block && to == ex->ee_block + ex->ee_len - 1) { + /* tail removal */ + unsigned long num, start; +@@ -2086,7 +2088,7 @@ ext3_remove_blocks(struct ext3_extents_t + bh = sb_find_get_block(tree->inode->i_sb, start + i); + ext3_forget(handle, 0, tree->inode, bh, start + i); + } +- ext3_free_blocks(handle, tree->inode, start, num); ++ ext3_free_blocks(handle, tree->inode, start, num, metadata); + } else if (from == ex->ee_block && to <= ex->ee_block + ex->ee_len - 1) { + printk("strange request: removal %lu-%lu from %u:%u\n", + from, to, ex->ee_block, ex->ee_len); +@@ -2177,11 +2179,8 @@ int ext3_ext_get_block(handle_t *handle, + struct ext3_extent *ex; + int goal, newblock, err = 0, depth; + struct ext3_extents_tree tree; +- unsigned long next; +- int allocated = 0; +- +- /* until we have multiblock allocation */ +- max_blocks = 1; ++ unsigned long allocated = 0; ++ struct ext3_allocation_request ar; + + clear_buffer_new(bh_result); + ext3_init_tree_desc(&tree, inode); +@@ -2253,18 +2252,33 @@ int ext3_ext_get_block(handle_t *handle, + goto out2; + } + ++ /* find neighbour allocated blocks */ ++ ar.lleft = iblock; ++ err = ext3_ext_search_left(&tree, path, &ar.lleft, &ar.pleft); ++ if (err) ++ goto out2; ++ ar.lright = iblock; ++ err = ext3_ext_search_right(&tree, path, &ar.lright, &ar.pright); ++ if (err) ++ goto out2; ++ + /* find next allocated block so that we know how many + * blocks we can allocate without ovelapping next extent */ +- EXT_ASSERT(iblock >= ex->ee_block + ex->ee_len); +- next = ext3_ext_next_allocated_block(path); +- EXT_ASSERT(next > iblock); +- allocated = next - iblock; ++ EXT_ASSERT(ar.pright == 0 || ar.lright > iblock); ++ if (ar.pright == 0) ++ allocated = EXT_MAX_BLOCK - iblock; ++ else ++ allocated = ar.lright - iblock; + if (allocated > max_blocks) + allocated = max_blocks; + + /* allocate new block */ +- goal = ext3_ext_find_goal(inode, path, iblock); +- newblock = ext3_new_block(handle, inode, goal, &err); ++ ar.inode = inode; ++ ar.goal = ext3_ext_find_goal(inode, path, iblock); ++ ar.logical = iblock; ++ ar.len = allocated; ++ ar.flags = EXT3_MB_HINT_DATA; ++ newblock = ext3_mb_new_blocks(handle, &ar, &err); + if (!newblock) + goto out2; + ext_debug(&tree, "allocate new block: goal %d, found %d\n", +@@ -2274,11 +2288,14 @@ int ext3_ext_get_block(handle_t *handle, + newex.ee_block = iblock; + newex.ee_start = newblock; + newex.ee_start_hi = 0; +- newex.ee_len = 1; ++ newex.ee_len = ar.len; + err = ext3_ext_insert_extent(handle, &tree, path, &newex); + if (err) { + /* free data blocks we just allocated */ +- ext3_free_blocks(handle, inode, newex.ee_start, newex.ee_len); ++ /* not a good idea to call discard here directly, ++ * but otherwise we'd need to call it every free() */ ++ ext3_mb_discard_inode_preallocations(inode); ++ ext3_free_blocks(handle, inode, newex.ee_start, newex.ee_len, 0); + goto out2; + } + +@@ -2287,6 +2304,7 @@ int ext3_ext_get_block(handle_t *handle, + + /* previous routine could use block we allocated */ + newblock = newex.ee_start; ++ allocated = newex.ee_len; + set_buffer_new(bh_result); + + ext3_ext_put_in_cache(&tree, newex.ee_block, newex.ee_len, +@@ -2339,6 +2357,9 @@ void ext3_ext_truncate(struct inode * in + down(&EXT3_I(inode)->truncate_sem); + ext3_ext_invalidate_cache(&tree); + ++ /* it's important to discard preallocations under truncate_sem */ ++ ext3_mb_discard_inode_preallocations(inode); ++ + /* + * TODO: optimization is possible here + * probably we need not scaning at all, +Index: linux-2.6.16.27-0.9-full/fs/ext3/Makefile +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/Makefile 2007-03-28 05:12:50.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/fs/ext3/Makefile 2007-03-28 16:03:20.000000000 +0400 +@@ -6,7 +6,7 @@ obj-$(CONFIG_EXT3_FS) += ext3.o + + ext3-y := balloc.o bitmap.o dir.o file.o fsync.o ialloc.o inode.o iopen.o \ + ioctl.o namei.o super.o symlink.o hash.o resize.o \ +- extents.o ++ extents.o mballoc.o + + ext3-$(CONFIG_EXT3_FS_XATTR) += xattr.o xattr_user.o xattr_trusted.o + ext3-$(CONFIG_EXT3_FS_POSIX_ACL) += acl.o +Index: linux-2.6.16.27-0.9-full/fs/ext3/xattr.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/xattr.c 2007-03-13 02:56:52.000000000 +0300 ++++ linux-2.6.16.27-0.9-full/fs/ext3/xattr.c 2007-03-28 16:03:20.000000000 +0400 +@@ -484,7 +484,7 @@ ext3_xattr_release_block(handle_t *handl + ea_bdebug(bh, "refcount now=0; freeing"); + if (ce) + mb_cache_entry_free(ce); +- ext3_free_blocks(handle, inode, bh->b_blocknr, 1); ++ ext3_free_blocks(handle, inode, bh->b_blocknr, 1, 1); + get_bh(bh); + ext3_forget(handle, 1, inode, bh, bh->b_blocknr); + } else { +@@ -804,7 +804,7 @@ inserted: + new_bh = sb_getblk(sb, block); + if (!new_bh) { + getblk_failed: +- ext3_free_blocks(handle, inode, block, 1); ++ ext3_free_blocks(handle, inode, block, 1, 1); + error = -EIO; + goto cleanup; + } +Index: linux-2.6.16.27-0.9-full/fs/ext3/balloc.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/balloc.c 2007-03-13 02:56:52.000000000 +0300 ++++ linux-2.6.16.27-0.9-full/fs/ext3/balloc.c 2007-03-28 16:03:20.000000000 +0400 +@@ -80,7 +80,7 @@ struct ext3_group_desc * ext3_get_group_ + * + * Return buffer_head on success or NULL in case of failure. + */ +-static struct buffer_head * ++struct buffer_head * + read_block_bitmap(struct super_block *sb, unsigned int block_group) + { + struct ext3_group_desc * desc; +@@ -296,6 +296,8 @@ void ext3_discard_reservation(struct ino + struct ext3_reserve_window_node *rsv; + spinlock_t *rsv_lock = &EXT3_SB(inode->i_sb)->s_rsv_window_lock; + ++ ext3_mb_discard_inode_preallocations(inode); ++ + if (!block_i) + return; + +@@ -491,21 +493,25 @@ error_return: + return; + } + +-/* Free given blocks, update quota and i_blocks field */ +-void ext3_free_blocks(handle_t *handle, struct inode *inode, +- unsigned long block, unsigned long count) ++void ext3_free_blocks(handle_t *handle, struct inode * inode, ++ unsigned long block, unsigned long count, int metadata) + { +- struct super_block * sb; +- int dquot_freed_blocks; ++ struct super_block *sb; ++ int freed; ++ ++ /* this isn't the right place to decide whether block is metadata ++ * inode.c/extents.c knows better, but for safety ... */ ++ if (S_ISDIR(inode->i_mode) || S_ISLNK(inode->i_mode) || ++ ext3_should_journal_data(inode)) ++ metadata = 1; + + sb = inode->i_sb; +- if (!sb) { +- printk ("ext3_free_blocks: nonexistent device"); +- return; +- } +- ext3_free_blocks_sb(handle, sb, block, count, &dquot_freed_blocks); +- if (dquot_freed_blocks) +- DQUOT_FREE_BLOCK(inode, dquot_freed_blocks); ++ if (!test_opt(sb, MBALLOC) || !EXT3_SB(sb)->s_group_info) ++ ext3_free_blocks_sb(handle, sb, block, count, &freed); ++ else ++ ext3_mb_free_blocks(handle, inode, block, count, metadata, &freed); ++ if (freed) ++ DQUOT_FREE_BLOCK(inode, freed); + return; + } + +@@ -1154,7 +1160,7 @@ int ext3_should_retry_alloc(struct super + * bitmap, and then for any free bit if that fails. + * This function also updates quota and i_blocks field. + */ +-int ext3_new_block(handle_t *handle, struct inode *inode, ++int ext3_new_block_old(handle_t *handle, struct inode *inode, + unsigned long goal, int *errp) + { + struct buffer_head *bitmap_bh = NULL; +Index: linux-2.6.16.27-0.9-full/fs/ext3/inode.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/inode.c 2007-03-28 05:13:38.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/fs/ext3/inode.c 2007-03-28 16:03:20.000000000 +0400 +@@ -568,7 +568,7 @@ static int ext3_alloc_branch(handle_t *h + ext3_journal_forget(handle, branch[i].bh); + } + for (i = 0; i < keys; i++) +- ext3_free_blocks(handle, inode, le32_to_cpu(branch[i].key), 1); ++ ext3_free_blocks(handle, inode, le32_to_cpu(branch[i].key), 1, 0); + return err; + } + +@@ -1865,7 +1865,7 @@ ext3_clear_blocks(handle_t *handle, stru + } + } + +- ext3_free_blocks(handle, inode, block_to_free, count); ++ ext3_free_blocks(handle, inode, block_to_free, count, 0); + } + + /** +@@ -2038,7 +2038,7 @@ static void ext3_free_branches(handle_t + ext3_journal_test_restart(handle, inode); + } + +- ext3_free_blocks(handle, inode, nr, 1); ++ ext3_free_blocks(handle, inode, nr, 1, 1); + + if (parent_bh) { + /* diff --git a/ldiskfs/kernel_patches/patches/ext3-mballoc3-suse.patch b/ldiskfs/kernel_patches/patches/ext3-mballoc3-suse.patch new file mode 100644 index 0000000000000000000000000000000000000000..dd071486882804bdb6dcace69eb7cadba5dcc164 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-mballoc3-suse.patch @@ -0,0 +1,397 @@ +Index: linux-2.6.5-7.283-full/include/linux/ext3_fs_i.h +=================================================================== +--- linux-2.6.5-7.283-full.orig/include/linux/ext3_fs_i.h 2007-03-28 02:13:37.000000000 +0400 ++++ linux-2.6.5-7.283-full/include/linux/ext3_fs_i.h 2007-03-28 15:46:02.000000000 +0400 +@@ -131,6 +131,10 @@ struct ext3_inode_info { + struct inode vfs_inode; + + struct ext3_ext_cache i_cached_extent; ++ ++ /* mballoc */ ++ struct list_head i_prealloc_list; ++ spinlock_t i_prealloc_lock; + }; + + #endif /* _LINUX_EXT3_FS_I */ +Index: linux-2.6.5-7.283-full/include/linux/ext3_fs_sb.h +=================================================================== +--- linux-2.6.5-7.283-full.orig/include/linux/ext3_fs_sb.h 2007-03-28 15:46:00.000000000 +0400 ++++ linux-2.6.5-7.283-full/include/linux/ext3_fs_sb.h 2007-03-28 15:46:02.000000000 +0400 +@@ -23,9 +23,16 @@ + #define EXT_INCLUDE + #include <linux/blockgroup_lock.h> + #include <linux/percpu_counter.h> ++#include <linux/list.h> + #endif + #endif + #include <linux/rbtree.h> ++#include <linux/proc_fs.h> ++ ++struct ext3_buddy_group_blocks; ++struct ext3_locality_group; ++struct ext3_mb_history; ++#define EXT3_BB_MAX_BLOCKS + + /* + * third extended-fs super-block data in memory +Index: linux-2.6.5-7.283-full/include/linux/ext3_fs.h +=================================================================== +--- linux-2.6.5-7.283-full.orig/include/linux/ext3_fs.h 2007-03-28 15:46:00.000000000 +0400 ++++ linux-2.6.5-7.283-full/include/linux/ext3_fs.h 2007-03-28 15:46:02.000000000 +0400 +@@ -363,6 +363,7 @@ struct ext3_inode { + #define EXT3_MOUNT_IOPEN_NOPRIV 0x100000/* Make iopen world-readable */ + #define EXT3_MOUNT_EXTENTS 0x200000/* Extents support */ + #define EXT3_MOUNT_EXTDEBUG 0x400000/* Extents debug */ ++#define EXT3_MOUNT_MBALLOC 0x800000/* Buddy allocation support */ + + /* Compatibility, for having both ext2_fs.h and ext3_fs.h included at once */ + #ifndef clear_opt +@@ -723,8 +724,9 @@ struct dir_private_info { + extern int ext3_bg_has_super(struct super_block *sb, int group); + extern unsigned long ext3_bg_num_gdb(struct super_block *sb, int group); + extern int ext3_new_block (handle_t *, struct inode *, unsigned long, int *); ++extern int ext3_new_block_old (handle_t *, struct inode *, unsigned long, int *); + extern void ext3_free_blocks (handle_t *, struct inode *, unsigned long, +- unsigned long); ++ unsigned long, int); + extern unsigned long ext3_count_free_blocks (struct super_block *); + extern void ext3_check_blocks_bitmap (struct super_block *); + extern struct ext3_group_desc * ext3_get_group_desc(struct super_block * sb, +Index: linux-2.6.5-7.283-full/fs/ext3/super.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/super.c 2007-03-28 15:46:00.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/super.c 2007-03-28 15:46:02.000000000 +0400 +@@ -622,6 +622,7 @@ enum { + Opt_err, + Opt_iopen, Opt_noiopen, Opt_iopen_nopriv, + Opt_extents, Opt_noextents, Opt_extdebug, ++ Opt_mballoc, Opt_nomballoc, Opt_stripe, + }; + + static match_table_t tokens = { +@@ -669,6 +670,9 @@ static match_table_t tokens = { + {Opt_noextents, "noextents"}, + {Opt_extdebug, "extdebug"}, + {Opt_barrier, "barrier=%u"}, ++ {Opt_mballoc, "mballoc"}, ++ {Opt_nomballoc, "nomballoc"}, ++ {Opt_stripe, "stripe=%u"}, + {Opt_err, NULL} + }; + +@@ -893,6 +897,19 @@ static int parse_options (char * options + case Opt_extdebug: + set_opt (sbi->s_mount_opt, EXTDEBUG); + break; ++ case Opt_mballoc: ++ set_opt(sbi->s_mount_opt, MBALLOC); ++ break; ++ case Opt_nomballoc: ++ clear_opt(sbi->s_mount_opt, MBALLOC); ++ break; ++ case Opt_stripe: ++ if (match_int(&args[0], &option)) ++ return 0; ++ if (option < 0) ++ return 0; ++ sbi->s_stripe = option; ++ break; + default: + printk (KERN_ERR + "EXT3-fs: Unrecognized mount option \"%s\" " +@@ -1548,6 +1565,7 @@ static int ext3_fill_super (struct super + ext3_count_dirs(sb)); + + ext3_ext_init(sb); ++ ext3_mb_init(sb, needs_recovery); + + return 0; + +Index: linux-2.6.5-7.283-full/fs/ext3/extents.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/extents.c 2007-03-28 03:18:19.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/extents.c 2007-03-28 15:46:02.000000000 +0400 +@@ -779,7 +779,7 @@ cleanup: + for (i = 0; i < depth; i++) { + if (!ablocks[i]) + continue; +- ext3_free_blocks(handle, tree->inode, ablocks[i], 1); ++ ext3_free_blocks(handle, tree->inode, ablocks[i], 1, 1); + } + } + kfree(ablocks); +@@ -1586,7 +1586,7 @@ int ext3_ext_rm_idx(handle_t *handle, st + path->p_idx->ei_leaf); + bh = sb_find_get_block(tree->inode->i_sb, path->p_idx->ei_leaf); + ext3_forget(handle, 1, tree->inode, bh, path->p_idx->ei_leaf); +- ext3_free_blocks(handle, tree->inode, path->p_idx->ei_leaf, 1); ++ ext3_free_blocks(handle, tree->inode, path->p_idx->ei_leaf, 1, 1); + return err; + } + +@@ -2071,10 +2071,12 @@ ext3_remove_blocks(struct ext3_extents_t + int needed = ext3_remove_blocks_credits(tree, ex, from, to); + handle_t *handle = ext3_journal_start(tree->inode, needed); + struct buffer_head *bh; +- int i; ++ int i, metadata = 0; + + if (IS_ERR(handle)) + return PTR_ERR(handle); ++ if (S_ISDIR(tree->inode->i_mode) || S_ISLNK(tree->inode->i_mode)) ++ metadata = 1; + if (from >= ex->ee_block && to == ex->ee_block + ex->ee_len - 1) { + /* tail removal */ + unsigned long num, start; +@@ -2086,7 +2088,7 @@ ext3_remove_blocks(struct ext3_extents_t + bh = sb_find_get_block(tree->inode->i_sb, start + i); + ext3_forget(handle, 0, tree->inode, bh, start + i); + } +- ext3_free_blocks(handle, tree->inode, start, num); ++ ext3_free_blocks(handle, tree->inode, start, num, metadata); + } else if (from == ex->ee_block && to <= ex->ee_block + ex->ee_len - 1) { + printk("strange request: removal %lu-%lu from %u:%u\n", + from, to, ex->ee_block, ex->ee_len); +@@ -2177,11 +2179,8 @@ int ext3_ext_get_block(handle_t *handle, + struct ext3_extent *ex; + int goal, newblock, err = 0, depth; + struct ext3_extents_tree tree; +- unsigned long next; +- int allocated = 0; +- +- /* until we have multiblock allocation */ +- max_blocks = 1; ++ unsigned long allocated = 0; ++ struct ext3_allocation_request ar; + + __clear_bit(BH_New, &bh_result->b_state); + ext3_init_tree_desc(&tree, inode); +@@ -2253,18 +2252,33 @@ int ext3_ext_get_block(handle_t *handle, + goto out2; + } + ++ /* find neighbour allocated blocks */ ++ ar.lleft = iblock; ++ err = ext3_ext_search_left(&tree, path, &ar.lleft, &ar.pleft); ++ if (err) ++ goto out2; ++ ar.lright = iblock; ++ err = ext3_ext_search_right(&tree, path, &ar.lright, &ar.pright); ++ if (err) ++ goto out2; ++ + /* find next allocated block so that we know how many + * blocks we can allocate without ovelapping next extent */ +- EXT_ASSERT(iblock >= ex->ee_block + ex->ee_len); +- next = ext3_ext_next_allocated_block(path); +- EXT_ASSERT(next > iblock); +- allocated = next - iblock; ++ EXT_ASSERT(ar.pright == 0 || ar.lright > iblock); ++ if (ar.pright == 0) ++ allocated = EXT_MAX_BLOCK - iblock; ++ else ++ allocated = ar.lright - iblock; + if (allocated > max_blocks) + allocated = max_blocks; + + /* allocate new block */ +- goal = ext3_ext_find_goal(inode, path, iblock); +- newblock = ext3_new_block(handle, inode, goal, &err); ++ ar.inode = inode; ++ ar.goal = ext3_ext_find_goal(inode, path, iblock); ++ ar.logical = iblock; ++ ar.len = allocated; ++ ar.flags = EXT3_MB_HINT_DATA; ++ newblock = ext3_mb_new_blocks(handle, &ar, &err); + if (!newblock) + goto out2; + ext_debug(&tree, "allocate new block: goal %d, found %d\n", +@@ -2274,11 +2288,14 @@ int ext3_ext_get_block(handle_t *handle, + newex.ee_block = iblock; + newex.ee_start = newblock; + newex.ee_start_hi = 0; +- newex.ee_len = 1; ++ newex.ee_len = ar.len; + err = ext3_ext_insert_extent(handle, &tree, path, &newex); + if (err) { + /* free data blocks we just allocated */ +- ext3_free_blocks(handle, inode, newex.ee_start, newex.ee_len); ++ /* not a good idea to call discard here directly, ++ * but otherwise we'd need to call it every free() */ ++ ext3_mb_discard_inode_preallocations(inode); ++ ext3_free_blocks(handle, inode, newex.ee_start, newex.ee_len, 0); + goto out2; + } + +@@ -2287,6 +2304,7 @@ int ext3_ext_get_block(handle_t *handle, + + /* previous routine could use block we allocated */ + newblock = newex.ee_start; ++ allocated = newex.ee_len; + __set_bit(BH_New, &bh_result->b_state); + + ext3_ext_put_in_cache(&tree, newex.ee_block, newex.ee_len, +@@ -2341,6 +2359,9 @@ void ext3_ext_truncate(struct inode * in + down(&EXT3_I(inode)->truncate_sem); + ext3_ext_invalidate_cache(&tree); + ++ /* it's important to discard preallocations under truncate_sem */ ++ ext3_mb_discard_inode_preallocations(inode); ++ + /* + * TODO: optimization is possible here + * probably we need not scaning at all, +Index: linux-2.6.5-7.283-full/fs/ext3/Makefile +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/Makefile 2007-03-28 15:27:39.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/Makefile 2007-03-28 15:46:02.000000000 +0400 +@@ -6,7 +6,7 @@ obj-$(CONFIG_EXT3_FS) += ext3.o + + ext3-y := balloc.o bitmap.o dir.o file.o fsync.o ialloc.o inode.o iopen.o \ + ioctl.o namei.o super.o symlink.o hash.o \ +- extents.o ++ extents.o mballoc.o + + ext3-$(CONFIG_EXT3_FS_XATTR) += xattr.o xattr_user.o xattr_trusted.o + ext3-$(CONFIG_EXT3_FS_POSIX_ACL) += acl.o +Index: linux-2.6.5-7.283-full/fs/ext3/xattr.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/xattr.c 2007-03-28 02:13:37.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/xattr.c 2007-03-28 15:46:02.000000000 +0400 +@@ -1371,7 +1371,7 @@ ext3_xattr_set_handle2(handle_t *handle, + new_bh = sb_getblk(sb, block); + if (!new_bh) { + getblk_failed: +- ext3_free_blocks(handle, inode, block, 1); ++ ext3_free_blocks(handle, inode, block, 1, 1); + error = -EIO; + goto cleanup; + } +@@ -1411,7 +1411,7 @@ getblk_failed: + if (HDR(old_bh)->h_refcount == cpu_to_le32(1)) { + /* Free the old block. */ + ea_bdebug(old_bh, "freeing"); +- ext3_free_blocks(handle, inode, old_bh->b_blocknr, 1); ++ ext3_free_blocks(handle, inode, old_bh->b_blocknr, 1, 1); + + /* ext3_forget() calls bforget() for us, but we + let our caller release old_bh, so we need to +@@ -1519,7 +1519,7 @@ ext3_xattr_delete_inode(handle_t *handle + mb_cache_entry_free(ce); + ce = NULL; + } +- ext3_free_blocks(handle, inode, EXT3_I(inode)->i_file_acl, 1); ++ ext3_free_blocks(handle, inode, EXT3_I(inode)->i_file_acl, 1, 1); + get_bh(bh); + ext3_forget(handle, 1, inode, bh, EXT3_I(inode)->i_file_acl); + } else { +Index: linux-2.6.5-7.283-full/fs/ext3/balloc.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/balloc.c 2006-12-01 18:39:48.000000000 +0300 ++++ linux-2.6.5-7.283-full/fs/ext3/balloc.c 2007-03-28 15:46:02.000000000 +0400 +@@ -78,7 +78,7 @@ struct ext3_group_desc * ext3_get_group_ + * + * Return buffer_head on success or NULL in case of failure. + */ +-static struct buffer_head * ++struct buffer_head * + read_block_bitmap(struct super_block *sb, unsigned int block_group) + { + struct ext3_group_desc * desc; +@@ -266,6 +266,8 @@ void ext3_discard_reservation(struct ino + struct reserve_window_node *rsv = &ei->i_rsv_window; + spinlock_t *rsv_lock = &EXT3_SB(inode->i_sb)->s_rsv_window_lock; + ++ ext3_mb_discard_inode_preallocations(inode); ++ + if (!rsv_is_empty(&rsv->rsv_window)) { + spin_lock(rsv_lock); + rsv_window_remove(inode->i_sb, rsv); +@@ -274,7 +276,7 @@ void ext3_discard_reservation(struct ino + } + + /* Free given blocks, update quota and i_blocks field */ +-void ext3_free_blocks(handle_t *handle, struct inode *inode, ++void ext3_free_blocks_old(handle_t *handle, struct inode *inode, + unsigned long block, unsigned long count) + { + struct buffer_head *bitmap_bh = NULL; +@@ -456,6 +458,29 @@ error_return: + return; + } + ++void ext3_free_blocks(handle_t *handle, struct inode * inode, ++ unsigned long block, unsigned long count, int metadata) ++{ ++ struct super_block *sb; ++ int freed; ++ ++ /* this isn't the right place to decide whether block is metadata ++ * inode.c/extents.c knows better, but for safety ... */ ++ if (S_ISDIR(inode->i_mode) || S_ISLNK(inode->i_mode) || ++ ext3_should_journal_data(inode)) ++ metadata = 1; ++ ++ sb = inode->i_sb; ++ if (!test_opt(sb, MBALLOC) || !EXT3_SB(sb)->s_group_info) ++ ext3_free_blocks_old(handle, inode, block, count); ++ else { ++ ext3_mb_free_blocks(handle, inode, block, count, metadata, &freed); ++ if (freed) ++ DQUOT_FREE_BLOCK(inode, freed); ++ } ++ return; ++} ++ + /* + * For ext3 allocations, we must not reuse any blocks which are + * allocated in the bitmap buffer's "last committed data" copy. This +@@ -1142,7 +1167,7 @@ int ext3_should_retry_alloc(struct super + * bitmap, and then for any free bit if that fails. + * This function also updates quota and i_blocks field. + */ +-int ext3_new_block(handle_t *handle, struct inode *inode, ++int ext3_new_block_old(handle_t *handle, struct inode *inode, + unsigned long goal, int *errp) + { + struct buffer_head *bitmap_bh = NULL; +Index: linux-2.6.5-7.283-full/fs/ext3/inode.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/inode.c 2007-03-28 02:50:19.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/inode.c 2007-03-28 15:46:02.000000000 +0400 +@@ -574,7 +574,7 @@ static int ext3_alloc_branch(handle_t *h + ext3_journal_forget(handle, branch[i].bh); + } + for (i = 0; i < keys; i++) +- ext3_free_blocks(handle, inode, le32_to_cpu(branch[i].key), 1); ++ ext3_free_blocks(handle, inode, le32_to_cpu(branch[i].key), 1, 0); + return err; + } + +@@ -675,7 +675,7 @@ err_out: + if (err == -EAGAIN) + for (i = 0; i < num; i++) + ext3_free_blocks(handle, inode, +- le32_to_cpu(where[i].key), 1); ++ le32_to_cpu(where[i].key), 1, 0); + return err; + } + +@@ -1839,7 +1839,7 @@ ext3_clear_blocks(handle_t *handle, stru + } + } + +- ext3_free_blocks(handle, inode, block_to_free, count); ++ ext3_free_blocks(handle, inode, block_to_free, count, 0); + } + + /** +@@ -2010,7 +2010,7 @@ static void ext3_free_branches(handle_t + ext3_journal_test_restart(handle, inode); + } + +- ext3_free_blocks(handle, inode, nr, 1); ++ ext3_free_blocks(handle, inode, nr, 1, 1); + + if (parent_bh) { + /* diff --git a/ldiskfs/kernel_patches/patches/ext3-uninit-2.6-sles10.patch b/ldiskfs/kernel_patches/patches/ext3-uninit-2.6-sles10.patch new file mode 100644 index 0000000000000000000000000000000000000000..62b1f50a9095134e9232f091ddfa131d52ed3906 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-uninit-2.6-sles10.patch @@ -0,0 +1,674 @@ +Add support for the uninit_groups feature to the kernel. + +Keep a high water mark of used inodes for each group to improve e2fsck time. +Block and inode bitmaps can be uninitialized on disk via a flag in the +group descriptor to avoid reading or scanning them at e2fsck time. +A checksum of each group descriptor is used to ensure that corruption in +the group descriptor's bit flags does not cause incorrect operation. + +Index: linux-2.6.16.27-0.9-full/include/linux/ext3_fs.h +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/include/linux/ext3_fs.h 2007-03-28 18:20:16.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/include/linux/ext3_fs.h 2007-03-28 18:30:06.000000000 +0400 +@@ -153,16 +153,22 @@ struct ext3_allocation_request { + */ + struct ext3_group_desc + { +- __le32 bg_block_bitmap; /* Blocks bitmap block */ +- __le32 bg_inode_bitmap; /* Inodes bitmap block */ ++ __le32 bg_block_bitmap; /* Blocks bitmap block */ ++ __le32 bg_inode_bitmap; /* Inodes bitmap block */ + __le32 bg_inode_table; /* Inodes table block */ + __le16 bg_free_blocks_count; /* Free blocks count */ + __le16 bg_free_inodes_count; /* Free inodes count */ + __le16 bg_used_dirs_count; /* Directories count */ +- __u16 bg_pad; +- __le32 bg_reserved[3]; ++ __le16 bg_flags; /* EXT3_BG_flags (UNINIT, etc) */ ++ __le32 bg_reserved[2]; /* Likely block/inode bitmap checksum */ ++ __le16 bg_itable_unused; /* Unused inodes count */ ++ __le16 bg_checksum; /* crc16(sb_uuid+group+desc) */ + }; + ++#define EXT3_BG_INODE_UNINIT 0x0001 /* Inode table/bitmap not in use */ ++#define EXT3_BG_BLOCK_UNINIT 0x0002 /* Block bitmap not in use */ ++#define EXT3_BG_INODE_ZEROED 0x0004 /* On-disk itable initialized to zero */ ++ + /* + * Macro-instructions used to manage group descriptors + */ +@@ -590,6 +596,7 @@ static inline struct ext3_inode_info *EX + #define EXT3_FEATURE_RO_COMPAT_SPARSE_SUPER 0x0001 + #define EXT3_FEATURE_RO_COMPAT_LARGE_FILE 0x0002 + #define EXT3_FEATURE_RO_COMPAT_BTREE_DIR 0x0004 ++#define EXT4_FEATURE_RO_COMPAT_GDT_CSUM 0x0010 + #define EXT4_FEATURE_RO_COMPAT_DIR_NLINK 0x0020 + + #define EXT3_FEATURE_INCOMPAT_COMPRESSION 0x0001 +@@ -606,6 +613,7 @@ static inline struct ext3_inode_info *EX + EXT3_FEATURE_INCOMPAT_EXTENTS) + #define EXT3_FEATURE_RO_COMPAT_SUPP (EXT3_FEATURE_RO_COMPAT_SPARSE_SUPER| \ + EXT3_FEATURE_RO_COMPAT_LARGE_FILE| \ ++ EXT4_FEATURE_RO_COMPAT_GDT_CSUM| \ + EXT4_FEATURE_RO_COMPAT_DIR_NLINK| \ + EXT3_FEATURE_RO_COMPAT_BTREE_DIR) + +Index: linux-2.6.16.27-0.9-full/fs/ext3/resize.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/resize.c 2007-03-13 02:56:52.000000000 +0300 ++++ linux-2.6.16.27-0.9-full/fs/ext3/resize.c 2007-03-28 18:30:06.000000000 +0400 +@@ -19,6 +19,7 @@ + #include <linux/errno.h> + #include <linux/slab.h> + ++#include "group.h" + + #define outside(b, first, last) ((b) < (first) || (b) >= (last)) + #define inside(b, first, last) ((b) >= (first) && (b) < (last)) +@@ -818,6 +819,7 @@ int ext3_group_add(struct super_block *s + gdp->bg_inode_table = cpu_to_le32(input->inode_table); + gdp->bg_free_blocks_count = cpu_to_le16(input->free_blocks_count); + gdp->bg_free_inodes_count = cpu_to_le16(EXT3_INODES_PER_GROUP(sb)); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, input->group, gdp); + + /* + * Make the new blocks and inodes valid next. We do this before +Index: linux-2.6.16.27-0.9-full/fs/ext3/super.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/super.c 2007-03-28 18:25:51.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/fs/ext3/super.c 2007-03-28 18:30:06.000000000 +0400 +@@ -42,6 +42,7 @@ + #include "xattr.h" + #include "acl.h" + #include "namei.h" ++#include "group.h" + + static int ext3_load_journal(struct super_block *, struct ext3_super_block *, + unsigned long journal_devnum); +@@ -1221,6 +1222,90 @@ static int ext3_setup_super(struct super + return res; + } + ++#if !defined(CONFIG_CRC16) && !defined(CONFIG_CRC16_MODULE) ++/** CRC table for the CRC-16. The poly is 0x8005 (x^16 + x^15 + x^2 + 1) */ ++__u16 const crc16_table[256] = { ++ 0x0000, 0xC0C1, 0xC181, 0x0140, 0xC301, 0x03C0, 0x0280, 0xC241, ++ 0xC601, 0x06C0, 0x0780, 0xC741, 0x0500, 0xC5C1, 0xC481, 0x0440, ++ 0xCC01, 0x0CC0, 0x0D80, 0xCD41, 0x0F00, 0xCFC1, 0xCE81, 0x0E40, ++ 0x0A00, 0xCAC1, 0xCB81, 0x0B40, 0xC901, 0x09C0, 0x0880, 0xC841, ++ 0xD801, 0x18C0, 0x1980, 0xD941, 0x1B00, 0xDBC1, 0xDA81, 0x1A40, ++ 0x1E00, 0xDEC1, 0xDF81, 0x1F40, 0xDD01, 0x1DC0, 0x1C80, 0xDC41, ++ 0x1400, 0xD4C1, 0xD581, 0x1540, 0xD701, 0x17C0, 0x1680, 0xD641, ++ 0xD201, 0x12C0, 0x1380, 0xD341, 0x1100, 0xD1C1, 0xD081, 0x1040, ++ 0xF001, 0x30C0, 0x3180, 0xF141, 0x3300, 0xF3C1, 0xF281, 0x3240, ++ 0x3600, 0xF6C1, 0xF781, 0x3740, 0xF501, 0x35C0, 0x3480, 0xF441, ++ 0x3C00, 0xFCC1, 0xFD81, 0x3D40, 0xFF01, 0x3FC0, 0x3E80, 0xFE41, ++ 0xFA01, 0x3AC0, 0x3B80, 0xFB41, 0x3900, 0xF9C1, 0xF881, 0x3840, ++ 0x2800, 0xE8C1, 0xE981, 0x2940, 0xEB01, 0x2BC0, 0x2A80, 0xEA41, ++ 0xEE01, 0x2EC0, 0x2F80, 0xEF41, 0x2D00, 0xEDC1, 0xEC81, 0x2C40, ++ 0xE401, 0x24C0, 0x2580, 0xE541, 0x2700, 0xE7C1, 0xE681, 0x2640, ++ 0x2200, 0xE2C1, 0xE381, 0x2340, 0xE101, 0x21C0, 0x2080, 0xE041, ++ 0xA001, 0x60C0, 0x6180, 0xA141, 0x6300, 0xA3C1, 0xA281, 0x6240, ++ 0x6600, 0xA6C1, 0xA781, 0x6740, 0xA501, 0x65C0, 0x6480, 0xA441, ++ 0x6C00, 0xACC1, 0xAD81, 0x6D40, 0xAF01, 0x6FC0, 0x6E80, 0xAE41, ++ 0xAA01, 0x6AC0, 0x6B80, 0xAB41, 0x6900, 0xA9C1, 0xA881, 0x6840, ++ 0x7800, 0xB8C1, 0xB981, 0x7940, 0xBB01, 0x7BC0, 0x7A80, 0xBA41, ++ 0xBE01, 0x7EC0, 0x7F80, 0xBF41, 0x7D00, 0xBDC1, 0xBC81, 0x7C40, ++ 0xB401, 0x74C0, 0x7580, 0xB541, 0x7700, 0xB7C1, 0xB681, 0x7640, ++ 0x7200, 0xB2C1, 0xB381, 0x7340, 0xB101, 0x71C0, 0x7080, 0xB041, ++ 0x5000, 0x90C1, 0x9181, 0x5140, 0x9301, 0x53C0, 0x5280, 0x9241, ++ 0x9601, 0x56C0, 0x5780, 0x9741, 0x5500, 0x95C1, 0x9481, 0x5440, ++ 0x9C01, 0x5CC0, 0x5D80, 0x9D41, 0x5F00, 0x9FC1, 0x9E81, 0x5E40, ++ 0x5A00, 0x9AC1, 0x9B81, 0x5B40, 0x9901, 0x59C0, 0x5880, 0x9841, ++ 0x8801, 0x48C0, 0x4980, 0x8941, 0x4B00, 0x8BC1, 0x8A81, 0x4A40, ++ 0x4E00, 0x8EC1, 0x8F81, 0x4F40, 0x8D01, 0x4DC0, 0x4C80, 0x8C41, ++ 0x4400, 0x84C1, 0x8581, 0x4540, 0x8701, 0x47C0, 0x4680, 0x8641, ++ 0x8201, 0x42C0, 0x4380, 0x8341, 0x4100, 0x81C1, 0x8081, 0x4040 ++}; ++ ++static inline __u16 crc16_byte(__u16 crc, const __u8 data) ++{ ++ return (crc >> 8) ^ crc16_table[(crc ^ data) & 0xff]; ++} ++ ++__u16 crc16(__u16 crc, __u8 const *buffer, size_t len) ++{ ++ while (len--) ++ crc = crc16_byte(crc, *buffer++); ++ return crc; ++} ++#endif ++ ++__le16 ext3_group_desc_csum(struct ext3_sb_info *sbi, __u32 block_group, ++ struct ext3_group_desc *gdp) ++{ ++ __u16 crc = 0; ++ ++ if (sbi->s_es->s_feature_ro_compat & ++ cpu_to_le32(EXT4_FEATURE_RO_COMPAT_GDT_CSUM)) { ++ int offset = offsetof(struct ext3_group_desc, bg_checksum); ++ __le32 le_group = cpu_to_le32(block_group); ++ ++ crc = crc16(~0, sbi->s_es->s_uuid, sizeof(sbi->s_es->s_uuid)); ++ crc = crc16(crc, (__u8 *)&le_group, sizeof(le_group)); ++ crc = crc16(crc, (__u8 *)gdp, offset); ++ offset += sizeof(gdp->bg_checksum); /* skip checksum */ ++ BUG_ON(offset != sizeof(*gdp)); /* XXX handle s_desc_size */ ++ /* for checksum of struct ext4_group_desc do the rest... ++ if (offset < sbi->s_es->s_desc_size) { ++ crc = crc16(crc, (__u8 *)gdp + offset, ++ sbi->s_es->s_desc_size - offset); ++ */ ++ } ++ ++ return cpu_to_le16(crc); ++} ++ ++int ext3_group_desc_csum_verify(struct ext3_sb_info *sbi, __u32 block_group, ++ struct ext3_group_desc *gdp) ++{ ++ if (gdp->bg_checksum != ext3_group_desc_csum(sbi, block_group, gdp)) ++ return 0; ++ ++ return 1; ++} ++ + /* Called at mount-time, super-block is locked */ + static int ext3_check_descriptors (struct super_block * sb) + { +@@ -1270,6 +1355,13 @@ static int ext3_check_descriptors (struc + le32_to_cpu(gdp->bg_inode_table)); + return 0; + } ++ if (!ext3_group_desc_csum_verify(sbi, i, gdp)) { ++ ext3_error(sb, __FUNCTION__, ++ "Checksum for group %d failed (%u!=%u)\n", i, ++ le16_to_cpu(ext3_group_desc_csum(sbi,i,gdp)), ++ le16_to_cpu(gdp->bg_checksum)); ++ return 0; ++ } + block += EXT3_BLOCKS_PER_GROUP(sb); + gdp++; + } +Index: linux-2.6.16.27-0.9-full/fs/ext3/group.h +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/group.h 2007-02-13 18:39:59.640066087 +0300 ++++ linux-2.6.16.27-0.9-full/fs/ext3/group.h 2007-03-28 18:30:06.000000000 +0400 +@@ -0,0 +1,29 @@ ++/* ++ * linux/fs/ext3/group.h ++ * ++ * Copyright (C) 2007 Cluster File Systems, Inc ++ * ++ * Author: Andreas Dilger <adilger@clusterfs.com> ++ */ ++ ++#ifndef _LINUX_EXT3_GROUP_H ++#define _LINUX_EXT3_GROUP_H ++#if defined(CONFIG_CRC16) || defined(CONFIG_CRC16_MODULE) ++#include <linux/crc16.h> ++#endif ++ ++extern __le16 ext3_group_desc_csum(struct ext3_sb_info *sbi, __u32 group, ++ struct ext3_group_desc *gdp); ++extern int ext3_group_desc_csum_verify(struct ext3_sb_info *sbi, __u32 group, ++ struct ext3_group_desc *gdp); ++struct buffer_head *read_block_bitmap(struct super_block *sb, ++ unsigned int block_group); ++extern unsigned ext3_init_block_bitmap(struct super_block *sb, ++ struct buffer_head *bh, int group, ++ struct ext3_group_desc *desc); ++#define ext3_free_blocks_after_init(sb, group, desc) \ ++ ext3_init_block_bitmap(sb, NULL, group, desc) ++extern unsigned ext3_init_inode_bitmap(struct super_block *sb, ++ struct buffer_head *bh, int group, ++ struct ext3_group_desc *desc); ++#endif /* _LINUX_EXT3_GROUP_H */ +Index: linux-2.6.16.27-0.9-full/fs/ext3/ialloc.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/ialloc.c 2007-03-28 18:20:17.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/fs/ext3/ialloc.c 2007-03-28 18:30:06.000000000 +0400 +@@ -28,6 +28,7 @@ + + #include "xattr.h" + #include "acl.h" ++#include "group.h" + + /* + * ialloc.c contains the inodes allocation and deallocation routines +@@ -43,6 +44,52 @@ + * the free blocks count in the block. + */ + ++/* ++ * To avoid calling the atomic setbit hundreds or thousands of times, we only ++ * need to use it within a single byte (to ensure we get endianness right). ++ * We can use memset for the rest of the bitmap as there are no other users. ++ */ ++static void mark_bitmap_end(int start_bit, int end_bit, char *bitmap) ++{ ++ int i; ++ ++ if (start_bit >= end_bit) ++ return; ++ ++ ext3_debug("mark end bits +%d through +%d used\n", start_bit, end_bit); ++ for (i = start_bit; i < ((start_bit + 7) & ~7UL); i++) ++ ext3_set_bit(i, bitmap); ++ if (i < end_bit) ++ memset(bitmap + (i >> 3), 0xff, (end_bit - i) >> 3); ++} ++ ++/* Initializes an uninitialized inode bitmap */ ++unsigned ext3_init_inode_bitmap(struct super_block *sb, ++ struct buffer_head *bh, int block_group, ++ struct ext3_group_desc *gdp) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ ++ J_ASSERT_BH(bh, buffer_locked(bh)); ++ ++ /* If checksum is bad mark all blocks and inodes use to prevent ++ * allocation, essentially implementing a per-group read-only flag. */ ++ if (!ext3_group_desc_csum_verify(sbi, block_group, gdp)) { ++ ext3_error(sb, __FUNCTION__, "Checksum bad for group %u\n", ++ block_group); ++ gdp->bg_free_blocks_count = 0; ++ gdp->bg_free_inodes_count = 0; ++ gdp->bg_itable_unused = 0; ++ memset(bh->b_data, 0xff, sb->s_blocksize); ++ return 0; ++ } ++ ++ memset(bh->b_data, 0, (EXT3_INODES_PER_GROUP(sb) + 7) / 8); ++ mark_bitmap_end(EXT3_INODES_PER_GROUP(sb), EXT3_BLOCKS_PER_GROUP(sb), ++ bh->b_data); ++ ++ return EXT3_INODES_PER_GROUP(sb); ++} + + /* + * Read the inode allocation bitmap for a given block_group, reading +@@ -59,8 +106,19 @@ read_inode_bitmap(struct super_block * s + desc = ext3_get_group_desc(sb, block_group, NULL); + if (!desc) + goto error_out; +- +- bh = sb_bread(sb, le32_to_cpu(desc->bg_inode_bitmap)); ++ if (desc->bg_flags & cpu_to_le16(EXT3_BG_INODE_UNINIT)) { ++ bh = sb_getblk(sb, le32_to_cpu(desc->bg_inode_bitmap)); ++ if (!buffer_uptodate(bh)) { ++ lock_buffer(bh); ++ if (!buffer_uptodate(bh)) { ++ ext3_init_inode_bitmap(sb, bh,block_group,desc); ++ set_buffer_uptodate(bh); ++ } ++ unlock_buffer(bh); ++ } ++ } else { ++ bh = sb_bread(sb, le32_to_cpu(desc->bg_inode_bitmap)); ++ } + if (!bh) + ext3_error(sb, "read_inode_bitmap", + "Cannot read inode bitmap - " +@@ -169,6 +227,8 @@ void ext3_free_inode (handle_t *handle, + if (is_directory) + gdp->bg_used_dirs_count = cpu_to_le16( + le16_to_cpu(gdp->bg_used_dirs_count) - 1); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi,block_group, ++ gdp); + spin_unlock(sb_bgl_lock(sbi, block_group)); + percpu_counter_inc(&sbi->s_freeinodes_counter); + if (is_directory) +@@ -453,7 +513,7 @@ struct inode *ext3_new_inode(handle_t *h + struct ext3_sb_info *sbi; + int err = 0; + struct inode *ret; +- int i; ++ int i, free = 0; + + /* Cannot create files in a deleted directory */ + if (!dir || !dir->i_nlink) +@@ -570,11 +630,13 @@ repeat_in_this_group: + goto out; + + got: +- ino += group * EXT3_INODES_PER_GROUP(sb) + 1; +- if (ino < EXT3_FIRST_INO(sb) || ino > le32_to_cpu(es->s_inodes_count)) { +- ext3_error (sb, "ext3_new_inode", +- "reserved inode or inode > inodes count - " +- "block_group = %d, inode=%lu", group, ino); ++ ino++; ++ if ((group == 0 && ino < EXT3_FIRST_INO(sb)) || ++ ino > EXT3_INODES_PER_GROUP(sb)) { ++ ext3_error(sb, __FUNCTION__, ++ "reserved inode or inode > inodes count - " ++ "block_group = %d, inode=%lu", group, ++ ino + group * EXT3_INODES_PER_GROUP(sb)); + err = -EIO; + goto fail; + } +@@ -582,13 +644,65 @@ got: + BUFFER_TRACE(bh2, "get_write_access"); + err = ext3_journal_get_write_access(handle, bh2); + if (err) goto fail; ++ ++ /* We may have to initialize the block bitmap if it isn't already */ ++ if (EXT3_HAS_RO_COMPAT_FEATURE(sb, EXT4_FEATURE_RO_COMPAT_GDT_CSUM) && ++ gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ struct buffer_head *block_bh = read_block_bitmap(sb, group); ++ ++ BUFFER_TRACE(block_bh, "get block bitmap access"); ++ err = ext3_journal_get_write_access(handle, block_bh); ++ if (err) { ++ brelse(block_bh); ++ goto fail; ++ } ++ ++ free = 0; ++ spin_lock(sb_bgl_lock(sbi, group)); ++ /* recheck and clear flag under lock if we still need to */ ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_BLOCK_UNINIT); ++ free = ext3_free_blocks_after_init(sb, group, gdp); ++ gdp->bg_free_blocks_count = cpu_to_le16(free); ++ } ++ spin_unlock(sb_bgl_lock(sbi, group)); ++ ++ /* Don't need to dirty bitmap block if we didn't change it */ ++ if (free) { ++ BUFFER_TRACE(block_bh, "dirty block bitmap"); ++ err = ext3_journal_dirty_metadata(handle, block_bh); ++ } ++ ++ brelse(block_bh); ++ if (err) ++ goto fail; ++ } ++ + spin_lock(sb_bgl_lock(sbi, group)); ++ /* If we didn't allocate from within the initialized part of the inode ++ * table then we need to initialize up to this inode. */ ++ if (EXT3_HAS_RO_COMPAT_FEATURE(sb, EXT4_FEATURE_RO_COMPAT_GDT_CSUM)) { ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_INODE_UNINIT)) { ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_INODE_UNINIT); ++ free = EXT3_INODES_PER_GROUP(sb); ++ } else { ++ free = EXT3_INODES_PER_GROUP(sb) - ++ le16_to_cpu(gdp->bg_itable_unused); ++ } ++ ++ if (ino > free) { ++ gdp->bg_itable_unused = ++ cpu_to_le16(EXT3_INODES_PER_GROUP(sb) - ino); ++ } ++ } ++ + gdp->bg_free_inodes_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_inodes_count) - 1); + if (S_ISDIR(mode)) { + gdp->bg_used_dirs_count = + cpu_to_le16(le16_to_cpu(gdp->bg_used_dirs_count) + 1); + } ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, group, gdp); + spin_unlock(sb_bgl_lock(sbi, group)); + BUFFER_TRACE(bh2, "call ext3_journal_dirty_metadata"); + err = ext3_journal_dirty_metadata(handle, bh2); +@@ -610,7 +724,7 @@ got: + inode->i_gid = current->fsgid; + inode->i_mode = mode; + +- inode->i_ino = ino; ++ inode->i_ino = ino + group * EXT3_INODES_PER_GROUP(sb); + /* This is the optimal IO size (for stat), not the fs block size */ + inode->i_blksize = PAGE_SIZE; + inode->i_blocks = 0; +Index: linux-2.6.16.27-0.9-full/fs/ext3/mballoc.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/mballoc.c 2007-03-28 16:03:19.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/fs/ext3/mballoc.c 2007-03-28 18:30:36.000000000 +0400 +@@ -36,6 +36,8 @@ + #include <linux/seq_file.h> + #include <linux/version.h> + ++#include "group.h" ++ + /* + * MUSTDO: + * - test ext3_ext_search_left() and ext3_ext_search_right() +@@ -323,6 +325,7 @@ struct ext3_group_info { + unsigned long bb_state; + unsigned long bb_tid; + struct ext3_free_metadata *bb_md_cur; ++ struct ext3_group_desc *bb_gdp; + unsigned short bb_first_free; + unsigned short bb_free; + unsigned short bb_fragments; +@@ -928,10 +931,7 @@ static int ext3_mb_init_cache(struct pag + if (first_group + i >= EXT3_SB(sb)->s_groups_count) + break; + +- err = -EIO; +- desc = ext3_get_group_desc(sb, first_group + i, NULL); +- if (desc == NULL) +- goto out; ++ desc = EXT3_GROUP_INFO(sb, first_group + i)->bb_gdp; + + err = -ENOMEM; + bh[i] = sb_getblk(sb, le32_to_cpu(desc->bg_block_bitmap)); +@@ -946,7 +946,12 @@ static int ext3_mb_init_cache(struct pag + unlock_buffer(bh[i]); + continue; + } +- ++ if (desc->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ ext3_init_block_bitmap(sb, bh[i], first_group + i,desc); ++ set_buffer_uptodate(bh[i]); ++ unlock_buffer(bh[i]); ++ continue; ++ } + get_bh(bh[i]); + bh[i]->b_end_io = end_buffer_read_sync; + submit_bh(READ, bh[i]); +@@ -1703,6 +1708,10 @@ static int ext3_mb_good_group(struct ext + switch (cr) { + case 0: + BUG_ON(ac->ac_2order == 0); ++ /* If this group is uninitialized, skip it initially */ ++ if (grp->bb_gdp->bg_flags & ++ cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) ++ return 0; + bits = ac->ac_sb->s_blocksize_bits + 1; + for (i = ac->ac_2order; i <= bits; i++) + if (grp->bb_counters[i] > 0) +@@ -1796,7 +1805,9 @@ repeat: + } + + ac->ac_groups_scanned++; +- if (cr == 0) ++ if (cr == 0 || (e3b.bd_info->bb_gdp->bg_flags & ++ cpu_to_le16(EXT3_BG_BLOCK_UNINIT) && ++ ac->ac_2order != 0)) + ext3_mb_simple_scan_group(ac, &e3b); + else if (cr == 1 && ac->ac_g_ex.fe_len == sbi->s_stripe) + ext3_mb_scan_aligned(ac, &e3b); +@@ -2267,12 +2278,13 @@ int ext3_mb_init_backend(struct super_bl + i--; + goto err_freebuddy; + } ++ memset(meta_group_info[j], 0, len); + desc = ext3_get_group_desc(sb, i, NULL); ++ meta_group_info[j]->bb_gdp = desc; + if (desc == NULL) { + printk(KERN_ERR"EXT3-fs: can't read descriptor %u\n",i); + goto err_freebuddy; + } +- memset(meta_group_info[j], 0, len); + set_bit(EXT3_GROUP_INFO_NEED_INIT_BIT, + &meta_group_info[j]->bb_state); + +@@ -2936,9 +2948,17 @@ int ext3_mb_mark_diskspace_used(struct e + mb_set_bits(bitmap_bh->b_data, ac->ac_b_ex.fe_start, ac->ac_b_ex.fe_len); + + spin_lock(sb_bgl_lock(sbi, ac->ac_b_ex.fe_group)); ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_BLOCK_UNINIT); ++ gdp->bg_free_blocks_count = ++ cpu_to_le16(ext3_free_blocks_after_init(sb, ++ ac->ac_b_ex.fe_group, ++ gdp)); ++ } + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) + - ac->ac_b_ex.fe_len); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, ac->ac_b_ex.fe_group, gdp); + spin_unlock(sb_bgl_lock(sbi, ac->ac_b_ex.fe_group)); + percpu_counter_mod(&sbi->s_freeblocks_counter, - ac->ac_b_ex.fe_len); + +@@ -4303,6 +4323,7 @@ do_more: + spin_lock(sb_bgl_lock(sbi, block_group)); + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) + count); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, block_group, gdp); + spin_unlock(sb_bgl_lock(sbi, block_group)); + percpu_counter_mod(&sbi->s_freeblocks_counter, count); + +Index: linux-2.6.16.27-0.9-full/fs/ext3/balloc.c +=================================================================== +--- linux-2.6.16.27-0.9-full.orig/fs/ext3/balloc.c 2007-03-28 16:03:20.000000000 +0400 ++++ linux-2.6.16.27-0.9-full/fs/ext3/balloc.c 2007-03-28 18:30:06.000000000 +0400 +@@ -21,6 +21,7 @@ + #include <linux/quotaops.h> + #include <linux/buffer_head.h> + ++#include "group.h" + /* + * balloc.c contains the blocks allocation and deallocation routines + */ +@@ -74,6 +75,75 @@ struct ext3_group_desc * ext3_get_group_ + return desc + offset; + } + ++/* Initializes an uninitialized block bitmap if given, and returns the ++ * number of blocks free in the group. */ ++unsigned ext3_init_block_bitmap(struct super_block *sb, struct buffer_head *bh, ++ int block_group, struct ext3_group_desc *gdp) ++{ ++ unsigned long start; ++ int bit, bit_max; ++ unsigned free_blocks; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ ++ if (bh) { ++ J_ASSERT_BH(bh, buffer_locked(bh)); ++ ++ /* If checksum is bad mark all blocks use to prevent allocation, ++ * essentially implementing a per-group read-only flag. */ ++ if (!ext3_group_desc_csum_verify(sbi, block_group, gdp)) { ++ ext3_error(sb, __FUNCTION__, ++ "Checksum bad for group %u\n", block_group); ++ gdp->bg_free_blocks_count = 0; ++ gdp->bg_free_inodes_count = 0; ++ gdp->bg_itable_unused = 0; ++ memset(bh->b_data, 0xff, sb->s_blocksize); ++ return 0; ++ } ++ memset(bh->b_data, 0, sb->s_blocksize); ++ } ++ ++ /* Check for superblock and gdt backups in this group */ ++ bit_max = ext3_bg_has_super(sb, block_group); ++ ++ if (!EXT3_HAS_INCOMPAT_FEATURE(sb, EXT3_FEATURE_INCOMPAT_META_BG) || ++ block_group < le32_to_cpu(sbi->s_es->s_first_meta_bg) * ++ sbi->s_desc_per_block) { ++ if (bit_max) { ++ bit_max += ext3_bg_num_gdb(sb, block_group); ++ bit_max +=le16_to_cpu(sbi->s_es->s_reserved_gdt_blocks); ++ } ++ } else { /* For META_BG_BLOCK_GROUPS */ ++ int group_rel = (block_group - ++ le32_to_cpu(sbi->s_es->s_first_meta_bg)) % ++ EXT3_DESC_PER_BLOCK(sb); ++ if (group_rel == 0 || group_rel == 1 || ++ (group_rel == EXT3_DESC_PER_BLOCK(sb) - 1)) ++ bit_max += 1; ++ } ++ ++ /* Last and first groups are always initialized */ ++ free_blocks = EXT3_BLOCKS_PER_GROUP(sb) - bit_max; ++ ++ if (bh) { ++ for (bit = 0; bit < bit_max; bit++) ++ ext3_set_bit(bit, bh->b_data); ++ ++ start = block_group * EXT3_BLOCKS_PER_GROUP(sb) + ++ le32_to_cpu(sbi->s_es->s_first_data_block); ++ ++ /* Set bits for block and inode bitmaps, and inode table */ ++ ext3_set_bit(le32_to_cpu(gdp->bg_block_bitmap) - start, ++ bh->b_data); ++ ext3_set_bit(le32_to_cpu(gdp->bg_inode_bitmap) - start, ++ bh->b_data); ++ for (bit = le32_to_cpu(gdp->bg_inode_table) - start, ++ bit_max = bit + sbi->s_itb_per_group; bit < bit_max; bit++) ++ ext3_set_bit(bit, bh->b_data); ++ } ++ ++ return free_blocks - sbi->s_itb_per_group - 2; ++} ++ + /* + * Read the bitmap for a given block_group, reading into the specified + * slot in the superblock's bitmap cache. +@@ -89,7 +159,19 @@ read_block_bitmap(struct super_block *sb + desc = ext3_get_group_desc (sb, block_group, NULL); + if (!desc) + goto error_out; +- bh = sb_bread(sb, le32_to_cpu(desc->bg_block_bitmap)); ++ if (desc->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ bh = sb_getblk(sb, le32_to_cpu(desc->bg_block_bitmap)); ++ if (!buffer_uptodate(bh)) { ++ lock_buffer(bh); ++ if (!buffer_uptodate(bh)) { ++ ext3_init_block_bitmap(sb, bh,block_group,desc); ++ set_buffer_uptodate(bh); ++ } ++ unlock_buffer(bh); ++ } ++ } else { ++ bh = sb_bread(sb, le32_to_cpu(desc->bg_block_bitmap)); ++ } + if (!bh) + ext3_error (sb, "read_block_bitmap", + "Cannot read block bitmap - " +@@ -468,6 +550,7 @@ do_more: + desc->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(desc->bg_free_blocks_count) + + group_freed); ++ desc->bg_checksum = ext3_group_desc_csum(sbi, block_group, desc); + spin_unlock(sb_bgl_lock(sbi, block_group)); + percpu_counter_mod(&sbi->s_freeblocks_counter, count); + +@@ -1378,8 +1461,11 @@ allocated: + ret_block, goal_hits, goal_attempts); + + spin_lock(sb_bgl_lock(sbi, group_no)); ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_BLOCK_UNINIT); + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) - 1); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, group_no, gdp); + spin_unlock(sb_bgl_lock(sbi, group_no)); + percpu_counter_mod(&sbi->s_freeblocks_counter, -1); + + +%diffstat + fs/ext3/balloc.c | 88 +++++++++++++++++++++++++++++ + fs/ext3/group.h | 38 ++++++++++++ + fs/ext3/ialloc.c | 144 +++++++++++++++++++++++++++++++++++++++++++----- + fs/ext3/mballoc.c | 35 +++++++++-- + fs/ext3/resize.c | 2 + fs/ext3/super.c | 92 ++++++++++++++++++++++++++++++ + include/linux/ext3_fs.h | 16 ++++- + 7 files changed, 388 insertions(+), 27 deletions(-) diff --git a/ldiskfs/kernel_patches/patches/ext3-uninit-2.6-suse.patch b/ldiskfs/kernel_patches/patches/ext3-uninit-2.6-suse.patch new file mode 100644 index 0000000000000000000000000000000000000000..8a34ea5afa32f5bde910da727052a339fc2b6015 --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-uninit-2.6-suse.patch @@ -0,0 +1,653 @@ +Add support for the uninit_groups feature to the kernel. + +Keep a high water mark of used inodes for each group to improve e2fsck time. +Block and inode bitmaps can be uninitialized on disk via a flag in the +group descriptor to avoid reading or scanning them at e2fsck time. +A checksum of each group descriptor is used to ensure that corruption in +the group descriptor's bit flags does not cause incorrect operation. + +Index: linux-2.6.5-7.283-full/include/linux/ext3_fs.h +=================================================================== +--- linux-2.6.5-7.283-full.orig/include/linux/ext3_fs.h 2007-03-28 17:33:05.000000000 +0400 ++++ linux-2.6.5-7.283-full/include/linux/ext3_fs.h 2007-03-28 18:33:35.000000000 +0400 +@@ -153,16 +153,22 @@ struct ext3_allocation_request { + */ + struct ext3_group_desc + { +- __u32 bg_block_bitmap; /* Blocks bitmap block */ +- __u32 bg_inode_bitmap; /* Inodes bitmap block */ ++ __u32 bg_block_bitmap; /* Blocks bitmap block */ ++ __u32 bg_inode_bitmap; /* Inodes bitmap block */ + __u32 bg_inode_table; /* Inodes table block */ + __u16 bg_free_blocks_count; /* Free blocks count */ + __u16 bg_free_inodes_count; /* Free inodes count */ + __u16 bg_used_dirs_count; /* Directories count */ +- __u16 bg_pad; +- __u32 bg_reserved[3]; ++ __u16 bg_flags; /* EXT3_BG_flags (UNINIT, etc) */ ++ __u32 bg_reserved[2]; /* Likely block/inode bitmap checksum */ ++ __u16 bg_itable_unused; /* Unused inodes count */ ++ __u16 bg_checksum; /* crc16(sb_uuid+group+desc) */ + }; + ++#define EXT3_BG_INODE_UNINIT 0x0001 /* Inode table/bitmap not in use */ ++#define EXT3_BG_BLOCK_UNINIT 0x0002 /* Block bitmap not in use */ ++#define EXT3_BG_INODE_ZEROED 0x0004 /* On-disk itable initialized to zero */ ++ + /* + * Macro-instructions used to manage group descriptors + */ +@@ -458,7 +464,7 @@ struct ext3_super_block { + */ + __u8 s_prealloc_blocks; /* Nr of blocks to try to preallocate*/ + __u8 s_prealloc_dir_blocks; /* Nr to preallocate for dirs */ +- __u16 s_padding1; ++ __u16 s_reserved_gdt_blocks; /* Per group desc for online growth */ + /* + * Journaling support valid if EXT3_FEATURE_COMPAT_HAS_JOURNAL set. + */ +@@ -546,6 +552,7 @@ static inline struct ext3_inode_info *EX + #define EXT3_FEATURE_RO_COMPAT_SPARSE_SUPER 0x0001 + #define EXT3_FEATURE_RO_COMPAT_LARGE_FILE 0x0002 + #define EXT3_FEATURE_RO_COMPAT_BTREE_DIR 0x0004 ++#define EXT4_FEATURE_RO_COMPAT_GDT_CSUM 0x0010 + #define EXT4_FEATURE_RO_COMPAT_DIR_NLINK 0x0020 + + #define EXT3_FEATURE_INCOMPAT_COMPRESSION 0x0001 +@@ -562,6 +569,7 @@ static inline struct ext3_inode_info *EX + EXT3_FEATURE_INCOMPAT_EXTENTS) + #define EXT3_FEATURE_RO_COMPAT_SUPP (EXT3_FEATURE_RO_COMPAT_SPARSE_SUPER| \ + EXT3_FEATURE_RO_COMPAT_LARGE_FILE| \ ++ EXT4_FEATURE_RO_COMPAT_GDT_CSUM| \ + EXT4_FEATURE_RO_COMPAT_DIR_NLINK| \ + EXT3_FEATURE_RO_COMPAT_BTREE_DIR) + +Index: linux-2.6.5-7.283-full/fs/ext3/super.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/super.c 2007-03-28 17:33:05.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/super.c 2007-03-28 18:33:35.000000000 +0400 +@@ -36,6 +36,7 @@ + #include <linux/quotaops.h> + #include "xattr.h" + #include "acl.h" ++#include "group.h" + + static int ext3_load_journal(struct super_block *, struct ext3_super_block *); + static int ext3_create_journal(struct super_block *, struct ext3_super_block *, +@@ -996,6 +997,90 @@ static int ext3_setup_super(struct super + return res; + } + ++#if !defined(CONFIG_CRC16) && !defined(CONFIG_CRC16_MODULE) ++/** CRC table for the CRC-16. The poly is 0x8005 (x^16 + x^15 + x^2 + 1) */ ++__u16 const crc16_table[256] = { ++ 0x0000, 0xC0C1, 0xC181, 0x0140, 0xC301, 0x03C0, 0x0280, 0xC241, ++ 0xC601, 0x06C0, 0x0780, 0xC741, 0x0500, 0xC5C1, 0xC481, 0x0440, ++ 0xCC01, 0x0CC0, 0x0D80, 0xCD41, 0x0F00, 0xCFC1, 0xCE81, 0x0E40, ++ 0x0A00, 0xCAC1, 0xCB81, 0x0B40, 0xC901, 0x09C0, 0x0880, 0xC841, ++ 0xD801, 0x18C0, 0x1980, 0xD941, 0x1B00, 0xDBC1, 0xDA81, 0x1A40, ++ 0x1E00, 0xDEC1, 0xDF81, 0x1F40, 0xDD01, 0x1DC0, 0x1C80, 0xDC41, ++ 0x1400, 0xD4C1, 0xD581, 0x1540, 0xD701, 0x17C0, 0x1680, 0xD641, ++ 0xD201, 0x12C0, 0x1380, 0xD341, 0x1100, 0xD1C1, 0xD081, 0x1040, ++ 0xF001, 0x30C0, 0x3180, 0xF141, 0x3300, 0xF3C1, 0xF281, 0x3240, ++ 0x3600, 0xF6C1, 0xF781, 0x3740, 0xF501, 0x35C0, 0x3480, 0xF441, ++ 0x3C00, 0xFCC1, 0xFD81, 0x3D40, 0xFF01, 0x3FC0, 0x3E80, 0xFE41, ++ 0xFA01, 0x3AC0, 0x3B80, 0xFB41, 0x3900, 0xF9C1, 0xF881, 0x3840, ++ 0x2800, 0xE8C1, 0xE981, 0x2940, 0xEB01, 0x2BC0, 0x2A80, 0xEA41, ++ 0xEE01, 0x2EC0, 0x2F80, 0xEF41, 0x2D00, 0xEDC1, 0xEC81, 0x2C40, ++ 0xE401, 0x24C0, 0x2580, 0xE541, 0x2700, 0xE7C1, 0xE681, 0x2640, ++ 0x2200, 0xE2C1, 0xE381, 0x2340, 0xE101, 0x21C0, 0x2080, 0xE041, ++ 0xA001, 0x60C0, 0x6180, 0xA141, 0x6300, 0xA3C1, 0xA281, 0x6240, ++ 0x6600, 0xA6C1, 0xA781, 0x6740, 0xA501, 0x65C0, 0x6480, 0xA441, ++ 0x6C00, 0xACC1, 0xAD81, 0x6D40, 0xAF01, 0x6FC0, 0x6E80, 0xAE41, ++ 0xAA01, 0x6AC0, 0x6B80, 0xAB41, 0x6900, 0xA9C1, 0xA881, 0x6840, ++ 0x7800, 0xB8C1, 0xB981, 0x7940, 0xBB01, 0x7BC0, 0x7A80, 0xBA41, ++ 0xBE01, 0x7EC0, 0x7F80, 0xBF41, 0x7D00, 0xBDC1, 0xBC81, 0x7C40, ++ 0xB401, 0x74C0, 0x7580, 0xB541, 0x7700, 0xB7C1, 0xB681, 0x7640, ++ 0x7200, 0xB2C1, 0xB381, 0x7340, 0xB101, 0x71C0, 0x7080, 0xB041, ++ 0x5000, 0x90C1, 0x9181, 0x5140, 0x9301, 0x53C0, 0x5280, 0x9241, ++ 0x9601, 0x56C0, 0x5780, 0x9741, 0x5500, 0x95C1, 0x9481, 0x5440, ++ 0x9C01, 0x5CC0, 0x5D80, 0x9D41, 0x5F00, 0x9FC1, 0x9E81, 0x5E40, ++ 0x5A00, 0x9AC1, 0x9B81, 0x5B40, 0x9901, 0x59C0, 0x5880, 0x9841, ++ 0x8801, 0x48C0, 0x4980, 0x8941, 0x4B00, 0x8BC1, 0x8A81, 0x4A40, ++ 0x4E00, 0x8EC1, 0x8F81, 0x4F40, 0x8D01, 0x4DC0, 0x4C80, 0x8C41, ++ 0x4400, 0x84C1, 0x8581, 0x4540, 0x8701, 0x47C0, 0x4680, 0x8641, ++ 0x8201, 0x42C0, 0x4380, 0x8341, 0x4100, 0x81C1, 0x8081, 0x4040 ++}; ++ ++static inline __u16 crc16_byte(__u16 crc, const __u8 data) ++{ ++ return (crc >> 8) ^ crc16_table[(crc ^ data) & 0xff]; ++} ++ ++__u16 crc16(__u16 crc, __u8 const *buffer, size_t len) ++{ ++ while (len--) ++ crc = crc16_byte(crc, *buffer++); ++ return crc; ++} ++#endif ++ ++__le16 ext3_group_desc_csum(struct ext3_sb_info *sbi, __u32 block_group, ++ struct ext3_group_desc *gdp) ++{ ++ __u16 crc = 0; ++ ++ if (sbi->s_es->s_feature_ro_compat & ++ cpu_to_le32(EXT4_FEATURE_RO_COMPAT_GDT_CSUM)) { ++ int offset = offsetof(struct ext3_group_desc, bg_checksum); ++ __le32 le_group = cpu_to_le32(block_group); ++ ++ crc = crc16(~0, sbi->s_es->s_uuid, sizeof(sbi->s_es->s_uuid)); ++ crc = crc16(crc, (__u8 *)&le_group, sizeof(le_group)); ++ crc = crc16(crc, (__u8 *)gdp, offset); ++ offset += sizeof(gdp->bg_checksum); /* skip checksum */ ++ BUG_ON(offset != sizeof(*gdp)); /* XXX handle s_desc_size */ ++ /* for checksum of struct ext4_group_desc do the rest... ++ if (offset < sbi->s_es->s_desc_size) { ++ crc = crc16(crc, (__u8 *)gdp + offset, ++ sbi->s_es->s_desc_size - offset); ++ */ ++ } ++ ++ return cpu_to_le16(crc); ++} ++ ++int ext3_group_desc_csum_verify(struct ext3_sb_info *sbi, __u32 block_group, ++ struct ext3_group_desc *gdp) ++{ ++ if (gdp->bg_checksum != ext3_group_desc_csum(sbi, block_group, gdp)) ++ return 0; ++ ++ return 1; ++} ++ + static int ext3_check_descriptors (struct super_block * sb) + { + struct ext3_sb_info *sbi = EXT3_SB(sb); +@@ -1044,6 +1129,13 @@ static int ext3_check_descriptors (struc + le32_to_cpu(gdp->bg_inode_table)); + return 0; + } ++ if (!ext3_group_desc_csum_verify(sbi, i, gdp)) { ++ ext3_error(sb, __FUNCTION__, ++ "Checksum for group %d failed (%u!=%u)\n", i, ++ le16_to_cpu(ext3_group_desc_csum(sbi,i,gdp)), ++ le16_to_cpu(gdp->bg_checksum)); ++ return 0; ++ } + block += EXT3_BLOCKS_PER_GROUP(sb); + gdp++; + } +Index: linux-2.6.5-7.283-full/fs/ext3/group.h +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/group.h 2007-02-13 18:39:59.640066087 +0300 ++++ linux-2.6.5-7.283-full/fs/ext3/group.h 2007-03-28 18:33:35.000000000 +0400 +@@ -0,0 +1,29 @@ ++/* ++ * linux/fs/ext3/group.h ++ * ++ * Copyright (C) 2007 Cluster File Systems, Inc ++ * ++ * Author: Andreas Dilger <adilger@clusterfs.com> ++ */ ++ ++#ifndef _LINUX_EXT3_GROUP_H ++#define _LINUX_EXT3_GROUP_H ++#if defined(CONFIG_CRC16) || defined(CONFIG_CRC16_MODULE) ++#include <linux/crc16.h> ++#endif ++ ++extern __le16 ext3_group_desc_csum(struct ext3_sb_info *sbi, __u32 group, ++ struct ext3_group_desc *gdp); ++extern int ext3_group_desc_csum_verify(struct ext3_sb_info *sbi, __u32 group, ++ struct ext3_group_desc *gdp); ++struct buffer_head *read_block_bitmap(struct super_block *sb, ++ unsigned int block_group); ++extern unsigned ext3_init_block_bitmap(struct super_block *sb, ++ struct buffer_head *bh, int group, ++ struct ext3_group_desc *desc); ++#define ext3_free_blocks_after_init(sb, group, desc) \ ++ ext3_init_block_bitmap(sb, NULL, group, desc) ++extern unsigned ext3_init_inode_bitmap(struct super_block *sb, ++ struct buffer_head *bh, int group, ++ struct ext3_group_desc *desc); ++#endif /* _LINUX_EXT3_GROUP_H */ +Index: linux-2.6.5-7.283-full/fs/ext3/ialloc.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/ialloc.c 2007-03-28 17:33:03.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/ialloc.c 2007-03-28 18:33:35.000000000 +0400 +@@ -28,6 +28,7 @@ + + #include "xattr.h" + #include "acl.h" ++#include "group.h" + + /* + * ialloc.c contains the inodes allocation and deallocation routines +@@ -43,6 +44,52 @@ + * the free blocks count in the block. + */ + ++/* ++ * To avoid calling the atomic setbit hundreds or thousands of times, we only ++ * need to use it within a single byte (to ensure we get endianness right). ++ * We can use memset for the rest of the bitmap as there are no other users. ++ */ ++static void mark_bitmap_end(int start_bit, int end_bit, char *bitmap) ++{ ++ int i; ++ ++ if (start_bit >= end_bit) ++ return; ++ ++ ext3_debug("mark end bits +%d through +%d used\n", start_bit, end_bit); ++ for (i = start_bit; i < ((start_bit + 7) & ~7UL); i++) ++ ext3_set_bit(i, bitmap); ++ if (i < end_bit) ++ memset(bitmap + (i >> 3), 0xff, (end_bit - i) >> 3); ++} ++ ++/* Initializes an uninitialized inode bitmap */ ++unsigned ext3_init_inode_bitmap(struct super_block *sb, ++ struct buffer_head *bh, int block_group, ++ struct ext3_group_desc *gdp) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ ++ J_ASSERT_BH(bh, buffer_locked(bh)); ++ ++ /* If checksum is bad mark all blocks and inodes use to prevent ++ * allocation, essentially implementing a per-group read-only flag. */ ++ if (!ext3_group_desc_csum_verify(sbi, block_group, gdp)) { ++ ext3_error(sb, __FUNCTION__, "Checksum bad for group %u\n", ++ block_group); ++ gdp->bg_free_blocks_count = 0; ++ gdp->bg_free_inodes_count = 0; ++ gdp->bg_itable_unused = 0; ++ memset(bh->b_data, 0xff, sb->s_blocksize); ++ return 0; ++ } ++ ++ memset(bh->b_data, 0, (EXT3_INODES_PER_GROUP(sb) + 7) / 8); ++ mark_bitmap_end(EXT3_INODES_PER_GROUP(sb), EXT3_BLOCKS_PER_GROUP(sb), ++ bh->b_data); ++ ++ return EXT3_INODES_PER_GROUP(sb); ++} + + /* + * Read the inode allocation bitmap for a given block_group, reading +@@ -59,8 +106,19 @@ read_inode_bitmap(struct super_block * s + desc = ext3_get_group_desc(sb, block_group, NULL); + if (!desc) + goto error_out; +- +- bh = sb_bread(sb, le32_to_cpu(desc->bg_inode_bitmap)); ++ if (desc->bg_flags & cpu_to_le16(EXT3_BG_INODE_UNINIT)) { ++ bh = sb_getblk(sb, le32_to_cpu(desc->bg_inode_bitmap)); ++ if (!buffer_uptodate(bh)) { ++ lock_buffer(bh); ++ if (!buffer_uptodate(bh)) { ++ ext3_init_inode_bitmap(sb, bh,block_group,desc); ++ set_buffer_uptodate(bh); ++ } ++ unlock_buffer(bh); ++ } ++ } else { ++ bh = sb_bread(sb, le32_to_cpu(desc->bg_inode_bitmap)); ++ } + if (!bh) + ext3_error(sb, "read_inode_bitmap", + "Cannot read inode bitmap - " +@@ -168,6 +226,8 @@ void ext3_free_inode (handle_t *handle, + if (is_directory) + gdp->bg_used_dirs_count = cpu_to_le16( + le16_to_cpu(gdp->bg_used_dirs_count) - 1); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi,block_group, ++ gdp); + spin_unlock(sb_bgl_lock(sbi, block_group)); + percpu_counter_inc(&sbi->s_freeinodes_counter); + if (is_directory) +@@ -454,7 +514,7 @@ struct inode *ext3_new_inode(handle_t *h + struct ext3_sb_info *sbi; + int err = 0; + struct inode *ret; +- int i; ++ int i, free = 0; + + /* Cannot create files in a deleted directory */ + if (!dir || !dir->i_nlink) +@@ -570,11 +630,13 @@ repeat_in_this_group: + goto out; + + got: +- ino += group * EXT3_INODES_PER_GROUP(sb) + 1; +- if (ino < EXT3_FIRST_INO(sb) || ino > le32_to_cpu(es->s_inodes_count)) { +- ext3_error (sb, "ext3_new_inode", +- "reserved inode or inode > inodes count - " +- "block_group = %d, inode=%lu", group, ino); ++ ino++; ++ if ((group == 0 && ino < EXT3_FIRST_INO(sb)) || ++ ino > EXT3_INODES_PER_GROUP(sb)) { ++ ext3_error(sb, __FUNCTION__, ++ "reserved inode or inode > inodes count - " ++ "block_group = %d, inode=%lu", group, ++ ino + group * EXT3_INODES_PER_GROUP(sb)); + err = -EIO; + goto fail; + } +@@ -582,13 +644,65 @@ got: + BUFFER_TRACE(bh2, "get_write_access"); + err = ext3_journal_get_write_access(handle, bh2); + if (err) goto fail; ++ ++ /* We may have to initialize the block bitmap if it isn't already */ ++ if (EXT3_HAS_RO_COMPAT_FEATURE(sb, EXT4_FEATURE_RO_COMPAT_GDT_CSUM) && ++ gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ struct buffer_head *block_bh = read_block_bitmap(sb, group); ++ ++ BUFFER_TRACE(block_bh, "get block bitmap access"); ++ err = ext3_journal_get_write_access(handle, block_bh); ++ if (err) { ++ brelse(block_bh); ++ goto fail; ++ } ++ ++ free = 0; ++ spin_lock(sb_bgl_lock(sbi, group)); ++ /* recheck and clear flag under lock if we still need to */ ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_BLOCK_UNINIT); ++ free = ext3_free_blocks_after_init(sb, group, gdp); ++ gdp->bg_free_blocks_count = cpu_to_le16(free); ++ } ++ spin_unlock(sb_bgl_lock(sbi, group)); ++ ++ /* Don't need to dirty bitmap block if we didn't change it */ ++ if (free) { ++ BUFFER_TRACE(block_bh, "dirty block bitmap"); ++ err = ext3_journal_dirty_metadata(handle, block_bh); ++ } ++ ++ brelse(block_bh); ++ if (err) ++ goto fail; ++ } ++ + spin_lock(sb_bgl_lock(sbi, group)); ++ /* If we didn't allocate from within the initialized part of the inode ++ * table then we need to initialize up to this inode. */ ++ if (EXT3_HAS_RO_COMPAT_FEATURE(sb, EXT4_FEATURE_RO_COMPAT_GDT_CSUM)) { ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_INODE_UNINIT)) { ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_INODE_UNINIT); ++ free = EXT3_INODES_PER_GROUP(sb); ++ } else { ++ free = EXT3_INODES_PER_GROUP(sb) - ++ le16_to_cpu(gdp->bg_itable_unused); ++ } ++ ++ if (ino > free) { ++ gdp->bg_itable_unused = ++ cpu_to_le16(EXT3_INODES_PER_GROUP(sb) - ino); ++ } ++ } ++ + gdp->bg_free_inodes_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_inodes_count) - 1); + if (S_ISDIR(mode)) { + gdp->bg_used_dirs_count = + cpu_to_le16(le16_to_cpu(gdp->bg_used_dirs_count) + 1); + } ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, group, gdp); + spin_unlock(sb_bgl_lock(sbi, group)); + BUFFER_TRACE(bh2, "call ext3_journal_dirty_metadata"); + err = ext3_journal_dirty_metadata(handle, bh2); +@@ -610,7 +724,7 @@ got: + inode->i_gid = current->fsgid; + inode->i_mode = mode; + +- inode->i_ino = ino; ++ inode->i_ino = ino + group * EXT3_INODES_PER_GROUP(sb); + /* This is the optimal IO size (for stat), not the fs block size */ + inode->i_blksize = PAGE_SIZE; + inode->i_blocks = 0; +Index: linux-2.6.5-7.283-full/fs/ext3/mballoc.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/mballoc.c 2007-03-28 15:46:00.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/mballoc.c 2007-03-28 18:33:35.000000000 +0400 +@@ -36,6 +36,8 @@ + #include <linux/seq_file.h> + #include <linux/version.h> + ++#include "group.h" ++ + /* + * MUSTDO: + * - test ext3_ext_search_left() and ext3_ext_search_right() +@@ -323,6 +325,7 @@ struct ext3_group_info { + unsigned long bb_state; + unsigned long bb_tid; + struct ext3_free_metadata *bb_md_cur; ++ struct ext3_group_desc *bb_gdp; + unsigned short bb_first_free; + unsigned short bb_free; + unsigned short bb_fragments; +@@ -928,10 +931,7 @@ static int ext3_mb_init_cache(struct pag + if (first_group + i >= EXT3_SB(sb)->s_groups_count) + break; + +- err = -EIO; +- desc = ext3_get_group_desc(sb, first_group + i, NULL); +- if (desc == NULL) +- goto out; ++ desc = EXT3_GROUP_INFO(sb, first_group + i)->bb_gdp; + + err = -ENOMEM; + bh[i] = sb_getblk(sb, le32_to_cpu(desc->bg_block_bitmap)); +@@ -946,7 +946,12 @@ static int ext3_mb_init_cache(struct pag + unlock_buffer(bh[i]); + continue; + } +- ++ if (desc->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ ext3_init_block_bitmap(sb, bh[i], first_group + i,desc); ++ set_buffer_uptodate(bh[i]); ++ unlock_buffer(bh[i]); ++ continue; ++ } + get_bh(bh[i]); + bh[i]->b_end_io = end_buffer_read_sync; + submit_bh(READ, bh[i]); +@@ -1703,6 +1708,10 @@ static int ext3_mb_good_group(struct ext + switch (cr) { + case 0: + BUG_ON(ac->ac_2order == 0); ++ /* If this group is uninitialized, skip it initially */ ++ if (grp->bb_gdp->bg_flags & ++ cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) ++ return 0; + bits = ac->ac_sb->s_blocksize_bits + 1; + for (i = ac->ac_2order; i <= bits; i++) + if (grp->bb_counters[i] > 0) +@@ -1796,7 +1805,9 @@ repeat: + } + + ac->ac_groups_scanned++; +- if (cr == 0) ++ if (cr == 0 || (e3b.bd_info->bb_gdp->bg_flags & ++ cpu_to_le16(EXT3_BG_BLOCK_UNINIT) && ++ ac->ac_2order != 0)) + ext3_mb_simple_scan_group(ac, &e3b); + else if (cr == 1 && ac->ac_g_ex.fe_len == sbi->s_stripe) + ext3_mb_scan_aligned(ac, &e3b); +@@ -2267,12 +2278,13 @@ int ext3_mb_init_backend(struct super_bl + i--; + goto err_freebuddy; + } ++ memset(meta_group_info[j], 0, len); + desc = ext3_get_group_desc(sb, i, NULL); ++ meta_group_info[j]->bb_gdp = desc; + if (desc == NULL) { + printk(KERN_ERR"EXT3-fs: can't read descriptor %u\n",i); + goto err_freebuddy; + } +- memset(meta_group_info[j], 0, len); + set_bit(EXT3_GROUP_INFO_NEED_INIT_BIT, + &meta_group_info[j]->bb_state); + +@@ -2936,9 +2948,17 @@ int ext3_mb_mark_diskspace_used(struct e + mb_set_bits(bitmap_bh->b_data, ac->ac_b_ex.fe_start, ac->ac_b_ex.fe_len); + + spin_lock(sb_bgl_lock(sbi, ac->ac_b_ex.fe_group)); ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_BLOCK_UNINIT); ++ gdp->bg_free_blocks_count = ++ cpu_to_le16(ext3_free_blocks_after_init(sb, ++ ac->ac_b_ex.fe_group, ++ gdp)); ++ } + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) + - ac->ac_b_ex.fe_len); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, ac->ac_b_ex.fe_group, gdp); + spin_unlock(sb_bgl_lock(sbi, ac->ac_b_ex.fe_group)); + percpu_counter_mod(&sbi->s_freeblocks_counter, - ac->ac_b_ex.fe_len); + +@@ -4303,6 +4323,7 @@ do_more: + spin_lock(sb_bgl_lock(sbi, block_group)); + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) + count); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, block_group, gdp); + spin_unlock(sb_bgl_lock(sbi, block_group)); + percpu_counter_mod(&sbi->s_freeblocks_counter, count); + +Index: linux-2.6.5-7.283-full/fs/ext3/balloc.c +=================================================================== +--- linux-2.6.5-7.283-full.orig/fs/ext3/balloc.c 2007-03-28 17:33:02.000000000 +0400 ++++ linux-2.6.5-7.283-full/fs/ext3/balloc.c 2007-03-28 18:33:35.000000000 +0400 +@@ -20,6 +20,7 @@ + #include <linux/quotaops.h> + #include <linux/buffer_head.h> + ++#include "group.h" + /* + * balloc.c contains the blocks allocation and deallocation routines + */ +@@ -72,6 +73,75 @@ struct ext3_group_desc * ext3_get_group_ + return gdp + desc; + } + ++/* Initializes an uninitialized block bitmap if given, and returns the ++ * number of blocks free in the group. */ ++unsigned ext3_init_block_bitmap(struct super_block *sb, struct buffer_head *bh, ++ int block_group, struct ext3_group_desc *gdp) ++{ ++ unsigned long start; ++ int bit, bit_max; ++ unsigned free_blocks; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ ++ if (bh) { ++ J_ASSERT_BH(bh, buffer_locked(bh)); ++ ++ /* If checksum is bad mark all blocks use to prevent allocation, ++ * essentially implementing a per-group read-only flag. */ ++ if (!ext3_group_desc_csum_verify(sbi, block_group, gdp)) { ++ ext3_error(sb, __FUNCTION__, ++ "Checksum bad for group %u\n", block_group); ++ gdp->bg_free_blocks_count = 0; ++ gdp->bg_free_inodes_count = 0; ++ gdp->bg_itable_unused = 0; ++ memset(bh->b_data, 0xff, sb->s_blocksize); ++ return 0; ++ } ++ memset(bh->b_data, 0, sb->s_blocksize); ++ } ++ ++ /* Check for superblock and gdt backups in this group */ ++ bit_max = ext3_bg_has_super(sb, block_group); ++ ++ if (!EXT3_HAS_INCOMPAT_FEATURE(sb, EXT3_FEATURE_INCOMPAT_META_BG) || ++ block_group < le32_to_cpu(sbi->s_es->s_first_meta_bg) * ++ sbi->s_desc_per_block) { ++ if (bit_max) { ++ bit_max += ext3_bg_num_gdb(sb, block_group); ++ bit_max +=le16_to_cpu(sbi->s_es->s_reserved_gdt_blocks); ++ } ++ } else { /* For META_BG_BLOCK_GROUPS */ ++ int group_rel = (block_group - ++ le32_to_cpu(sbi->s_es->s_first_meta_bg)) % ++ EXT3_DESC_PER_BLOCK(sb); ++ if (group_rel == 0 || group_rel == 1 || ++ (group_rel == EXT3_DESC_PER_BLOCK(sb) - 1)) ++ bit_max += 1; ++ } ++ ++ /* Last and first groups are always initialized */ ++ free_blocks = EXT3_BLOCKS_PER_GROUP(sb) - bit_max; ++ ++ if (bh) { ++ for (bit = 0; bit < bit_max; bit++) ++ ext3_set_bit(bit, bh->b_data); ++ ++ start = block_group * EXT3_BLOCKS_PER_GROUP(sb) + ++ le32_to_cpu(sbi->s_es->s_first_data_block); ++ ++ /* Set bits for block and inode bitmaps, and inode table */ ++ ext3_set_bit(le32_to_cpu(gdp->bg_block_bitmap) - start, ++ bh->b_data); ++ ext3_set_bit(le32_to_cpu(gdp->bg_inode_bitmap) - start, ++ bh->b_data); ++ for (bit = le32_to_cpu(gdp->bg_inode_table) - start, ++ bit_max = bit + sbi->s_itb_per_group; bit < bit_max; bit++) ++ ext3_set_bit(bit, bh->b_data); ++ } ++ ++ return free_blocks - sbi->s_itb_per_group - 2; ++} ++ + /* + * Read the bitmap for a given block_group, reading into the specified + * slot in the superblock's bitmap cache. +@@ -87,7 +157,19 @@ read_block_bitmap(struct super_block *sb + desc = ext3_get_group_desc (sb, block_group, NULL); + if (!desc) + goto error_out; +- bh = sb_bread(sb, le32_to_cpu(desc->bg_block_bitmap)); ++ if (desc->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ bh = sb_getblk(sb, le32_to_cpu(desc->bg_block_bitmap)); ++ if (!buffer_uptodate(bh)) { ++ lock_buffer(bh); ++ if (!buffer_uptodate(bh)) { ++ ext3_init_block_bitmap(sb, bh,block_group,desc); ++ set_buffer_uptodate(bh); ++ } ++ unlock_buffer(bh); ++ } ++ } else { ++ bh = sb_bread(sb, le32_to_cpu(desc->bg_block_bitmap)); ++ } + if (!bh) + ext3_error (sb, "read_block_bitmap", + "Cannot read block bitmap - " +@@ -432,6 +514,7 @@ do_more: + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) + + dquot_freed_blocks); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, block_group, gdp); + spin_unlock(sb_bgl_lock(sbi, block_group)); + percpu_counter_mod(&sbi->s_freeblocks_counter, count); + +@@ -1372,8 +1455,11 @@ allocated: + ret_block, goal_hits, goal_attempts); + + spin_lock(sb_bgl_lock(sbi, group_no)); ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_BLOCK_UNINIT); + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) - 1); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, group_no, gdp); + spin_unlock(sb_bgl_lock(sbi, group_no)); + percpu_counter_mod(&sbi->s_freeblocks_counter, -1); + diff --git a/ldiskfs/kernel_patches/patches/ext3-uninit-2.6.9.patch b/ldiskfs/kernel_patches/patches/ext3-uninit-2.6.9.patch new file mode 100644 index 0000000000000000000000000000000000000000..2dbeb80a2bc763a513ac080824510ac7e8a9132b --- /dev/null +++ b/ldiskfs/kernel_patches/patches/ext3-uninit-2.6.9.patch @@ -0,0 +1,664 @@ +Add support for the uninit_groups feature to the kernel. + +Keep a high water mark of used inodes for each group to improve e2fsck time. +Block and inode bitmaps can be uninitialized on disk via a flag in the +group descriptor to avoid reading or scanning them at e2fsck time. +A checksum of each group descriptor is used to ensure that corruption in +the group descriptor's bit flags does not cause incorrect operation. + +Index: linux-2.6.9-full/include/linux/ext3_fs.h +=================================================================== +--- linux-2.6.9-full.orig/include/linux/ext3_fs.h 2007-03-28 18:35:41.000000000 +0400 ++++ linux-2.6.9-full/include/linux/ext3_fs.h 2007-03-28 18:36:16.000000000 +0400 +@@ -153,16 +153,22 @@ struct ext3_allocation_request { + */ + struct ext3_group_desc + { +- __le32 bg_block_bitmap; /* Blocks bitmap block */ +- __le32 bg_inode_bitmap; /* Inodes bitmap block */ ++ __le32 bg_block_bitmap; /* Blocks bitmap block */ ++ __le32 bg_inode_bitmap; /* Inodes bitmap block */ + __le32 bg_inode_table; /* Inodes table block */ + __le16 bg_free_blocks_count; /* Free blocks count */ + __le16 bg_free_inodes_count; /* Free inodes count */ + __le16 bg_used_dirs_count; /* Directories count */ +- __u16 bg_pad; +- __le32 bg_reserved[3]; ++ __le16 bg_flags; /* EXT3_BG_flags (UNINIT, etc) */ ++ __le32 bg_reserved[2]; /* Likely block/inode bitmap checksum */ ++ __le16 bg_itable_unused; /* Unused inodes count */ ++ __le16 bg_checksum; /* crc16(sb_uuid+group+desc) */ + }; + ++#define EXT3_BG_INODE_UNINIT 0x0001 /* Inode table/bitmap not in use */ ++#define EXT3_BG_BLOCK_UNINIT 0x0002 /* Block bitmap not in use */ ++#define EXT3_BG_INODE_ZEROED 0x0004 /* On-disk itable initialized to zero */ ++ + /* + * Macro-instructions used to manage group descriptors + */ +@@ -572,6 +578,7 @@ static inline struct ext3_inode_info *EX + #define EXT3_FEATURE_RO_COMPAT_SPARSE_SUPER 0x0001 + #define EXT3_FEATURE_RO_COMPAT_LARGE_FILE 0x0002 + #define EXT3_FEATURE_RO_COMPAT_BTREE_DIR 0x0004 ++#define EXT4_FEATURE_RO_COMPAT_GDT_CSUM 0x0010 + #define EXT4_FEATURE_RO_COMPAT_DIR_NLINK 0x0020 + + #define EXT3_FEATURE_INCOMPAT_COMPRESSION 0x0001 +@@ -588,6 +595,7 @@ static inline struct ext3_inode_info *EX + EXT3_FEATURE_INCOMPAT_EXTENTS) + #define EXT3_FEATURE_RO_COMPAT_SUPP (EXT3_FEATURE_RO_COMPAT_SPARSE_SUPER| \ + EXT3_FEATURE_RO_COMPAT_LARGE_FILE| \ ++ EXT4_FEATURE_RO_COMPAT_GDT_CSUM| \ + EXT4_FEATURE_RO_COMPAT_DIR_NLINK| \ + EXT3_FEATURE_RO_COMPAT_BTREE_DIR) + +Index: linux-2.6.9-full/fs/ext3/resize.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/resize.c 2006-03-10 18:20:03.000000000 +0300 ++++ linux-2.6.9-full/fs/ext3/resize.c 2007-03-28 18:36:16.000000000 +0400 +@@ -19,6 +19,7 @@ + #include <linux/errno.h> + #include <linux/slab.h> + ++#include "group.h" + + #define outside(b, first, last) ((b) < (first) || (b) >= (last)) + #define inside(b, first, last) ((b) >= (first) && (b) < (last)) +@@ -807,6 +808,7 @@ int ext3_group_add(struct super_block *s + gdp->bg_inode_table = cpu_to_le32(input->inode_table); + gdp->bg_free_blocks_count = cpu_to_le16(input->free_blocks_count); + gdp->bg_free_inodes_count = cpu_to_le16(EXT3_INODES_PER_GROUP(sb)); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, input->group, gdp); + + /* + * Make the new blocks and inodes valid next. We do this before +Index: linux-2.6.9-full/fs/ext3/super.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/super.c 2007-03-28 18:35:42.000000000 +0400 ++++ linux-2.6.9-full/fs/ext3/super.c 2007-03-28 18:36:16.000000000 +0400 +@@ -38,6 +38,7 @@ + #include <asm/uaccess.h> + #include "xattr.h" + #include "acl.h" ++#include "group.h" + + static int ext3_load_journal(struct super_block *, struct ext3_super_block *, + unsigned long journal_devnum); +@@ -1090,6 +1091,90 @@ static int ext3_setup_super(struct super + return res; + } + ++#if !defined(CONFIG_CRC16) && !defined(CONFIG_CRC16_MODULE) ++/** CRC table for the CRC-16. The poly is 0x8005 (x^16 + x^15 + x^2 + 1) */ ++__u16 const crc16_table[256] = { ++ 0x0000, 0xC0C1, 0xC181, 0x0140, 0xC301, 0x03C0, 0x0280, 0xC241, ++ 0xC601, 0x06C0, 0x0780, 0xC741, 0x0500, 0xC5C1, 0xC481, 0x0440, ++ 0xCC01, 0x0CC0, 0x0D80, 0xCD41, 0x0F00, 0xCFC1, 0xCE81, 0x0E40, ++ 0x0A00, 0xCAC1, 0xCB81, 0x0B40, 0xC901, 0x09C0, 0x0880, 0xC841, ++ 0xD801, 0x18C0, 0x1980, 0xD941, 0x1B00, 0xDBC1, 0xDA81, 0x1A40, ++ 0x1E00, 0xDEC1, 0xDF81, 0x1F40, 0xDD01, 0x1DC0, 0x1C80, 0xDC41, ++ 0x1400, 0xD4C1, 0xD581, 0x1540, 0xD701, 0x17C0, 0x1680, 0xD641, ++ 0xD201, 0x12C0, 0x1380, 0xD341, 0x1100, 0xD1C1, 0xD081, 0x1040, ++ 0xF001, 0x30C0, 0x3180, 0xF141, 0x3300, 0xF3C1, 0xF281, 0x3240, ++ 0x3600, 0xF6C1, 0xF781, 0x3740, 0xF501, 0x35C0, 0x3480, 0xF441, ++ 0x3C00, 0xFCC1, 0xFD81, 0x3D40, 0xFF01, 0x3FC0, 0x3E80, 0xFE41, ++ 0xFA01, 0x3AC0, 0x3B80, 0xFB41, 0x3900, 0xF9C1, 0xF881, 0x3840, ++ 0x2800, 0xE8C1, 0xE981, 0x2940, 0xEB01, 0x2BC0, 0x2A80, 0xEA41, ++ 0xEE01, 0x2EC0, 0x2F80, 0xEF41, 0x2D00, 0xEDC1, 0xEC81, 0x2C40, ++ 0xE401, 0x24C0, 0x2580, 0xE541, 0x2700, 0xE7C1, 0xE681, 0x2640, ++ 0x2200, 0xE2C1, 0xE381, 0x2340, 0xE101, 0x21C0, 0x2080, 0xE041, ++ 0xA001, 0x60C0, 0x6180, 0xA141, 0x6300, 0xA3C1, 0xA281, 0x6240, ++ 0x6600, 0xA6C1, 0xA781, 0x6740, 0xA501, 0x65C0, 0x6480, 0xA441, ++ 0x6C00, 0xACC1, 0xAD81, 0x6D40, 0xAF01, 0x6FC0, 0x6E80, 0xAE41, ++ 0xAA01, 0x6AC0, 0x6B80, 0xAB41, 0x6900, 0xA9C1, 0xA881, 0x6840, ++ 0x7800, 0xB8C1, 0xB981, 0x7940, 0xBB01, 0x7BC0, 0x7A80, 0xBA41, ++ 0xBE01, 0x7EC0, 0x7F80, 0xBF41, 0x7D00, 0xBDC1, 0xBC81, 0x7C40, ++ 0xB401, 0x74C0, 0x7580, 0xB541, 0x7700, 0xB7C1, 0xB681, 0x7640, ++ 0x7200, 0xB2C1, 0xB381, 0x7340, 0xB101, 0x71C0, 0x7080, 0xB041, ++ 0x5000, 0x90C1, 0x9181, 0x5140, 0x9301, 0x53C0, 0x5280, 0x9241, ++ 0x9601, 0x56C0, 0x5780, 0x9741, 0x5500, 0x95C1, 0x9481, 0x5440, ++ 0x9C01, 0x5CC0, 0x5D80, 0x9D41, 0x5F00, 0x9FC1, 0x9E81, 0x5E40, ++ 0x5A00, 0x9AC1, 0x9B81, 0x5B40, 0x9901, 0x59C0, 0x5880, 0x9841, ++ 0x8801, 0x48C0, 0x4980, 0x8941, 0x4B00, 0x8BC1, 0x8A81, 0x4A40, ++ 0x4E00, 0x8EC1, 0x8F81, 0x4F40, 0x8D01, 0x4DC0, 0x4C80, 0x8C41, ++ 0x4400, 0x84C1, 0x8581, 0x4540, 0x8701, 0x47C0, 0x4680, 0x8641, ++ 0x8201, 0x42C0, 0x4380, 0x8341, 0x4100, 0x81C1, 0x8081, 0x4040 ++}; ++ ++static inline __u16 crc16_byte(__u16 crc, const __u8 data) ++{ ++ return (crc >> 8) ^ crc16_table[(crc ^ data) & 0xff]; ++} ++ ++__u16 crc16(__u16 crc, __u8 const *buffer, size_t len) ++{ ++ while (len--) ++ crc = crc16_byte(crc, *buffer++); ++ return crc; ++} ++#endif ++ ++__le16 ext3_group_desc_csum(struct ext3_sb_info *sbi, __u32 block_group, ++ struct ext3_group_desc *gdp) ++{ ++ __u16 crc = 0; ++ ++ if (sbi->s_es->s_feature_ro_compat & ++ cpu_to_le32(EXT4_FEATURE_RO_COMPAT_GDT_CSUM)) { ++ int offset = offsetof(struct ext3_group_desc, bg_checksum); ++ __le32 le_group = cpu_to_le32(block_group); ++ ++ crc = crc16(~0, sbi->s_es->s_uuid, sizeof(sbi->s_es->s_uuid)); ++ crc = crc16(crc, (__u8 *)&le_group, sizeof(le_group)); ++ crc = crc16(crc, (__u8 *)gdp, offset); ++ offset += sizeof(gdp->bg_checksum); /* skip checksum */ ++ BUG_ON(offset != sizeof(*gdp)); /* XXX handle s_desc_size */ ++ /* for checksum of struct ext4_group_desc do the rest... ++ if (offset < sbi->s_es->s_desc_size) { ++ crc = crc16(crc, (__u8 *)gdp + offset, ++ sbi->s_es->s_desc_size - offset); ++ */ ++ } ++ ++ return cpu_to_le16(crc); ++} ++ ++int ext3_group_desc_csum_verify(struct ext3_sb_info *sbi, __u32 block_group, ++ struct ext3_group_desc *gdp) ++{ ++ if (gdp->bg_checksum != ext3_group_desc_csum(sbi, block_group, gdp)) ++ return 0; ++ ++ return 1; ++} ++ + /* Called at mount-time, super-block is locked */ + static int ext3_check_descriptors (struct super_block * sb) + { +@@ -1139,6 +1224,13 @@ static int ext3_check_descriptors (struc + le32_to_cpu(gdp->bg_inode_table)); + return 0; + } ++ if (!ext3_group_desc_csum_verify(sbi, i, gdp)) { ++ ext3_error(sb, __FUNCTION__, ++ "Checksum for group %d failed (%u!=%u)\n", i, ++ le16_to_cpu(ext3_group_desc_csum(sbi,i,gdp)), ++ le16_to_cpu(gdp->bg_checksum)); ++ return 0; ++ } + block += EXT3_BLOCKS_PER_GROUP(sb); + gdp++; + } +Index: linux-2.6.9-full/fs/ext3/group.h +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/group.h 2007-02-13 18:39:59.640066087 +0300 ++++ linux-2.6.9-full/fs/ext3/group.h 2007-03-28 18:36:16.000000000 +0400 +@@ -0,0 +1,29 @@ ++/* ++ * linux/fs/ext3/group.h ++ * ++ * Copyright (C) 2007 Cluster File Systems, Inc ++ * ++ * Author: Andreas Dilger <adilger@clusterfs.com> ++ */ ++ ++#ifndef _LINUX_EXT3_GROUP_H ++#define _LINUX_EXT3_GROUP_H ++#if defined(CONFIG_CRC16) || defined(CONFIG_CRC16_MODULE) ++#include <linux/crc16.h> ++#endif ++ ++extern __le16 ext3_group_desc_csum(struct ext3_sb_info *sbi, __u32 group, ++ struct ext3_group_desc *gdp); ++extern int ext3_group_desc_csum_verify(struct ext3_sb_info *sbi, __u32 group, ++ struct ext3_group_desc *gdp); ++struct buffer_head *read_block_bitmap(struct super_block *sb, ++ unsigned int block_group); ++extern unsigned ext3_init_block_bitmap(struct super_block *sb, ++ struct buffer_head *bh, int group, ++ struct ext3_group_desc *desc); ++#define ext3_free_blocks_after_init(sb, group, desc) \ ++ ext3_init_block_bitmap(sb, NULL, group, desc) ++extern unsigned ext3_init_inode_bitmap(struct super_block *sb, ++ struct buffer_head *bh, int group, ++ struct ext3_group_desc *desc); ++#endif /* _LINUX_EXT3_GROUP_H */ +Index: linux-2.6.9-full/fs/ext3/ialloc.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/ialloc.c 2007-03-28 18:35:38.000000000 +0400 ++++ linux-2.6.9-full/fs/ext3/ialloc.c 2007-03-28 18:36:16.000000000 +0400 +@@ -28,6 +28,7 @@ + + #include "xattr.h" + #include "acl.h" ++#include "group.h" + + /* + * ialloc.c contains the inodes allocation and deallocation routines +@@ -43,6 +44,52 @@ + * the free blocks count in the block. + */ + ++/* ++ * To avoid calling the atomic setbit hundreds or thousands of times, we only ++ * need to use it within a single byte (to ensure we get endianness right). ++ * We can use memset for the rest of the bitmap as there are no other users. ++ */ ++static void mark_bitmap_end(int start_bit, int end_bit, char *bitmap) ++{ ++ int i; ++ ++ if (start_bit >= end_bit) ++ return; ++ ++ ext3_debug("mark end bits +%d through +%d used\n", start_bit, end_bit); ++ for (i = start_bit; i < ((start_bit + 7) & ~7UL); i++) ++ ext3_set_bit(i, bitmap); ++ if (i < end_bit) ++ memset(bitmap + (i >> 3), 0xff, (end_bit - i) >> 3); ++} ++ ++/* Initializes an uninitialized inode bitmap */ ++unsigned ext3_init_inode_bitmap(struct super_block *sb, ++ struct buffer_head *bh, int block_group, ++ struct ext3_group_desc *gdp) ++{ ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ ++ J_ASSERT_BH(bh, buffer_locked(bh)); ++ ++ /* If checksum is bad mark all blocks and inodes use to prevent ++ * allocation, essentially implementing a per-group read-only flag. */ ++ if (!ext3_group_desc_csum_verify(sbi, block_group, gdp)) { ++ ext3_error(sb, __FUNCTION__, "Checksum bad for group %u\n", ++ block_group); ++ gdp->bg_free_blocks_count = 0; ++ gdp->bg_free_inodes_count = 0; ++ gdp->bg_itable_unused = 0; ++ memset(bh->b_data, 0xff, sb->s_blocksize); ++ return 0; ++ } ++ ++ memset(bh->b_data, 0, (EXT3_INODES_PER_GROUP(sb) + 7) / 8); ++ mark_bitmap_end(EXT3_INODES_PER_GROUP(sb), EXT3_BLOCKS_PER_GROUP(sb), ++ bh->b_data); ++ ++ return EXT3_INODES_PER_GROUP(sb); ++} + + /* + * Read the inode allocation bitmap for a given block_group, reading +@@ -59,8 +106,19 @@ read_inode_bitmap(struct super_block * s + desc = ext3_get_group_desc(sb, block_group, NULL); + if (!desc) + goto error_out; +- +- bh = sb_bread(sb, le32_to_cpu(desc->bg_inode_bitmap)); ++ if (desc->bg_flags & cpu_to_le16(EXT3_BG_INODE_UNINIT)) { ++ bh = sb_getblk(sb, le32_to_cpu(desc->bg_inode_bitmap)); ++ if (!buffer_uptodate(bh)) { ++ lock_buffer(bh); ++ if (!buffer_uptodate(bh)) { ++ ext3_init_inode_bitmap(sb, bh,block_group,desc); ++ set_buffer_uptodate(bh); ++ } ++ unlock_buffer(bh); ++ } ++ } else { ++ bh = sb_bread(sb, le32_to_cpu(desc->bg_inode_bitmap)); ++ } + if (!bh) + ext3_error(sb, "read_inode_bitmap", + "Cannot read inode bitmap - " +@@ -169,6 +227,8 @@ void ext3_free_inode (handle_t *handle, + if (is_directory) + gdp->bg_used_dirs_count = cpu_to_le16( + le16_to_cpu(gdp->bg_used_dirs_count) - 1); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi,block_group, ++ gdp); + spin_unlock(sb_bgl_lock(sbi, block_group)); + percpu_counter_inc(&sbi->s_freeinodes_counter); + if (is_directory) +@@ -453,7 +513,7 @@ struct inode *ext3_new_inode(handle_t *h + struct ext3_sb_info *sbi; + int err = 0; + struct inode *ret; +- int i; ++ int i, free = 0; + + /* Cannot create files in a deleted directory */ + if (!dir || !dir->i_nlink) +@@ -566,11 +626,13 @@ repeat_in_this_group: + goto out; + + got: +- ino += group * EXT3_INODES_PER_GROUP(sb) + 1; +- if (ino < EXT3_FIRST_INO(sb) || ino > le32_to_cpu(es->s_inodes_count)) { +- ext3_error (sb, "ext3_new_inode", +- "reserved inode or inode > inodes count - " +- "block_group = %d, inode=%lu", group, ino); ++ ino++; ++ if ((group == 0 && ino < EXT3_FIRST_INO(sb)) || ++ ino > EXT3_INODES_PER_GROUP(sb)) { ++ ext3_error(sb, __FUNCTION__, ++ "reserved inode or inode > inodes count - " ++ "block_group = %d, inode=%lu", group, ++ ino + group * EXT3_INODES_PER_GROUP(sb)); + err = -EIO; + goto fail; + } +@@ -578,13 +640,65 @@ got: + BUFFER_TRACE(bh2, "get_write_access"); + err = ext3_journal_get_write_access(handle, bh2); + if (err) goto fail; ++ ++ /* We may have to initialize the block bitmap if it isn't already */ ++ if (EXT3_HAS_RO_COMPAT_FEATURE(sb, EXT4_FEATURE_RO_COMPAT_GDT_CSUM) && ++ gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ struct buffer_head *block_bh = read_block_bitmap(sb, group); ++ ++ BUFFER_TRACE(block_bh, "get block bitmap access"); ++ err = ext3_journal_get_write_access(handle, block_bh); ++ if (err) { ++ brelse(block_bh); ++ goto fail; ++ } ++ ++ free = 0; ++ spin_lock(sb_bgl_lock(sbi, group)); ++ /* recheck and clear flag under lock if we still need to */ ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_BLOCK_UNINIT); ++ free = ext3_free_blocks_after_init(sb, group, gdp); ++ gdp->bg_free_blocks_count = cpu_to_le16(free); ++ } ++ spin_unlock(sb_bgl_lock(sbi, group)); ++ ++ /* Don't need to dirty bitmap block if we didn't change it */ ++ if (free) { ++ BUFFER_TRACE(block_bh, "dirty block bitmap"); ++ err = ext3_journal_dirty_metadata(handle, block_bh); ++ } ++ ++ brelse(block_bh); ++ if (err) ++ goto fail; ++ } ++ + spin_lock(sb_bgl_lock(sbi, group)); ++ /* If we didn't allocate from within the initialized part of the inode ++ * table then we need to initialize up to this inode. */ ++ if (EXT3_HAS_RO_COMPAT_FEATURE(sb, EXT4_FEATURE_RO_COMPAT_GDT_CSUM)) { ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_INODE_UNINIT)) { ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_INODE_UNINIT); ++ free = EXT3_INODES_PER_GROUP(sb); ++ } else { ++ free = EXT3_INODES_PER_GROUP(sb) - ++ le16_to_cpu(gdp->bg_itable_unused); ++ } ++ ++ if (ino > free) { ++ gdp->bg_itable_unused = ++ cpu_to_le16(EXT3_INODES_PER_GROUP(sb) - ino); ++ } ++ } ++ + gdp->bg_free_inodes_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_inodes_count) - 1); + if (S_ISDIR(mode)) { + gdp->bg_used_dirs_count = + cpu_to_le16(le16_to_cpu(gdp->bg_used_dirs_count) + 1); + } ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, group, gdp); + spin_unlock(sb_bgl_lock(sbi, group)); + BUFFER_TRACE(bh2, "call ext3_journal_dirty_metadata"); + err = ext3_journal_dirty_metadata(handle, bh2); +@@ -606,7 +720,7 @@ got: + inode->i_gid = current->fsgid; + inode->i_mode = mode; + +- inode->i_ino = ino; ++ inode->i_ino = ino + group * EXT3_INODES_PER_GROUP(sb); + /* This is the optimal IO size (for stat), not the fs block size */ + inode->i_blksize = PAGE_SIZE; + inode->i_blocks = 0; +Index: linux-2.6.9-full/fs/ext3/mballoc.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/mballoc.c 2007-03-28 15:42:45.000000000 +0400 ++++ linux-2.6.9-full/fs/ext3/mballoc.c 2007-03-28 18:36:16.000000000 +0400 +@@ -36,6 +36,8 @@ + #include <linux/seq_file.h> + #include <linux/version.h> + ++#include "group.h" ++ + /* + * MUSTDO: + * - test ext3_ext_search_left() and ext3_ext_search_right() +@@ -323,6 +325,7 @@ struct ext3_group_info { + unsigned long bb_state; + unsigned long bb_tid; + struct ext3_free_metadata *bb_md_cur; ++ struct ext3_group_desc *bb_gdp; + unsigned short bb_first_free; + unsigned short bb_free; + unsigned short bb_fragments; +@@ -928,10 +931,7 @@ static int ext3_mb_init_cache(struct pag + if (first_group + i >= EXT3_SB(sb)->s_groups_count) + break; + +- err = -EIO; +- desc = ext3_get_group_desc(sb, first_group + i, NULL); +- if (desc == NULL) +- goto out; ++ desc = EXT3_GROUP_INFO(sb, first_group + i)->bb_gdp; + + err = -ENOMEM; + bh[i] = sb_getblk(sb, le32_to_cpu(desc->bg_block_bitmap)); +@@ -946,7 +946,12 @@ static int ext3_mb_init_cache(struct pag + unlock_buffer(bh[i]); + continue; + } +- ++ if (desc->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ ext3_init_block_bitmap(sb, bh[i], first_group + i,desc); ++ set_buffer_uptodate(bh[i]); ++ unlock_buffer(bh[i]); ++ continue; ++ } + get_bh(bh[i]); + bh[i]->b_end_io = end_buffer_read_sync; + submit_bh(READ, bh[i]); +@@ -1703,6 +1708,10 @@ static int ext3_mb_good_group(struct ext + switch (cr) { + case 0: + BUG_ON(ac->ac_2order == 0); ++ /* If this group is uninitialized, skip it initially */ ++ if (grp->bb_gdp->bg_flags & ++ cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) ++ return 0; + bits = ac->ac_sb->s_blocksize_bits + 1; + for (i = ac->ac_2order; i <= bits; i++) + if (grp->bb_counters[i] > 0) +@@ -1796,7 +1805,9 @@ repeat: + } + + ac->ac_groups_scanned++; +- if (cr == 0) ++ if (cr == 0 || (e3b.bd_info->bb_gdp->bg_flags & ++ cpu_to_le16(EXT3_BG_BLOCK_UNINIT) && ++ ac->ac_2order != 0)) + ext3_mb_simple_scan_group(ac, &e3b); + else if (cr == 1 && ac->ac_g_ex.fe_len == sbi->s_stripe) + ext3_mb_scan_aligned(ac, &e3b); +@@ -2267,12 +2278,13 @@ int ext3_mb_init_backend(struct super_bl + i--; + goto err_freebuddy; + } ++ memset(meta_group_info[j], 0, len); + desc = ext3_get_group_desc(sb, i, NULL); ++ meta_group_info[j]->bb_gdp = desc; + if (desc == NULL) { + printk(KERN_ERR"EXT3-fs: can't read descriptor %u\n",i); + goto err_freebuddy; + } +- memset(meta_group_info[j], 0, len); + set_bit(EXT3_GROUP_INFO_NEED_INIT_BIT, + &meta_group_info[j]->bb_state); + +@@ -2936,9 +2948,17 @@ int ext3_mb_mark_diskspace_used(struct e + mb_set_bits(bitmap_bh->b_data, ac->ac_b_ex.fe_start, ac->ac_b_ex.fe_len); + + spin_lock(sb_bgl_lock(sbi, ac->ac_b_ex.fe_group)); ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_BLOCK_UNINIT); ++ gdp->bg_free_blocks_count = ++ cpu_to_le16(ext3_free_blocks_after_init(sb, ++ ac->ac_b_ex.fe_group, ++ gdp)); ++ } + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) + - ac->ac_b_ex.fe_len); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, ac->ac_b_ex.fe_group, gdp); + spin_unlock(sb_bgl_lock(sbi, ac->ac_b_ex.fe_group)); + percpu_counter_mod(&sbi->s_freeblocks_counter, - ac->ac_b_ex.fe_len); + +@@ -4303,6 +4323,7 @@ do_more: + spin_lock(sb_bgl_lock(sbi, block_group)); + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) + count); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, block_group, gdp); + spin_unlock(sb_bgl_lock(sbi, block_group)); + percpu_counter_mod(&sbi->s_freeblocks_counter, count); + +Index: linux-2.6.9-full/fs/ext3/balloc.c +=================================================================== +--- linux-2.6.9-full.orig/fs/ext3/balloc.c 2007-03-28 15:45:41.000000000 +0400 ++++ linux-2.6.9-full/fs/ext3/balloc.c 2007-03-28 18:36:16.000000000 +0400 +@@ -20,6 +20,7 @@ + #include <linux/quotaops.h> + #include <linux/buffer_head.h> + ++#include "group.h" + /* + * balloc.c contains the blocks allocation and deallocation routines + */ +@@ -73,6 +74,75 @@ struct ext3_group_desc * ext3_get_group_ + return gdp + desc; + } + ++/* Initializes an uninitialized block bitmap if given, and returns the ++ * number of blocks free in the group. */ ++unsigned ext3_init_block_bitmap(struct super_block *sb, struct buffer_head *bh, ++ int block_group, struct ext3_group_desc *gdp) ++{ ++ unsigned long start; ++ int bit, bit_max; ++ unsigned free_blocks; ++ struct ext3_sb_info *sbi = EXT3_SB(sb); ++ ++ if (bh) { ++ J_ASSERT_BH(bh, buffer_locked(bh)); ++ ++ /* If checksum is bad mark all blocks use to prevent allocation, ++ * essentially implementing a per-group read-only flag. */ ++ if (!ext3_group_desc_csum_verify(sbi, block_group, gdp)) { ++ ext3_error(sb, __FUNCTION__, ++ "Checksum bad for group %u\n", block_group); ++ gdp->bg_free_blocks_count = 0; ++ gdp->bg_free_inodes_count = 0; ++ gdp->bg_itable_unused = 0; ++ memset(bh->b_data, 0xff, sb->s_blocksize); ++ return 0; ++ } ++ memset(bh->b_data, 0, sb->s_blocksize); ++ } ++ ++ /* Check for superblock and gdt backups in this group */ ++ bit_max = ext3_bg_has_super(sb, block_group); ++ ++ if (!EXT3_HAS_INCOMPAT_FEATURE(sb, EXT3_FEATURE_INCOMPAT_META_BG) || ++ block_group < le32_to_cpu(sbi->s_es->s_first_meta_bg) * ++ sbi->s_desc_per_block) { ++ if (bit_max) { ++ bit_max += ext3_bg_num_gdb(sb, block_group); ++ bit_max +=le16_to_cpu(sbi->s_es->s_reserved_gdt_blocks); ++ } ++ } else { /* For META_BG_BLOCK_GROUPS */ ++ int group_rel = (block_group - ++ le32_to_cpu(sbi->s_es->s_first_meta_bg)) % ++ EXT3_DESC_PER_BLOCK(sb); ++ if (group_rel == 0 || group_rel == 1 || ++ (group_rel == EXT3_DESC_PER_BLOCK(sb) - 1)) ++ bit_max += 1; ++ } ++ ++ /* Last and first groups are always initialized */ ++ free_blocks = EXT3_BLOCKS_PER_GROUP(sb) - bit_max; ++ ++ if (bh) { ++ for (bit = 0; bit < bit_max; bit++) ++ ext3_set_bit(bit, bh->b_data); ++ ++ start = block_group * EXT3_BLOCKS_PER_GROUP(sb) + ++ le32_to_cpu(sbi->s_es->s_first_data_block); ++ ++ /* Set bits for block and inode bitmaps, and inode table */ ++ ext3_set_bit(le32_to_cpu(gdp->bg_block_bitmap) - start, ++ bh->b_data); ++ ext3_set_bit(le32_to_cpu(gdp->bg_inode_bitmap) - start, ++ bh->b_data); ++ for (bit = le32_to_cpu(gdp->bg_inode_table) - start, ++ bit_max = bit + sbi->s_itb_per_group; bit < bit_max; bit++) ++ ext3_set_bit(bit, bh->b_data); ++ } ++ ++ return free_blocks - sbi->s_itb_per_group - 2; ++} ++ + /* + * Read the bitmap for a given block_group, reading into the specified + * slot in the superblock's bitmap cache. +@@ -88,7 +158,19 @@ read_block_bitmap(struct super_block *sb + desc = ext3_get_group_desc (sb, block_group, NULL); + if (!desc) + goto error_out; +- bh = sb_bread(sb, le32_to_cpu(desc->bg_block_bitmap)); ++ if (desc->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) { ++ bh = sb_getblk(sb, le32_to_cpu(desc->bg_block_bitmap)); ++ if (!buffer_uptodate(bh)) { ++ lock_buffer(bh); ++ if (!buffer_uptodate(bh)) { ++ ext3_init_block_bitmap(sb, bh,block_group,desc); ++ set_buffer_uptodate(bh); ++ } ++ unlock_buffer(bh); ++ } ++ } else { ++ bh = sb_bread(sb, le32_to_cpu(desc->bg_block_bitmap)); ++ } + if (!bh) + ext3_error (sb, "read_block_bitmap", + "Cannot read block bitmap - " +@@ -429,6 +511,7 @@ do_more: + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) + + *pdquot_freed_blocks); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, block_group, gdp); + spin_unlock(sb_bgl_lock(sbi, block_group)); + percpu_counter_mod(&sbi->s_freeblocks_counter, count); + +@@ -1330,8 +1413,11 @@ allocated: + ret_block, goal_hits, goal_attempts); + + spin_lock(sb_bgl_lock(sbi, group_no)); ++ if (gdp->bg_flags & cpu_to_le16(EXT3_BG_BLOCK_UNINIT)) ++ gdp->bg_flags &= cpu_to_le16(~EXT3_BG_BLOCK_UNINIT); + gdp->bg_free_blocks_count = + cpu_to_le16(le16_to_cpu(gdp->bg_free_blocks_count) - 1); ++ gdp->bg_checksum = ext3_group_desc_csum(sbi, group_no, gdp); + spin_unlock(sb_bgl_lock(sbi, group_no)); + percpu_counter_mod(&sbi->s_freeblocks_counter, -1); + diff --git a/ldiskfs/kernel_patches/series/ldiskfs-2.6-rhel4.series b/ldiskfs/kernel_patches/series/ldiskfs-2.6-rhel4.series index ee07d11812531f35bb8c2eb35f0902957b993e5c..7c2246f2714b4467a6820d3af953b24c653c81f5 100644 --- a/ldiskfs/kernel_patches/series/ldiskfs-2.6-rhel4.series +++ b/ldiskfs/kernel_patches/series/ldiskfs-2.6-rhel4.series @@ -7,11 +7,15 @@ ext3-ea-in-inode-2.6-rhel4.patch export-ext3-2.6-rhel4.patch ext3-include-fixes-2.6-rhel4.patch ext3-extents-2.6.9-rhel4.patch -ext3-mballoc2-2.6.9-rhel4.patch +ext3-extents-fixes-2.6.9-rhel4.patch +ext3-extents-multiblock-directio-2.6.9-rhel4.patch +ext3-extents-search-2.6.9-rhel4.patch +ext3-mballoc3-core.patch +ext3-mballoc3-rhel4.patch ext3-nlinks-2.6.9.patch ext3-ialloc-2.6.patch ext3-lookup-dotdot-2.6.9.patch ext3-sector_t-overflow-2.6.9-rhel4.patch ext3-check-jbd-errors-2.6.9.patch +ext3-uninit-2.6.9.patch ext3-nanosecond-2.6-rhel4.patch -ext3-extents-bug11324.patch diff --git a/ldiskfs/kernel_patches/series/ldiskfs-2.6-sles10.series b/ldiskfs/kernel_patches/series/ldiskfs-2.6-sles10.series index e9f3f1fb5a26d81e1e90a540349d5c955e67e354..982bc45d7b873e7435a1f37567c9240eac41d301 100644 --- a/ldiskfs/kernel_patches/series/ldiskfs-2.6-sles10.series +++ b/ldiskfs/kernel_patches/series/ldiskfs-2.6-sles10.series @@ -5,11 +5,16 @@ ext3-map_inode_page-2.6-suse.patch export-ext3-2.6-rhel4.patch ext3-include-fixes-2.6-rhel4.patch ext3-extents-2.6.16-sles10.patch -ext3-mballoc2-2.6-fc5.patch +ext3-extents-fixes-2.6.9-rhel4.patch +ext3-extents-multiblock-directio-2.6.9-rhel4.patch +ext3-extents-search-2.6.9-rhel4.patch +ext3-mballoc3-core.patch +ext3-mballoc3-sles10.patch ext3-nlinks-2.6.9.patch ext3-ialloc-2.6.patch ext3-remove-cond_resched-calls-2.6.12.patch -ext3-filterdata-2.6.15.patch +ext3-filterdata-sles10.patch ext3-disable-write-bar-by-default-2.6-sles10.patch +ext3-uninit-2.6-sles10.patch ext3-nanosecond-2.6-sles10.patch ext3-inode-version-2.6-sles10.patch diff --git a/ldiskfs/kernel_patches/series/ldiskfs-2.6-suse.series b/ldiskfs/kernel_patches/series/ldiskfs-2.6-suse.series index e27e861c1bd42c8d8e695e94e17f699c46d35d69..ad01cd9b8953e0add15511501f2fe2905ddd85e2 100644 --- a/ldiskfs/kernel_patches/series/ldiskfs-2.6-suse.series +++ b/ldiskfs/kernel_patches/series/ldiskfs-2.6-suse.series @@ -7,11 +7,16 @@ ext3-ea-in-inode-2.6-suse.patch export-ext3-2.6-suse.patch ext3-include-fixes-2.6-suse.patch ext3-extents-2.6.5.patch -ext3-mballoc2-2.6-suse.patch +ext3-extents-fixes-2.6.9-rhel4.patch +ext3-extents-multiblock-directio-2.6.5-suse.patch +ext3-extents-search-2.6.9-rhel4.patch +ext3-mballoc3-core.patch +ext3-mballoc3-suse.patch ext3-nlinks-2.6.7.patch ext3-rename-reserve-2.6-suse.patch ext3-ialloc-2.6.patch ext3-lookup-dotdot-2.6.9.patch ext3-sector_t-overflow-2.6.5-suse.patch ext3-check-jbd-errors-2.6.5.patch +ext3-uninit-2.6-suse.patch ext3-nanosecond-2.6-suse.patch diff --git a/ldiskfs/ldiskfs/Makefile.in b/ldiskfs/ldiskfs/Makefile.in index 9f9e5154107bac4e5a93712d7bc565bccc9bc9f3..f945ff3de0b43d4ad8de6dad68ca0c762400789c 100644 --- a/ldiskfs/ldiskfs/Makefile.in +++ b/ldiskfs/ldiskfs/Makefile.in @@ -9,7 +9,7 @@ ext3_headers := $(wildcard @LINUX@/fs/ext3/*.h) linux_headers := $(wildcard @LINUX@/include/linux/ext3*.h) ext3_sources := $(filter-out %.mod.c,$(wildcard @LINUX@/fs/ext3/*.c)) -new_sources := iopen.c iopen.h extents.c mballoc.c +new_sources := iopen.c iopen.h extents.c mballoc.c group.h new_headers := ext3_extents.h ldiskfs_patched_sources := $(notdir $(ext3_sources) $(ext3_headers)) $(new_sources) $(new_headers) ldiskfs_sources := $(ldiskfs_patched_sources)