2005-04-17 06:20:36 +08:00
|
|
|
/*
|
2005-11-02 11:58:39 +08:00
|
|
|
* Copyright (c) 2000-2001,2005 Silicon Graphics, Inc.
|
|
|
|
* All Rights Reserved.
|
2005-04-17 06:20:36 +08:00
|
|
|
*
|
2005-11-02 11:58:39 +08:00
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License as
|
2005-04-17 06:20:36 +08:00
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*
|
2005-11-02 11:58:39 +08:00
|
|
|
* This program is distributed in the hope that it would be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
* GNU General Public License for more details.
|
2005-04-17 06:20:36 +08:00
|
|
|
*
|
2005-11-02 11:58:39 +08:00
|
|
|
* You should have received a copy of the GNU General Public License
|
|
|
|
* along with this program; if not, write the Free Software Foundation,
|
|
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
|
2005-04-17 06:20:36 +08:00
|
|
|
*/
|
|
|
|
#include "xfs.h"
|
2005-11-02 11:38:42 +08:00
|
|
|
#include "xfs_fs.h"
|
2005-04-17 06:20:36 +08:00
|
|
|
#include "xfs_types.h"
|
2005-11-02 11:38:42 +08:00
|
|
|
#include "xfs_bit.h"
|
2005-04-17 06:20:36 +08:00
|
|
|
#include "xfs_log.h"
|
2005-11-02 11:38:42 +08:00
|
|
|
#include "xfs_inum.h"
|
2005-04-17 06:20:36 +08:00
|
|
|
#include "xfs_trans.h"
|
|
|
|
#include "xfs_sb.h"
|
|
|
|
#include "xfs_ag.h"
|
2005-11-02 11:38:42 +08:00
|
|
|
#include "xfs_dir2.h"
|
2005-04-17 06:20:36 +08:00
|
|
|
#include "xfs_dmapi.h"
|
|
|
|
#include "xfs_mount.h"
|
2005-11-02 11:38:42 +08:00
|
|
|
#include "xfs_bmap_btree.h"
|
2005-04-17 06:20:36 +08:00
|
|
|
#include "xfs_alloc_btree.h"
|
|
|
|
#include "xfs_ialloc_btree.h"
|
2005-11-02 11:38:42 +08:00
|
|
|
#include "xfs_dir2_sf.h"
|
|
|
|
#include "xfs_attr_sf.h"
|
|
|
|
#include "xfs_dinode.h"
|
|
|
|
#include "xfs_inode.h"
|
2005-04-17 06:20:36 +08:00
|
|
|
#include "xfs_btree.h"
|
|
|
|
#include "xfs_ialloc.h"
|
|
|
|
#include "xfs_alloc.h"
|
|
|
|
#include "xfs_error.h"
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Prototypes for internal functions.
|
|
|
|
*/
|
|
|
|
|
|
|
|
STATIC void xfs_alloc_log_block(xfs_trans_t *, xfs_buf_t *, int);
|
|
|
|
STATIC void xfs_alloc_log_keys(xfs_btree_cur_t *, xfs_buf_t *, int, int);
|
|
|
|
STATIC void xfs_alloc_log_ptrs(xfs_btree_cur_t *, xfs_buf_t *, int, int);
|
|
|
|
STATIC void xfs_alloc_log_recs(xfs_btree_cur_t *, xfs_buf_t *, int, int);
|
|
|
|
STATIC int xfs_alloc_lshift(xfs_btree_cur_t *, int, int *);
|
|
|
|
STATIC int xfs_alloc_newroot(xfs_btree_cur_t *, int *);
|
|
|
|
STATIC int xfs_alloc_rshift(xfs_btree_cur_t *, int, int *);
|
|
|
|
STATIC int xfs_alloc_split(xfs_btree_cur_t *, int, xfs_agblock_t *,
|
|
|
|
xfs_alloc_key_t *, xfs_btree_cur_t **, int *);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Internal functions.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Single level of the xfs_alloc_delete record deletion routine.
|
|
|
|
* Delete record pointed to by cur/level.
|
|
|
|
* Remove the record from its block then rebalance the tree.
|
|
|
|
* Return 0 for error, 1 for done, 2 to go on to the next level.
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_alloc_delrec(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
int level, /* level removing record from */
|
|
|
|
int *stat) /* fail/done/go-on */
|
|
|
|
{
|
|
|
|
xfs_agf_t *agf; /* allocation group freelist header */
|
|
|
|
xfs_alloc_block_t *block; /* btree block record/key lives in */
|
|
|
|
xfs_agblock_t bno; /* btree block number */
|
|
|
|
xfs_buf_t *bp; /* buffer for block */
|
|
|
|
int error; /* error return value */
|
|
|
|
int i; /* loop index */
|
|
|
|
xfs_alloc_key_t key; /* kp points here if block is level 0 */
|
|
|
|
xfs_agblock_t lbno; /* left block's block number */
|
|
|
|
xfs_buf_t *lbp; /* left block's buffer pointer */
|
|
|
|
xfs_alloc_block_t *left; /* left btree block */
|
|
|
|
xfs_alloc_key_t *lkp=NULL; /* left block key pointer */
|
|
|
|
xfs_alloc_ptr_t *lpp=NULL; /* left block address pointer */
|
|
|
|
int lrecs=0; /* number of records in left block */
|
|
|
|
xfs_alloc_rec_t *lrp; /* left block record pointer */
|
|
|
|
xfs_mount_t *mp; /* mount structure */
|
|
|
|
int ptr; /* index in btree block for this rec */
|
|
|
|
xfs_agblock_t rbno; /* right block's block number */
|
|
|
|
xfs_buf_t *rbp; /* right block's buffer pointer */
|
|
|
|
xfs_alloc_block_t *right; /* right btree block */
|
|
|
|
xfs_alloc_key_t *rkp; /* right block key pointer */
|
|
|
|
xfs_alloc_ptr_t *rpp; /* right block address pointer */
|
|
|
|
int rrecs=0; /* number of records in right block */
|
2006-09-28 09:05:40 +08:00
|
|
|
int numrecs;
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_rec_t *rrp; /* right block record pointer */
|
|
|
|
xfs_btree_cur_t *tcur; /* temporary btree cursor */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Get the index of the entry being deleted, check for nothing there.
|
|
|
|
*/
|
|
|
|
ptr = cur->bc_ptrs[level];
|
|
|
|
if (ptr == 0) {
|
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Get the buffer & block containing the record or key/ptr.
|
|
|
|
*/
|
|
|
|
bp = cur->bc_bufs[level];
|
|
|
|
block = XFS_BUF_TO_ALLOC_BLOCK(bp);
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, block, level, bp)))
|
|
|
|
return error;
|
|
|
|
#endif
|
|
|
|
/*
|
|
|
|
* Fail if we're off the end of the block.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
numrecs = be16_to_cpu(block->bb_numrecs);
|
|
|
|
if (ptr > numrecs) {
|
2005-04-17 06:20:36 +08:00
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
XFS_STATS_INC(xs_abt_delrec);
|
|
|
|
/*
|
|
|
|
* It's a nonleaf. Excise the key and ptr being deleted, by
|
|
|
|
* sliding the entries past them down one.
|
|
|
|
* Log the changed areas of the block.
|
|
|
|
*/
|
|
|
|
if (level > 0) {
|
|
|
|
lkp = XFS_ALLOC_KEY_ADDR(block, 1, cur);
|
|
|
|
lpp = XFS_ALLOC_PTR_ADDR(block, 1, cur);
|
|
|
|
#ifdef DEBUG
|
2006-09-28 09:05:40 +08:00
|
|
|
for (i = ptr; i < numrecs; i++) {
|
2005-11-02 12:11:25 +08:00
|
|
|
if ((error = xfs_btree_check_sptr(cur, be32_to_cpu(lpp[i]), level)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
#endif
|
2006-09-28 09:05:40 +08:00
|
|
|
if (ptr < numrecs) {
|
2005-04-17 06:20:36 +08:00
|
|
|
memmove(&lkp[ptr - 1], &lkp[ptr],
|
2006-09-28 09:05:40 +08:00
|
|
|
(numrecs - ptr) * sizeof(*lkp));
|
2005-04-17 06:20:36 +08:00
|
|
|
memmove(&lpp[ptr - 1], &lpp[ptr],
|
2006-09-28 09:05:40 +08:00
|
|
|
(numrecs - ptr) * sizeof(*lpp));
|
|
|
|
xfs_alloc_log_ptrs(cur, bp, ptr, numrecs - 1);
|
|
|
|
xfs_alloc_log_keys(cur, bp, ptr, numrecs - 1);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* It's a leaf. Excise the record being deleted, by sliding the
|
|
|
|
* entries past it down one. Log the changed areas of the block.
|
|
|
|
*/
|
|
|
|
else {
|
|
|
|
lrp = XFS_ALLOC_REC_ADDR(block, 1, cur);
|
2006-09-28 09:05:40 +08:00
|
|
|
if (ptr < numrecs) {
|
2005-04-17 06:20:36 +08:00
|
|
|
memmove(&lrp[ptr - 1], &lrp[ptr],
|
2006-09-28 09:05:40 +08:00
|
|
|
(numrecs - ptr) * sizeof(*lrp));
|
|
|
|
xfs_alloc_log_recs(cur, bp, ptr, numrecs - 1);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If it's the first record in the block, we'll need a key
|
|
|
|
* structure to pass up to the next level (updkey).
|
|
|
|
*/
|
|
|
|
if (ptr == 1) {
|
2005-11-02 12:11:25 +08:00
|
|
|
key.ar_startblock = lrp->ar_startblock;
|
|
|
|
key.ar_blockcount = lrp->ar_blockcount;
|
2005-04-17 06:20:36 +08:00
|
|
|
lkp = &key;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Decrement and log the number of entries in the block.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
numrecs--;
|
|
|
|
block->bb_numrecs = cpu_to_be16(numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_block(cur->bc_tp, bp, XFS_BB_NUMRECS);
|
|
|
|
/*
|
|
|
|
* See if the longest free extent in the allocation group was
|
|
|
|
* changed by this operation. True if it's the by-size btree, and
|
|
|
|
* this is the leaf level, and there is no right sibling block,
|
|
|
|
* and this was the last record.
|
|
|
|
*/
|
|
|
|
agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp);
|
|
|
|
mp = cur->bc_mp;
|
|
|
|
|
|
|
|
if (level == 0 &&
|
|
|
|
cur->bc_btnum == XFS_BTNUM_CNT &&
|
2005-11-02 12:11:25 +08:00
|
|
|
be32_to_cpu(block->bb_rightsib) == NULLAGBLOCK &&
|
2006-09-28 09:05:40 +08:00
|
|
|
ptr > numrecs) {
|
|
|
|
ASSERT(ptr == numrecs + 1);
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* There are still records in the block. Grab the size
|
|
|
|
* from the last one.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
if (numrecs) {
|
|
|
|
rrp = XFS_ALLOC_REC_ADDR(block, numrecs, cur);
|
2005-11-02 12:11:25 +08:00
|
|
|
agf->agf_longest = rrp->ar_blockcount;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
/*
|
|
|
|
* No free extents left.
|
|
|
|
*/
|
|
|
|
else
|
|
|
|
agf->agf_longest = 0;
|
2005-11-02 12:11:25 +08:00
|
|
|
mp->m_perag[be32_to_cpu(agf->agf_seqno)].pagf_longest =
|
|
|
|
be32_to_cpu(agf->agf_longest);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_agf(cur->bc_tp, cur->bc_private.a.agbp,
|
|
|
|
XFS_AGF_LONGEST);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Is this the root level? If so, we're almost done.
|
|
|
|
*/
|
|
|
|
if (level == cur->bc_nlevels - 1) {
|
|
|
|
/*
|
|
|
|
* If this is the root level,
|
|
|
|
* and there's only one entry left,
|
|
|
|
* and it's NOT the leaf level,
|
|
|
|
* then we can get rid of this level.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
if (numrecs == 1 && level > 0) {
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* lpp is still set to the first pointer in the block.
|
|
|
|
* Make it the new root of the btree.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
bno = be32_to_cpu(agf->agf_roots[cur->bc_btnum]);
|
|
|
|
agf->agf_roots[cur->bc_btnum] = *lpp;
|
2008-02-14 07:03:29 +08:00
|
|
|
be32_add_cpu(&agf->agf_levels[cur->bc_btnum], -1);
|
2005-11-02 12:11:25 +08:00
|
|
|
mp->m_perag[be32_to_cpu(agf->agf_seqno)].pagf_levels[cur->bc_btnum]--;
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Put this buffer/block on the ag's freelist.
|
|
|
|
*/
|
[XFS] Lazy Superblock Counters
When we have a couple of hundred transactions on the fly at once, they all
typically modify the on disk superblock in some way.
create/unclink/mkdir/rmdir modify inode counts, allocation/freeing modify
free block counts.
When these counts are modified in a transaction, they must eventually lock
the superblock buffer and apply the mods. The buffer then remains locked
until the transaction is committed into the incore log buffer. The result
of this is that with enough transactions on the fly the incore superblock
buffer becomes a bottleneck.
The result of contention on the incore superblock buffer is that
transaction rates fall - the more pressure that is put on the superblock
buffer, the slower things go.
The key to removing the contention is to not require the superblock fields
in question to be locked. We do that by not marking the superblock dirty
in the transaction. IOWs, we modify the incore superblock but do not
modify the cached superblock buffer. In short, we do not log superblock
modifications to critical fields in the superblock on every transaction.
In fact we only do it just before we write the superblock to disk every
sync period or just before unmount.
This creates an interesting problem - if we don't log or write out the
fields in every transaction, then how do the values get recovered after a
crash? the answer is simple - we keep enough duplicate, logged information
in other structures that we can reconstruct the correct count after log
recovery has been performed.
It is the AGF and AGI structures that contain the duplicate information;
after recovery, we walk every AGI and AGF and sum their individual
counters to get the correct value, and we do a transaction into the log to
correct them. An optimisation of this is that if we have a clean unmount
record, we know the value in the superblock is correct, so we can avoid
the summation walk under normal conditions and so mount/recovery times do
not change under normal operation.
One wrinkle that was discovered during development was that the blocks
used in the freespace btrees are never accounted for in the AGF counters.
This was once a valid optimisation to make; when the filesystem is full,
the free space btrees are empty and consume no space. Hence when it
matters, the "accounting" is correct. But that means the when we do the
AGF summations, we would not have a correct count and xfs_check would
complain. Hence a new counter was added to track the number of blocks used
by the free space btrees. This is an *on-disk format change*.
As a result of this, lazy superblock counters are a mkfs option and at the
moment on linux there is no way to convert an old filesystem. This is
possible - xfs_db can be used to twiddle the right bits and then
xfs_repair will do the format conversion for you. Similarly, you can
convert backwards as well. At some point we'll add functionality to
xfs_admin to do the bit twiddling easily....
SGI-PV: 964999
SGI-Modid: xfs-linux-melb:xfs-kern:28652a
Signed-off-by: David Chinner <dgc@sgi.com>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Tim Shimmin <tes@sgi.com>
2007-05-24 13:26:31 +08:00
|
|
|
error = xfs_alloc_put_freelist(cur->bc_tp,
|
|
|
|
cur->bc_private.a.agbp, NULL, bno, 1);
|
|
|
|
if (error)
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* Since blocks move to the free list without the
|
|
|
|
* coordination used in xfs_bmap_finish, we can't allow
|
|
|
|
* block to be available for reallocation and
|
|
|
|
* non-transaction writing (user data) until we know
|
|
|
|
* that the transaction that moved it to the free list
|
|
|
|
* is permanently on disk. We track the blocks by
|
|
|
|
* declaring these blocks as "busy"; the busy list is
|
|
|
|
* maintained on a per-ag basis and each transaction
|
|
|
|
* records which entries should be removed when the
|
|
|
|
* iclog commits to disk. If a busy block is
|
|
|
|
* allocated, the iclog is pushed up to the LSN
|
|
|
|
* that freed the block.
|
|
|
|
*/
|
|
|
|
xfs_alloc_mark_busy(cur->bc_tp,
|
2005-11-02 12:11:25 +08:00
|
|
|
be32_to_cpu(agf->agf_seqno), bno, 1);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
xfs_trans_agbtree_delta(cur->bc_tp, -1);
|
|
|
|
xfs_alloc_log_agf(cur->bc_tp, cur->bc_private.a.agbp,
|
|
|
|
XFS_AGF_ROOTS | XFS_AGF_LEVELS);
|
|
|
|
/*
|
|
|
|
* Update the cursor so there's one fewer level.
|
|
|
|
*/
|
|
|
|
xfs_btree_setbuf(cur, level, NULL);
|
|
|
|
cur->bc_nlevels--;
|
|
|
|
} else if (level > 0 &&
|
2008-10-30 13:55:58 +08:00
|
|
|
(error = xfs_btree_decrement(cur, level, &i)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If we deleted the leftmost entry in the block, update the
|
|
|
|
* key values above us in the tree.
|
|
|
|
*/
|
2008-10-30 13:56:22 +08:00
|
|
|
if (ptr == 1 && (error = xfs_btree_updkey(cur, (union xfs_btree_key *)lkp, level + 1)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* If the number of records remaining in the block is at least
|
|
|
|
* the minimum, we're done.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
if (numrecs >= XFS_ALLOC_BLOCK_MINRECS(level, cur)) {
|
2008-10-30 13:55:58 +08:00
|
|
|
if (level > 0 && (error = xfs_btree_decrement(cur, level, &i)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Otherwise, we have to move some records around to keep the
|
|
|
|
* tree balanced. Look at the left and right sibling blocks to
|
|
|
|
* see if we can re-balance by moving only one record.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
rbno = be32_to_cpu(block->bb_rightsib);
|
|
|
|
lbno = be32_to_cpu(block->bb_leftsib);
|
2005-04-17 06:20:36 +08:00
|
|
|
bno = NULLAGBLOCK;
|
|
|
|
ASSERT(rbno != NULLAGBLOCK || lbno != NULLAGBLOCK);
|
|
|
|
/*
|
|
|
|
* Duplicate the cursor so our btree manipulations here won't
|
|
|
|
* disrupt the next level up.
|
|
|
|
*/
|
|
|
|
if ((error = xfs_btree_dup_cursor(cur, &tcur)))
|
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* If there's a right sibling, see if it's ok to shift an entry
|
|
|
|
* out of it.
|
|
|
|
*/
|
|
|
|
if (rbno != NULLAGBLOCK) {
|
|
|
|
/*
|
|
|
|
* Move the temp cursor to the last entry in the next block.
|
|
|
|
* Actually any entry but the first would suffice.
|
|
|
|
*/
|
|
|
|
i = xfs_btree_lastrec(tcur, level);
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(i == 1, error0);
|
2008-10-30 13:55:45 +08:00
|
|
|
if ((error = xfs_btree_increment(tcur, level, &i)))
|
2005-04-17 06:20:36 +08:00
|
|
|
goto error0;
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(i == 1, error0);
|
|
|
|
i = xfs_btree_lastrec(tcur, level);
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(i == 1, error0);
|
|
|
|
/*
|
|
|
|
* Grab a pointer to the block.
|
|
|
|
*/
|
|
|
|
rbp = tcur->bc_bufs[level];
|
|
|
|
right = XFS_BUF_TO_ALLOC_BLOCK(rbp);
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, right, level, rbp)))
|
|
|
|
goto error0;
|
|
|
|
#endif
|
|
|
|
/*
|
|
|
|
* Grab the current block number, for future use.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
bno = be32_to_cpu(right->bb_leftsib);
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* If right block is full enough so that removing one entry
|
|
|
|
* won't make it too empty, and left-shifting an entry out
|
|
|
|
* of right to us works, we're done.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (be16_to_cpu(right->bb_numrecs) - 1 >=
|
2005-04-17 06:20:36 +08:00
|
|
|
XFS_ALLOC_BLOCK_MINRECS(level, cur)) {
|
|
|
|
if ((error = xfs_alloc_lshift(tcur, level, &i)))
|
|
|
|
goto error0;
|
|
|
|
if (i) {
|
2005-11-02 12:11:25 +08:00
|
|
|
ASSERT(be16_to_cpu(block->bb_numrecs) >=
|
2005-04-17 06:20:36 +08:00
|
|
|
XFS_ALLOC_BLOCK_MINRECS(level, cur));
|
|
|
|
xfs_btree_del_cursor(tcur,
|
|
|
|
XFS_BTREE_NOERROR);
|
|
|
|
if (level > 0 &&
|
2008-10-30 13:55:58 +08:00
|
|
|
(error = xfs_btree_decrement(cur, level,
|
2005-04-17 06:20:36 +08:00
|
|
|
&i)))
|
|
|
|
return error;
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Otherwise, grab the number of records in right for
|
|
|
|
* future reference, and fix up the temp cursor to point
|
|
|
|
* to our block again (last record).
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
rrecs = be16_to_cpu(right->bb_numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
if (lbno != NULLAGBLOCK) {
|
|
|
|
i = xfs_btree_firstrec(tcur, level);
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(i == 1, error0);
|
2008-10-30 13:55:58 +08:00
|
|
|
if ((error = xfs_btree_decrement(tcur, level, &i)))
|
2005-04-17 06:20:36 +08:00
|
|
|
goto error0;
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(i == 1, error0);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If there's a left sibling, see if it's ok to shift an entry
|
|
|
|
* out of it.
|
|
|
|
*/
|
|
|
|
if (lbno != NULLAGBLOCK) {
|
|
|
|
/*
|
|
|
|
* Move the temp cursor to the first entry in the
|
|
|
|
* previous block.
|
|
|
|
*/
|
|
|
|
i = xfs_btree_firstrec(tcur, level);
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(i == 1, error0);
|
2008-10-30 13:55:58 +08:00
|
|
|
if ((error = xfs_btree_decrement(tcur, level, &i)))
|
2005-04-17 06:20:36 +08:00
|
|
|
goto error0;
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(i == 1, error0);
|
|
|
|
xfs_btree_firstrec(tcur, level);
|
|
|
|
/*
|
|
|
|
* Grab a pointer to the block.
|
|
|
|
*/
|
|
|
|
lbp = tcur->bc_bufs[level];
|
|
|
|
left = XFS_BUF_TO_ALLOC_BLOCK(lbp);
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, left, level, lbp)))
|
|
|
|
goto error0;
|
|
|
|
#endif
|
|
|
|
/*
|
|
|
|
* Grab the current block number, for future use.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
bno = be32_to_cpu(left->bb_rightsib);
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* If left block is full enough so that removing one entry
|
|
|
|
* won't make it too empty, and right-shifting an entry out
|
|
|
|
* of left to us works, we're done.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (be16_to_cpu(left->bb_numrecs) - 1 >=
|
2005-04-17 06:20:36 +08:00
|
|
|
XFS_ALLOC_BLOCK_MINRECS(level, cur)) {
|
|
|
|
if ((error = xfs_alloc_rshift(tcur, level, &i)))
|
|
|
|
goto error0;
|
|
|
|
if (i) {
|
2005-11-02 12:11:25 +08:00
|
|
|
ASSERT(be16_to_cpu(block->bb_numrecs) >=
|
2005-04-17 06:20:36 +08:00
|
|
|
XFS_ALLOC_BLOCK_MINRECS(level, cur));
|
|
|
|
xfs_btree_del_cursor(tcur,
|
|
|
|
XFS_BTREE_NOERROR);
|
|
|
|
if (level == 0)
|
|
|
|
cur->bc_ptrs[0]++;
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Otherwise, grab the number of records in right for
|
|
|
|
* future reference.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
lrecs = be16_to_cpu(left->bb_numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Delete the temp cursor, we're done with it.
|
|
|
|
*/
|
|
|
|
xfs_btree_del_cursor(tcur, XFS_BTREE_NOERROR);
|
|
|
|
/*
|
|
|
|
* If here, we need to do a join to keep the tree balanced.
|
|
|
|
*/
|
|
|
|
ASSERT(bno != NULLAGBLOCK);
|
|
|
|
/*
|
|
|
|
* See if we can join with the left neighbor block.
|
|
|
|
*/
|
|
|
|
if (lbno != NULLAGBLOCK &&
|
2006-09-28 09:05:40 +08:00
|
|
|
lrecs + numrecs <= XFS_ALLOC_BLOCK_MAXRECS(level, cur)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Set "right" to be the starting block,
|
|
|
|
* "left" to be the left neighbor.
|
|
|
|
*/
|
|
|
|
rbno = bno;
|
|
|
|
right = block;
|
2006-09-28 09:05:40 +08:00
|
|
|
rrecs = be16_to_cpu(right->bb_numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
rbp = bp;
|
|
|
|
if ((error = xfs_btree_read_bufs(mp, cur->bc_tp,
|
|
|
|
cur->bc_private.a.agno, lbno, 0, &lbp,
|
|
|
|
XFS_ALLOC_BTREE_REF)))
|
|
|
|
return error;
|
|
|
|
left = XFS_BUF_TO_ALLOC_BLOCK(lbp);
|
2006-09-28 09:05:40 +08:00
|
|
|
lrecs = be16_to_cpu(left->bb_numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
if ((error = xfs_btree_check_sblock(cur, left, level, lbp)))
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If that won't work, see if we can join with the right neighbor block.
|
|
|
|
*/
|
|
|
|
else if (rbno != NULLAGBLOCK &&
|
2006-09-28 09:05:40 +08:00
|
|
|
rrecs + numrecs <= XFS_ALLOC_BLOCK_MAXRECS(level, cur)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Set "left" to be the starting block,
|
|
|
|
* "right" to be the right neighbor.
|
|
|
|
*/
|
|
|
|
lbno = bno;
|
|
|
|
left = block;
|
2006-09-28 09:05:40 +08:00
|
|
|
lrecs = be16_to_cpu(left->bb_numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
lbp = bp;
|
|
|
|
if ((error = xfs_btree_read_bufs(mp, cur->bc_tp,
|
|
|
|
cur->bc_private.a.agno, rbno, 0, &rbp,
|
|
|
|
XFS_ALLOC_BTREE_REF)))
|
|
|
|
return error;
|
|
|
|
right = XFS_BUF_TO_ALLOC_BLOCK(rbp);
|
2006-09-28 09:05:40 +08:00
|
|
|
rrecs = be16_to_cpu(right->bb_numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
if ((error = xfs_btree_check_sblock(cur, right, level, rbp)))
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Otherwise, we can't fix the imbalance.
|
|
|
|
* Just return. This is probably a logic error, but it's not fatal.
|
|
|
|
*/
|
|
|
|
else {
|
2008-10-30 13:55:58 +08:00
|
|
|
if (level > 0 && (error = xfs_btree_decrement(cur, level, &i)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* We're now going to join "left" and "right" by moving all the stuff
|
|
|
|
* in "right" to "left" and deleting "right".
|
|
|
|
*/
|
|
|
|
if (level > 0) {
|
|
|
|
/*
|
|
|
|
* It's a non-leaf. Move keys and pointers.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
lkp = XFS_ALLOC_KEY_ADDR(left, lrecs + 1, cur);
|
|
|
|
lpp = XFS_ALLOC_PTR_ADDR(left, lrecs + 1, cur);
|
2005-04-17 06:20:36 +08:00
|
|
|
rkp = XFS_ALLOC_KEY_ADDR(right, 1, cur);
|
|
|
|
rpp = XFS_ALLOC_PTR_ADDR(right, 1, cur);
|
|
|
|
#ifdef DEBUG
|
2006-09-28 09:05:40 +08:00
|
|
|
for (i = 0; i < rrecs; i++) {
|
2005-11-02 12:11:25 +08:00
|
|
|
if ((error = xfs_btree_check_sptr(cur, be32_to_cpu(rpp[i]), level)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
#endif
|
2006-09-28 09:05:40 +08:00
|
|
|
memcpy(lkp, rkp, rrecs * sizeof(*lkp));
|
|
|
|
memcpy(lpp, rpp, rrecs * sizeof(*lpp));
|
|
|
|
xfs_alloc_log_keys(cur, lbp, lrecs + 1, lrecs + rrecs);
|
|
|
|
xfs_alloc_log_ptrs(cur, lbp, lrecs + 1, lrecs + rrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* It's a leaf. Move records.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
lrp = XFS_ALLOC_REC_ADDR(left, lrecs + 1, cur);
|
2005-04-17 06:20:36 +08:00
|
|
|
rrp = XFS_ALLOC_REC_ADDR(right, 1, cur);
|
2006-09-28 09:05:40 +08:00
|
|
|
memcpy(lrp, rrp, rrecs * sizeof(*lrp));
|
|
|
|
xfs_alloc_log_recs(cur, lbp, lrecs + 1, lrecs + rrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If we joined with the left neighbor, set the buffer in the
|
|
|
|
* cursor to the left block, and fix up the index.
|
|
|
|
*/
|
|
|
|
if (bp != lbp) {
|
|
|
|
xfs_btree_setbuf(cur, level, lbp);
|
2006-09-28 09:05:40 +08:00
|
|
|
cur->bc_ptrs[level] += lrecs;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If we joined with the right neighbor and there's a level above
|
|
|
|
* us, increment the cursor at that level.
|
|
|
|
*/
|
|
|
|
else if (level + 1 < cur->bc_nlevels &&
|
2008-10-30 13:55:45 +08:00
|
|
|
(error = xfs_btree_increment(cur, level + 1, &i)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* Fix up the number of records in the surviving block.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
lrecs += rrecs;
|
|
|
|
left->bb_numrecs = cpu_to_be16(lrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Fix up the right block pointer in the surviving block, and log it.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
left->bb_rightsib = right->bb_rightsib;
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_block(cur->bc_tp, lbp, XFS_BB_NUMRECS | XFS_BB_RIGHTSIB);
|
|
|
|
/*
|
|
|
|
* If there is a right sibling now, make it point to the
|
|
|
|
* remaining block.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (be32_to_cpu(left->bb_rightsib) != NULLAGBLOCK) {
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_block_t *rrblock;
|
|
|
|
xfs_buf_t *rrbp;
|
|
|
|
|
|
|
|
if ((error = xfs_btree_read_bufs(mp, cur->bc_tp,
|
2005-11-02 12:11:25 +08:00
|
|
|
cur->bc_private.a.agno, be32_to_cpu(left->bb_rightsib), 0,
|
2005-04-17 06:20:36 +08:00
|
|
|
&rrbp, XFS_ALLOC_BTREE_REF)))
|
|
|
|
return error;
|
|
|
|
rrblock = XFS_BUF_TO_ALLOC_BLOCK(rrbp);
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, rrblock, level, rrbp)))
|
|
|
|
return error;
|
2005-11-02 12:11:25 +08:00
|
|
|
rrblock->bb_leftsib = cpu_to_be32(lbno);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_block(cur->bc_tp, rrbp, XFS_BB_LEFTSIB);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Free the deleting block by putting it on the freelist.
|
|
|
|
*/
|
[XFS] Lazy Superblock Counters
When we have a couple of hundred transactions on the fly at once, they all
typically modify the on disk superblock in some way.
create/unclink/mkdir/rmdir modify inode counts, allocation/freeing modify
free block counts.
When these counts are modified in a transaction, they must eventually lock
the superblock buffer and apply the mods. The buffer then remains locked
until the transaction is committed into the incore log buffer. The result
of this is that with enough transactions on the fly the incore superblock
buffer becomes a bottleneck.
The result of contention on the incore superblock buffer is that
transaction rates fall - the more pressure that is put on the superblock
buffer, the slower things go.
The key to removing the contention is to not require the superblock fields
in question to be locked. We do that by not marking the superblock dirty
in the transaction. IOWs, we modify the incore superblock but do not
modify the cached superblock buffer. In short, we do not log superblock
modifications to critical fields in the superblock on every transaction.
In fact we only do it just before we write the superblock to disk every
sync period or just before unmount.
This creates an interesting problem - if we don't log or write out the
fields in every transaction, then how do the values get recovered after a
crash? the answer is simple - we keep enough duplicate, logged information
in other structures that we can reconstruct the correct count after log
recovery has been performed.
It is the AGF and AGI structures that contain the duplicate information;
after recovery, we walk every AGI and AGF and sum their individual
counters to get the correct value, and we do a transaction into the log to
correct them. An optimisation of this is that if we have a clean unmount
record, we know the value in the superblock is correct, so we can avoid
the summation walk under normal conditions and so mount/recovery times do
not change under normal operation.
One wrinkle that was discovered during development was that the blocks
used in the freespace btrees are never accounted for in the AGF counters.
This was once a valid optimisation to make; when the filesystem is full,
the free space btrees are empty and consume no space. Hence when it
matters, the "accounting" is correct. But that means the when we do the
AGF summations, we would not have a correct count and xfs_check would
complain. Hence a new counter was added to track the number of blocks used
by the free space btrees. This is an *on-disk format change*.
As a result of this, lazy superblock counters are a mkfs option and at the
moment on linux there is no way to convert an old filesystem. This is
possible - xfs_db can be used to twiddle the right bits and then
xfs_repair will do the format conversion for you. Similarly, you can
convert backwards as well. At some point we'll add functionality to
xfs_admin to do the bit twiddling easily....
SGI-PV: 964999
SGI-Modid: xfs-linux-melb:xfs-kern:28652a
Signed-off-by: David Chinner <dgc@sgi.com>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Tim Shimmin <tes@sgi.com>
2007-05-24 13:26:31 +08:00
|
|
|
error = xfs_alloc_put_freelist(cur->bc_tp,
|
|
|
|
cur->bc_private.a.agbp, NULL, rbno, 1);
|
|
|
|
if (error)
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* Since blocks move to the free list without the coordination
|
|
|
|
* used in xfs_bmap_finish, we can't allow block to be available
|
|
|
|
* for reallocation and non-transaction writing (user data)
|
|
|
|
* until we know that the transaction that moved it to the free
|
|
|
|
* list is permanently on disk. We track the blocks by declaring
|
|
|
|
* these blocks as "busy"; the busy list is maintained on a
|
|
|
|
* per-ag basis and each transaction records which entries
|
|
|
|
* should be removed when the iclog commits to disk. If a
|
|
|
|
* busy block is allocated, the iclog is pushed up to the
|
|
|
|
* LSN that freed the block.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
xfs_alloc_mark_busy(cur->bc_tp, be32_to_cpu(agf->agf_seqno), bno, 1);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_trans_agbtree_delta(cur->bc_tp, -1);
|
2005-11-02 12:11:25 +08:00
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Adjust the current level's cursor so that we're left referring
|
|
|
|
* to the right node, after we're done.
|
|
|
|
* If this leaves the ptr value 0 our caller will fix it up.
|
|
|
|
*/
|
|
|
|
if (level > 0)
|
|
|
|
cur->bc_ptrs[level]--;
|
|
|
|
/*
|
|
|
|
* Return value means the next level up has something to do.
|
|
|
|
*/
|
|
|
|
*stat = 2;
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
error0:
|
|
|
|
xfs_btree_del_cursor(tcur, XFS_BTREE_ERROR);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Insert one record/level. Return information to the caller
|
|
|
|
* allowing the next level up to proceed if necessary.
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_alloc_insrec(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
int level, /* level to insert record at */
|
|
|
|
xfs_agblock_t *bnop, /* i/o: block number inserted */
|
|
|
|
xfs_alloc_rec_t *recp, /* i/o: record data inserted */
|
|
|
|
xfs_btree_cur_t **curp, /* output: new cursor replacing cur */
|
|
|
|
int *stat) /* output: success/failure */
|
|
|
|
{
|
|
|
|
xfs_agf_t *agf; /* allocation group freelist header */
|
|
|
|
xfs_alloc_block_t *block; /* btree block record/key lives in */
|
|
|
|
xfs_buf_t *bp; /* buffer for block */
|
|
|
|
int error; /* error return value */
|
|
|
|
int i; /* loop index */
|
|
|
|
xfs_alloc_key_t key; /* key value being inserted */
|
|
|
|
xfs_alloc_key_t *kp; /* pointer to btree keys */
|
|
|
|
xfs_agblock_t nbno; /* block number of allocated block */
|
|
|
|
xfs_btree_cur_t *ncur; /* new cursor to be used at next lvl */
|
|
|
|
xfs_alloc_key_t nkey; /* new key value, from split */
|
|
|
|
xfs_alloc_rec_t nrec; /* new record value, for caller */
|
2006-09-28 09:05:40 +08:00
|
|
|
int numrecs;
|
2005-04-17 06:20:36 +08:00
|
|
|
int optr; /* old ptr value */
|
|
|
|
xfs_alloc_ptr_t *pp; /* pointer to btree addresses */
|
|
|
|
int ptr; /* index in btree block for this rec */
|
|
|
|
xfs_alloc_rec_t *rp; /* pointer to btree records */
|
|
|
|
|
2005-11-02 12:11:25 +08:00
|
|
|
ASSERT(be32_to_cpu(recp->ar_blockcount) > 0);
|
2005-11-02 12:06:18 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* GCC doesn't understand the (arguably complex) control flow in
|
|
|
|
* this function and complains about uninitialized structure fields
|
|
|
|
* without this.
|
|
|
|
*/
|
|
|
|
memset(&nrec, 0, sizeof(nrec));
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* If we made it to the root level, allocate a new root block
|
|
|
|
* and we're done.
|
|
|
|
*/
|
|
|
|
if (level >= cur->bc_nlevels) {
|
|
|
|
XFS_STATS_INC(xs_abt_insrec);
|
|
|
|
if ((error = xfs_alloc_newroot(cur, &i)))
|
|
|
|
return error;
|
|
|
|
*bnop = NULLAGBLOCK;
|
|
|
|
*stat = i;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Make a key out of the record data to be inserted, and save it.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
key.ar_startblock = recp->ar_startblock;
|
|
|
|
key.ar_blockcount = recp->ar_blockcount;
|
2005-04-17 06:20:36 +08:00
|
|
|
optr = ptr = cur->bc_ptrs[level];
|
|
|
|
/*
|
|
|
|
* If we're off the left edge, return failure.
|
|
|
|
*/
|
|
|
|
if (ptr == 0) {
|
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
XFS_STATS_INC(xs_abt_insrec);
|
|
|
|
/*
|
|
|
|
* Get pointers to the btree buffer and block.
|
|
|
|
*/
|
|
|
|
bp = cur->bc_bufs[level];
|
|
|
|
block = XFS_BUF_TO_ALLOC_BLOCK(bp);
|
2006-09-28 09:05:40 +08:00
|
|
|
numrecs = be16_to_cpu(block->bb_numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, block, level, bp)))
|
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* Check that the new entry is being inserted in the right place.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
if (ptr <= numrecs) {
|
2005-04-17 06:20:36 +08:00
|
|
|
if (level == 0) {
|
|
|
|
rp = XFS_ALLOC_REC_ADDR(block, ptr, cur);
|
|
|
|
xfs_btree_check_rec(cur->bc_btnum, recp, rp);
|
|
|
|
} else {
|
|
|
|
kp = XFS_ALLOC_KEY_ADDR(block, ptr, cur);
|
|
|
|
xfs_btree_check_key(cur->bc_btnum, &key, kp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
nbno = NULLAGBLOCK;
|
2006-09-28 08:58:40 +08:00
|
|
|
ncur = NULL;
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* If the block is full, we can't insert the new entry until we
|
|
|
|
* make the block un-full.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
if (numrecs == XFS_ALLOC_BLOCK_MAXRECS(level, cur)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* First, try shifting an entry to the right neighbor.
|
|
|
|
*/
|
|
|
|
if ((error = xfs_alloc_rshift(cur, level, &i)))
|
|
|
|
return error;
|
|
|
|
if (i) {
|
|
|
|
/* nothing */
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Next, try shifting an entry to the left neighbor.
|
|
|
|
*/
|
|
|
|
else {
|
|
|
|
if ((error = xfs_alloc_lshift(cur, level, &i)))
|
|
|
|
return error;
|
|
|
|
if (i)
|
|
|
|
optr = ptr = cur->bc_ptrs[level];
|
|
|
|
else {
|
|
|
|
/*
|
|
|
|
* Next, try splitting the current block in
|
|
|
|
* half. If this works we have to re-set our
|
|
|
|
* variables because we could be in a
|
|
|
|
* different block now.
|
|
|
|
*/
|
|
|
|
if ((error = xfs_alloc_split(cur, level, &nbno,
|
|
|
|
&nkey, &ncur, &i)))
|
|
|
|
return error;
|
|
|
|
if (i) {
|
|
|
|
bp = cur->bc_bufs[level];
|
|
|
|
block = XFS_BUF_TO_ALLOC_BLOCK(bp);
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error =
|
|
|
|
xfs_btree_check_sblock(cur,
|
|
|
|
block, level, bp)))
|
|
|
|
return error;
|
|
|
|
#endif
|
|
|
|
ptr = cur->bc_ptrs[level];
|
2005-11-02 12:11:25 +08:00
|
|
|
nrec.ar_startblock = nkey.ar_startblock;
|
|
|
|
nrec.ar_blockcount = nkey.ar_blockcount;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Otherwise the insert fails.
|
|
|
|
*/
|
|
|
|
else {
|
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* At this point we know there's room for our new entry in the block
|
|
|
|
* we're pointing at.
|
|
|
|
*/
|
2006-09-28 09:05:40 +08:00
|
|
|
numrecs = be16_to_cpu(block->bb_numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
if (level > 0) {
|
|
|
|
/*
|
|
|
|
* It's a non-leaf entry. Make a hole for the new data
|
|
|
|
* in the key and ptr regions of the block.
|
|
|
|
*/
|
|
|
|
kp = XFS_ALLOC_KEY_ADDR(block, 1, cur);
|
|
|
|
pp = XFS_ALLOC_PTR_ADDR(block, 1, cur);
|
|
|
|
#ifdef DEBUG
|
2006-09-28 09:05:40 +08:00
|
|
|
for (i = numrecs; i >= ptr; i--) {
|
2005-11-02 12:11:25 +08:00
|
|
|
if ((error = xfs_btree_check_sptr(cur, be32_to_cpu(pp[i - 1]), level)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
memmove(&kp[ptr], &kp[ptr - 1],
|
2006-09-28 09:05:40 +08:00
|
|
|
(numrecs - ptr + 1) * sizeof(*kp));
|
2005-04-17 06:20:36 +08:00
|
|
|
memmove(&pp[ptr], &pp[ptr - 1],
|
2006-09-28 09:05:40 +08:00
|
|
|
(numrecs - ptr + 1) * sizeof(*pp));
|
2005-04-17 06:20:36 +08:00
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sptr(cur, *bnop, level)))
|
|
|
|
return error;
|
|
|
|
#endif
|
|
|
|
/*
|
|
|
|
* Now stuff the new data in, bump numrecs and log the new data.
|
|
|
|
*/
|
|
|
|
kp[ptr - 1] = key;
|
2005-11-02 12:11:25 +08:00
|
|
|
pp[ptr - 1] = cpu_to_be32(*bnop);
|
2006-09-28 09:05:40 +08:00
|
|
|
numrecs++;
|
|
|
|
block->bb_numrecs = cpu_to_be16(numrecs);
|
|
|
|
xfs_alloc_log_keys(cur, bp, ptr, numrecs);
|
|
|
|
xfs_alloc_log_ptrs(cur, bp, ptr, numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
#ifdef DEBUG
|
2006-09-28 09:05:40 +08:00
|
|
|
if (ptr < numrecs)
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_btree_check_key(cur->bc_btnum, kp + ptr - 1,
|
|
|
|
kp + ptr);
|
|
|
|
#endif
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* It's a leaf entry. Make a hole for the new record.
|
|
|
|
*/
|
|
|
|
rp = XFS_ALLOC_REC_ADDR(block, 1, cur);
|
|
|
|
memmove(&rp[ptr], &rp[ptr - 1],
|
2006-09-28 09:05:40 +08:00
|
|
|
(numrecs - ptr + 1) * sizeof(*rp));
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Now stuff the new record in, bump numrecs
|
|
|
|
* and log the new data.
|
|
|
|
*/
|
2006-09-28 08:57:17 +08:00
|
|
|
rp[ptr - 1] = *recp;
|
2006-09-28 09:05:40 +08:00
|
|
|
numrecs++;
|
|
|
|
block->bb_numrecs = cpu_to_be16(numrecs);
|
|
|
|
xfs_alloc_log_recs(cur, bp, ptr, numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
#ifdef DEBUG
|
2006-09-28 09:05:40 +08:00
|
|
|
if (ptr < numrecs)
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_btree_check_rec(cur->bc_btnum, rp + ptr - 1,
|
|
|
|
rp + ptr);
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Log the new number of records in the btree header.
|
|
|
|
*/
|
|
|
|
xfs_alloc_log_block(cur->bc_tp, bp, XFS_BB_NUMRECS);
|
|
|
|
/*
|
|
|
|
* If we inserted at the start of a block, update the parents' keys.
|
|
|
|
*/
|
2008-10-30 13:56:22 +08:00
|
|
|
if (optr == 1 && (error = xfs_btree_updkey(cur, (union xfs_btree_key *)&key, level + 1)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* Look to see if the longest extent in the allocation group
|
|
|
|
* needs to be updated.
|
|
|
|
*/
|
|
|
|
|
|
|
|
agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp);
|
|
|
|
if (level == 0 &&
|
|
|
|
cur->bc_btnum == XFS_BTNUM_CNT &&
|
2005-11-02 12:11:25 +08:00
|
|
|
be32_to_cpu(block->bb_rightsib) == NULLAGBLOCK &&
|
|
|
|
be32_to_cpu(recp->ar_blockcount) > be32_to_cpu(agf->agf_longest)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* If this is a leaf in the by-size btree and there
|
|
|
|
* is no right sibling block and this block is bigger
|
|
|
|
* than the previous longest block, update it.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
agf->agf_longest = recp->ar_blockcount;
|
|
|
|
cur->bc_mp->m_perag[be32_to_cpu(agf->agf_seqno)].pagf_longest
|
|
|
|
= be32_to_cpu(recp->ar_blockcount);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_agf(cur->bc_tp, cur->bc_private.a.agbp,
|
|
|
|
XFS_AGF_LONGEST);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Return the new block number, if any.
|
|
|
|
* If there is one, give back a record value and a cursor too.
|
|
|
|
*/
|
|
|
|
*bnop = nbno;
|
|
|
|
if (nbno != NULLAGBLOCK) {
|
2006-09-28 08:57:17 +08:00
|
|
|
*recp = nrec;
|
|
|
|
*curp = ncur;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Log header fields from a btree block.
|
|
|
|
*/
|
|
|
|
STATIC void
|
|
|
|
xfs_alloc_log_block(
|
|
|
|
xfs_trans_t *tp, /* transaction pointer */
|
|
|
|
xfs_buf_t *bp, /* buffer containing btree block */
|
|
|
|
int fields) /* mask of fields: XFS_BB_... */
|
|
|
|
{
|
|
|
|
int first; /* first byte offset logged */
|
|
|
|
int last; /* last byte offset logged */
|
|
|
|
static const short offsets[] = { /* table of offsets */
|
|
|
|
offsetof(xfs_alloc_block_t, bb_magic),
|
|
|
|
offsetof(xfs_alloc_block_t, bb_level),
|
|
|
|
offsetof(xfs_alloc_block_t, bb_numrecs),
|
|
|
|
offsetof(xfs_alloc_block_t, bb_leftsib),
|
|
|
|
offsetof(xfs_alloc_block_t, bb_rightsib),
|
|
|
|
sizeof(xfs_alloc_block_t)
|
|
|
|
};
|
|
|
|
|
|
|
|
xfs_btree_offsets(fields, offsets, XFS_BB_NUM_BITS, &first, &last);
|
|
|
|
xfs_trans_log_buf(tp, bp, first, last);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Log keys from a btree block (nonleaf).
|
|
|
|
*/
|
|
|
|
STATIC void
|
|
|
|
xfs_alloc_log_keys(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
xfs_buf_t *bp, /* buffer containing btree block */
|
|
|
|
int kfirst, /* index of first key to log */
|
|
|
|
int klast) /* index of last key to log */
|
|
|
|
{
|
|
|
|
xfs_alloc_block_t *block; /* btree block to log from */
|
|
|
|
int first; /* first byte offset logged */
|
|
|
|
xfs_alloc_key_t *kp; /* key pointer in btree block */
|
|
|
|
int last; /* last byte offset logged */
|
|
|
|
|
|
|
|
block = XFS_BUF_TO_ALLOC_BLOCK(bp);
|
|
|
|
kp = XFS_ALLOC_KEY_ADDR(block, 1, cur);
|
|
|
|
first = (int)((xfs_caddr_t)&kp[kfirst - 1] - (xfs_caddr_t)block);
|
|
|
|
last = (int)(((xfs_caddr_t)&kp[klast] - 1) - (xfs_caddr_t)block);
|
|
|
|
xfs_trans_log_buf(cur->bc_tp, bp, first, last);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Log block pointer fields from a btree block (nonleaf).
|
|
|
|
*/
|
|
|
|
STATIC void
|
|
|
|
xfs_alloc_log_ptrs(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
xfs_buf_t *bp, /* buffer containing btree block */
|
|
|
|
int pfirst, /* index of first pointer to log */
|
|
|
|
int plast) /* index of last pointer to log */
|
|
|
|
{
|
|
|
|
xfs_alloc_block_t *block; /* btree block to log from */
|
|
|
|
int first; /* first byte offset logged */
|
|
|
|
int last; /* last byte offset logged */
|
|
|
|
xfs_alloc_ptr_t *pp; /* block-pointer pointer in btree blk */
|
|
|
|
|
|
|
|
block = XFS_BUF_TO_ALLOC_BLOCK(bp);
|
|
|
|
pp = XFS_ALLOC_PTR_ADDR(block, 1, cur);
|
|
|
|
first = (int)((xfs_caddr_t)&pp[pfirst - 1] - (xfs_caddr_t)block);
|
|
|
|
last = (int)(((xfs_caddr_t)&pp[plast] - 1) - (xfs_caddr_t)block);
|
|
|
|
xfs_trans_log_buf(cur->bc_tp, bp, first, last);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Log records from a btree block (leaf).
|
|
|
|
*/
|
|
|
|
STATIC void
|
|
|
|
xfs_alloc_log_recs(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
xfs_buf_t *bp, /* buffer containing btree block */
|
|
|
|
int rfirst, /* index of first record to log */
|
|
|
|
int rlast) /* index of last record to log */
|
|
|
|
{
|
|
|
|
xfs_alloc_block_t *block; /* btree block to log from */
|
|
|
|
int first; /* first byte offset logged */
|
|
|
|
int last; /* last byte offset logged */
|
|
|
|
xfs_alloc_rec_t *rp; /* record pointer for btree block */
|
|
|
|
|
|
|
|
|
|
|
|
block = XFS_BUF_TO_ALLOC_BLOCK(bp);
|
|
|
|
rp = XFS_ALLOC_REC_ADDR(block, 1, cur);
|
|
|
|
#ifdef DEBUG
|
|
|
|
{
|
|
|
|
xfs_agf_t *agf;
|
|
|
|
xfs_alloc_rec_t *p;
|
|
|
|
|
|
|
|
agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp);
|
|
|
|
for (p = &rp[rfirst - 1]; p <= &rp[rlast - 1]; p++)
|
2005-11-02 12:11:25 +08:00
|
|
|
ASSERT(be32_to_cpu(p->ar_startblock) +
|
|
|
|
be32_to_cpu(p->ar_blockcount) <=
|
|
|
|
be32_to_cpu(agf->agf_length));
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
first = (int)((xfs_caddr_t)&rp[rfirst - 1] - (xfs_caddr_t)block);
|
|
|
|
last = (int)(((xfs_caddr_t)&rp[rlast] - 1) - (xfs_caddr_t)block);
|
|
|
|
xfs_trans_log_buf(cur->bc_tp, bp, first, last);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Move 1 record left from cur/level if possible.
|
|
|
|
* Update cur to reflect the new path.
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_alloc_lshift(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
int level, /* level to shift record on */
|
|
|
|
int *stat) /* success/failure */
|
|
|
|
{
|
|
|
|
int error; /* error return value */
|
|
|
|
#ifdef DEBUG
|
|
|
|
int i; /* loop index */
|
|
|
|
#endif
|
|
|
|
xfs_alloc_key_t key; /* key value for leaf level upward */
|
|
|
|
xfs_buf_t *lbp; /* buffer for left neighbor block */
|
|
|
|
xfs_alloc_block_t *left; /* left neighbor btree block */
|
|
|
|
int nrec; /* new number of left block entries */
|
|
|
|
xfs_buf_t *rbp; /* buffer for right (current) block */
|
|
|
|
xfs_alloc_block_t *right; /* right (current) btree block */
|
|
|
|
xfs_alloc_key_t *rkp=NULL; /* key pointer for right block */
|
|
|
|
xfs_alloc_ptr_t *rpp=NULL; /* address pointer for right block */
|
|
|
|
xfs_alloc_rec_t *rrp=NULL; /* record pointer for right block */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Set up variables for this block as "right".
|
|
|
|
*/
|
|
|
|
rbp = cur->bc_bufs[level];
|
|
|
|
right = XFS_BUF_TO_ALLOC_BLOCK(rbp);
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, right, level, rbp)))
|
|
|
|
return error;
|
|
|
|
#endif
|
|
|
|
/*
|
|
|
|
* If we've got no left sibling then we can't shift an entry left.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (be32_to_cpu(right->bb_leftsib) == NULLAGBLOCK) {
|
2005-04-17 06:20:36 +08:00
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If the cursor entry is the one that would be moved, don't
|
|
|
|
* do it... it's too complicated.
|
|
|
|
*/
|
|
|
|
if (cur->bc_ptrs[level] <= 1) {
|
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Set up the left neighbor as "left".
|
|
|
|
*/
|
|
|
|
if ((error = xfs_btree_read_bufs(cur->bc_mp, cur->bc_tp,
|
2005-11-02 12:11:25 +08:00
|
|
|
cur->bc_private.a.agno, be32_to_cpu(right->bb_leftsib),
|
|
|
|
0, &lbp, XFS_ALLOC_BTREE_REF)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
left = XFS_BUF_TO_ALLOC_BLOCK(lbp);
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, left, level, lbp)))
|
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* If it's full, it can't take another entry.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (be16_to_cpu(left->bb_numrecs) == XFS_ALLOC_BLOCK_MAXRECS(level, cur)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
2005-11-02 12:11:25 +08:00
|
|
|
nrec = be16_to_cpu(left->bb_numrecs) + 1;
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* If non-leaf, copy a key and a ptr to the left block.
|
|
|
|
*/
|
|
|
|
if (level > 0) {
|
|
|
|
xfs_alloc_key_t *lkp; /* key pointer for left block */
|
|
|
|
xfs_alloc_ptr_t *lpp; /* address pointer for left block */
|
|
|
|
|
|
|
|
lkp = XFS_ALLOC_KEY_ADDR(left, nrec, cur);
|
|
|
|
rkp = XFS_ALLOC_KEY_ADDR(right, 1, cur);
|
|
|
|
*lkp = *rkp;
|
|
|
|
xfs_alloc_log_keys(cur, lbp, nrec, nrec);
|
|
|
|
lpp = XFS_ALLOC_PTR_ADDR(left, nrec, cur);
|
|
|
|
rpp = XFS_ALLOC_PTR_ADDR(right, 1, cur);
|
|
|
|
#ifdef DEBUG
|
2005-11-02 12:11:25 +08:00
|
|
|
if ((error = xfs_btree_check_sptr(cur, be32_to_cpu(*rpp), level)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
#endif
|
2006-09-28 08:57:17 +08:00
|
|
|
*lpp = *rpp;
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_ptrs(cur, lbp, nrec, nrec);
|
|
|
|
xfs_btree_check_key(cur->bc_btnum, lkp - 1, lkp);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If leaf, copy a record to the left block.
|
|
|
|
*/
|
|
|
|
else {
|
|
|
|
xfs_alloc_rec_t *lrp; /* record pointer for left block */
|
|
|
|
|
|
|
|
lrp = XFS_ALLOC_REC_ADDR(left, nrec, cur);
|
|
|
|
rrp = XFS_ALLOC_REC_ADDR(right, 1, cur);
|
|
|
|
*lrp = *rrp;
|
|
|
|
xfs_alloc_log_recs(cur, lbp, nrec, nrec);
|
|
|
|
xfs_btree_check_rec(cur->bc_btnum, lrp - 1, lrp);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Bump and log left's numrecs, decrement and log right's numrecs.
|
|
|
|
*/
|
2008-02-14 07:03:29 +08:00
|
|
|
be16_add_cpu(&left->bb_numrecs, 1);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_block(cur->bc_tp, lbp, XFS_BB_NUMRECS);
|
2008-02-14 07:03:29 +08:00
|
|
|
be16_add_cpu(&right->bb_numrecs, -1);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_block(cur->bc_tp, rbp, XFS_BB_NUMRECS);
|
|
|
|
/*
|
|
|
|
* Slide the contents of right down one entry.
|
|
|
|
*/
|
|
|
|
if (level > 0) {
|
|
|
|
#ifdef DEBUG
|
2005-11-02 12:11:25 +08:00
|
|
|
for (i = 0; i < be16_to_cpu(right->bb_numrecs); i++) {
|
|
|
|
if ((error = xfs_btree_check_sptr(cur, be32_to_cpu(rpp[i + 1]),
|
2005-04-17 06:20:36 +08:00
|
|
|
level)))
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
#endif
|
2005-11-02 12:11:25 +08:00
|
|
|
memmove(rkp, rkp + 1, be16_to_cpu(right->bb_numrecs) * sizeof(*rkp));
|
|
|
|
memmove(rpp, rpp + 1, be16_to_cpu(right->bb_numrecs) * sizeof(*rpp));
|
|
|
|
xfs_alloc_log_keys(cur, rbp, 1, be16_to_cpu(right->bb_numrecs));
|
|
|
|
xfs_alloc_log_ptrs(cur, rbp, 1, be16_to_cpu(right->bb_numrecs));
|
2005-04-17 06:20:36 +08:00
|
|
|
} else {
|
2005-11-02 12:11:25 +08:00
|
|
|
memmove(rrp, rrp + 1, be16_to_cpu(right->bb_numrecs) * sizeof(*rrp));
|
|
|
|
xfs_alloc_log_recs(cur, rbp, 1, be16_to_cpu(right->bb_numrecs));
|
|
|
|
key.ar_startblock = rrp->ar_startblock;
|
|
|
|
key.ar_blockcount = rrp->ar_blockcount;
|
2005-04-17 06:20:36 +08:00
|
|
|
rkp = &key;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Update the parent key values of right.
|
|
|
|
*/
|
2008-10-30 13:56:22 +08:00
|
|
|
if ((error = xfs_btree_updkey(cur, (union xfs_btree_key *)rkp, level + 1)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* Slide the cursor value left one.
|
|
|
|
*/
|
|
|
|
cur->bc_ptrs[level]--;
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate a new root block, fill it in.
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_alloc_newroot(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
int *stat) /* success/failure */
|
|
|
|
{
|
|
|
|
int error; /* error return value */
|
|
|
|
xfs_agblock_t lbno; /* left block number */
|
|
|
|
xfs_buf_t *lbp; /* left btree buffer */
|
|
|
|
xfs_alloc_block_t *left; /* left btree block */
|
|
|
|
xfs_mount_t *mp; /* mount structure */
|
|
|
|
xfs_agblock_t nbno; /* new block number */
|
|
|
|
xfs_buf_t *nbp; /* new (root) buffer */
|
|
|
|
xfs_alloc_block_t *new; /* new (root) btree block */
|
|
|
|
int nptr; /* new value for key index, 1 or 2 */
|
|
|
|
xfs_agblock_t rbno; /* right block number */
|
|
|
|
xfs_buf_t *rbp; /* right btree buffer */
|
|
|
|
xfs_alloc_block_t *right; /* right btree block */
|
|
|
|
|
|
|
|
mp = cur->bc_mp;
|
|
|
|
|
|
|
|
ASSERT(cur->bc_nlevels < XFS_AG_MAXLEVELS(mp));
|
|
|
|
/*
|
|
|
|
* Get a buffer from the freelist blocks, for the new root.
|
|
|
|
*/
|
[XFS] Lazy Superblock Counters
When we have a couple of hundred transactions on the fly at once, they all
typically modify the on disk superblock in some way.
create/unclink/mkdir/rmdir modify inode counts, allocation/freeing modify
free block counts.
When these counts are modified in a transaction, they must eventually lock
the superblock buffer and apply the mods. The buffer then remains locked
until the transaction is committed into the incore log buffer. The result
of this is that with enough transactions on the fly the incore superblock
buffer becomes a bottleneck.
The result of contention on the incore superblock buffer is that
transaction rates fall - the more pressure that is put on the superblock
buffer, the slower things go.
The key to removing the contention is to not require the superblock fields
in question to be locked. We do that by not marking the superblock dirty
in the transaction. IOWs, we modify the incore superblock but do not
modify the cached superblock buffer. In short, we do not log superblock
modifications to critical fields in the superblock on every transaction.
In fact we only do it just before we write the superblock to disk every
sync period or just before unmount.
This creates an interesting problem - if we don't log or write out the
fields in every transaction, then how do the values get recovered after a
crash? the answer is simple - we keep enough duplicate, logged information
in other structures that we can reconstruct the correct count after log
recovery has been performed.
It is the AGF and AGI structures that contain the duplicate information;
after recovery, we walk every AGI and AGF and sum their individual
counters to get the correct value, and we do a transaction into the log to
correct them. An optimisation of this is that if we have a clean unmount
record, we know the value in the superblock is correct, so we can avoid
the summation walk under normal conditions and so mount/recovery times do
not change under normal operation.
One wrinkle that was discovered during development was that the blocks
used in the freespace btrees are never accounted for in the AGF counters.
This was once a valid optimisation to make; when the filesystem is full,
the free space btrees are empty and consume no space. Hence when it
matters, the "accounting" is correct. But that means the when we do the
AGF summations, we would not have a correct count and xfs_check would
complain. Hence a new counter was added to track the number of blocks used
by the free space btrees. This is an *on-disk format change*.
As a result of this, lazy superblock counters are a mkfs option and at the
moment on linux there is no way to convert an old filesystem. This is
possible - xfs_db can be used to twiddle the right bits and then
xfs_repair will do the format conversion for you. Similarly, you can
convert backwards as well. At some point we'll add functionality to
xfs_admin to do the bit twiddling easily....
SGI-PV: 964999
SGI-Modid: xfs-linux-melb:xfs-kern:28652a
Signed-off-by: David Chinner <dgc@sgi.com>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Tim Shimmin <tes@sgi.com>
2007-05-24 13:26:31 +08:00
|
|
|
error = xfs_alloc_get_freelist(cur->bc_tp,
|
|
|
|
cur->bc_private.a.agbp, &nbno, 1);
|
|
|
|
if (error)
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* None available, we fail.
|
|
|
|
*/
|
|
|
|
if (nbno == NULLAGBLOCK) {
|
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
xfs_trans_agbtree_delta(cur->bc_tp, 1);
|
|
|
|
nbp = xfs_btree_get_bufs(mp, cur->bc_tp, cur->bc_private.a.agno, nbno,
|
|
|
|
0);
|
|
|
|
new = XFS_BUF_TO_ALLOC_BLOCK(nbp);
|
|
|
|
/*
|
|
|
|
* Set the root data in the a.g. freespace structure.
|
|
|
|
*/
|
|
|
|
{
|
|
|
|
xfs_agf_t *agf; /* a.g. freespace header */
|
|
|
|
xfs_agnumber_t seqno;
|
|
|
|
|
|
|
|
agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp);
|
2005-11-02 12:11:25 +08:00
|
|
|
agf->agf_roots[cur->bc_btnum] = cpu_to_be32(nbno);
|
2008-02-14 07:03:29 +08:00
|
|
|
be32_add_cpu(&agf->agf_levels[cur->bc_btnum], 1);
|
2005-11-02 12:11:25 +08:00
|
|
|
seqno = be32_to_cpu(agf->agf_seqno);
|
2005-04-17 06:20:36 +08:00
|
|
|
mp->m_perag[seqno].pagf_levels[cur->bc_btnum]++;
|
|
|
|
xfs_alloc_log_agf(cur->bc_tp, cur->bc_private.a.agbp,
|
|
|
|
XFS_AGF_ROOTS | XFS_AGF_LEVELS);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* At the previous root level there are now two blocks: the old
|
|
|
|
* root, and the new block generated when it was split.
|
|
|
|
* We don't know which one the cursor is pointing at, so we
|
|
|
|
* set up variables "left" and "right" for each case.
|
|
|
|
*/
|
|
|
|
lbp = cur->bc_bufs[cur->bc_nlevels - 1];
|
|
|
|
left = XFS_BUF_TO_ALLOC_BLOCK(lbp);
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, left, cur->bc_nlevels - 1, lbp)))
|
|
|
|
return error;
|
|
|
|
#endif
|
2005-11-02 12:11:25 +08:00
|
|
|
if (be32_to_cpu(left->bb_rightsib) != NULLAGBLOCK) {
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Our block is left, pick up the right block.
|
|
|
|
*/
|
|
|
|
lbno = XFS_DADDR_TO_AGBNO(mp, XFS_BUF_ADDR(lbp));
|
2005-11-02 12:11:25 +08:00
|
|
|
rbno = be32_to_cpu(left->bb_rightsib);
|
2005-04-17 06:20:36 +08:00
|
|
|
if ((error = xfs_btree_read_bufs(mp, cur->bc_tp,
|
|
|
|
cur->bc_private.a.agno, rbno, 0, &rbp,
|
|
|
|
XFS_ALLOC_BTREE_REF)))
|
|
|
|
return error;
|
|
|
|
right = XFS_BUF_TO_ALLOC_BLOCK(rbp);
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, right,
|
|
|
|
cur->bc_nlevels - 1, rbp)))
|
|
|
|
return error;
|
|
|
|
nptr = 1;
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* Our block is right, pick up the left block.
|
|
|
|
*/
|
|
|
|
rbp = lbp;
|
|
|
|
right = left;
|
|
|
|
rbno = XFS_DADDR_TO_AGBNO(mp, XFS_BUF_ADDR(rbp));
|
2005-11-02 12:11:25 +08:00
|
|
|
lbno = be32_to_cpu(right->bb_leftsib);
|
2005-04-17 06:20:36 +08:00
|
|
|
if ((error = xfs_btree_read_bufs(mp, cur->bc_tp,
|
|
|
|
cur->bc_private.a.agno, lbno, 0, &lbp,
|
|
|
|
XFS_ALLOC_BTREE_REF)))
|
|
|
|
return error;
|
|
|
|
left = XFS_BUF_TO_ALLOC_BLOCK(lbp);
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, left,
|
|
|
|
cur->bc_nlevels - 1, lbp)))
|
|
|
|
return error;
|
|
|
|
nptr = 2;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Fill in the new block's btree header and log it.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
new->bb_magic = cpu_to_be32(xfs_magics[cur->bc_btnum]);
|
|
|
|
new->bb_level = cpu_to_be16(cur->bc_nlevels);
|
|
|
|
new->bb_numrecs = cpu_to_be16(2);
|
|
|
|
new->bb_leftsib = cpu_to_be32(NULLAGBLOCK);
|
|
|
|
new->bb_rightsib = cpu_to_be32(NULLAGBLOCK);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_block(cur->bc_tp, nbp, XFS_BB_ALL_BITS);
|
|
|
|
ASSERT(lbno != NULLAGBLOCK && rbno != NULLAGBLOCK);
|
|
|
|
/*
|
|
|
|
* Fill in the key data in the new root.
|
|
|
|
*/
|
|
|
|
{
|
|
|
|
xfs_alloc_key_t *kp; /* btree key pointer */
|
|
|
|
|
|
|
|
kp = XFS_ALLOC_KEY_ADDR(new, 1, cur);
|
2005-11-02 12:11:25 +08:00
|
|
|
if (be16_to_cpu(left->bb_level) > 0) {
|
2006-09-28 08:57:17 +08:00
|
|
|
kp[0] = *XFS_ALLOC_KEY_ADDR(left, 1, cur);
|
|
|
|
kp[1] = *XFS_ALLOC_KEY_ADDR(right, 1, cur);
|
2005-04-17 06:20:36 +08:00
|
|
|
} else {
|
|
|
|
xfs_alloc_rec_t *rp; /* btree record pointer */
|
|
|
|
|
|
|
|
rp = XFS_ALLOC_REC_ADDR(left, 1, cur);
|
2005-11-02 12:11:25 +08:00
|
|
|
kp[0].ar_startblock = rp->ar_startblock;
|
|
|
|
kp[0].ar_blockcount = rp->ar_blockcount;
|
2005-04-17 06:20:36 +08:00
|
|
|
rp = XFS_ALLOC_REC_ADDR(right, 1, cur);
|
2005-11-02 12:11:25 +08:00
|
|
|
kp[1].ar_startblock = rp->ar_startblock;
|
|
|
|
kp[1].ar_blockcount = rp->ar_blockcount;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
xfs_alloc_log_keys(cur, nbp, 1, 2);
|
|
|
|
/*
|
|
|
|
* Fill in the pointer data in the new root.
|
|
|
|
*/
|
|
|
|
{
|
|
|
|
xfs_alloc_ptr_t *pp; /* btree address pointer */
|
|
|
|
|
|
|
|
pp = XFS_ALLOC_PTR_ADDR(new, 1, cur);
|
2005-11-02 12:11:25 +08:00
|
|
|
pp[0] = cpu_to_be32(lbno);
|
|
|
|
pp[1] = cpu_to_be32(rbno);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
xfs_alloc_log_ptrs(cur, nbp, 1, 2);
|
|
|
|
/*
|
|
|
|
* Fix up the cursor.
|
|
|
|
*/
|
|
|
|
xfs_btree_setbuf(cur, cur->bc_nlevels, nbp);
|
|
|
|
cur->bc_ptrs[cur->bc_nlevels] = nptr;
|
|
|
|
cur->bc_nlevels++;
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Move 1 record right from cur/level if possible.
|
|
|
|
* Update cur to reflect the new path.
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_alloc_rshift(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
int level, /* level to shift record on */
|
|
|
|
int *stat) /* success/failure */
|
|
|
|
{
|
|
|
|
int error; /* error return value */
|
|
|
|
int i; /* loop index */
|
|
|
|
xfs_alloc_key_t key; /* key value for leaf level upward */
|
|
|
|
xfs_buf_t *lbp; /* buffer for left (current) block */
|
|
|
|
xfs_alloc_block_t *left; /* left (current) btree block */
|
|
|
|
xfs_buf_t *rbp; /* buffer for right neighbor block */
|
|
|
|
xfs_alloc_block_t *right; /* right neighbor btree block */
|
|
|
|
xfs_alloc_key_t *rkp; /* key pointer for right block */
|
|
|
|
xfs_btree_cur_t *tcur; /* temporary cursor */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Set up variables for this block as "left".
|
|
|
|
*/
|
|
|
|
lbp = cur->bc_bufs[level];
|
|
|
|
left = XFS_BUF_TO_ALLOC_BLOCK(lbp);
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, left, level, lbp)))
|
|
|
|
return error;
|
|
|
|
#endif
|
|
|
|
/*
|
|
|
|
* If we've got no right sibling then we can't shift an entry right.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (be32_to_cpu(left->bb_rightsib) == NULLAGBLOCK) {
|
2005-04-17 06:20:36 +08:00
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If the cursor entry is the one that would be moved, don't
|
|
|
|
* do it... it's too complicated.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (cur->bc_ptrs[level] >= be16_to_cpu(left->bb_numrecs)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Set up the right neighbor as "right".
|
|
|
|
*/
|
|
|
|
if ((error = xfs_btree_read_bufs(cur->bc_mp, cur->bc_tp,
|
2005-11-02 12:11:25 +08:00
|
|
|
cur->bc_private.a.agno, be32_to_cpu(left->bb_rightsib),
|
|
|
|
0, &rbp, XFS_ALLOC_BTREE_REF)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
right = XFS_BUF_TO_ALLOC_BLOCK(rbp);
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, right, level, rbp)))
|
|
|
|
return error;
|
|
|
|
/*
|
|
|
|
* If it's full, it can't take another entry.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (be16_to_cpu(right->bb_numrecs) == XFS_ALLOC_BLOCK_MAXRECS(level, cur)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Make a hole at the start of the right neighbor block, then
|
|
|
|
* copy the last left block entry to the hole.
|
|
|
|
*/
|
|
|
|
if (level > 0) {
|
|
|
|
xfs_alloc_key_t *lkp; /* key pointer for left block */
|
|
|
|
xfs_alloc_ptr_t *lpp; /* address pointer for left block */
|
|
|
|
xfs_alloc_ptr_t *rpp; /* address pointer for right block */
|
|
|
|
|
2005-11-02 12:11:25 +08:00
|
|
|
lkp = XFS_ALLOC_KEY_ADDR(left, be16_to_cpu(left->bb_numrecs), cur);
|
|
|
|
lpp = XFS_ALLOC_PTR_ADDR(left, be16_to_cpu(left->bb_numrecs), cur);
|
2005-04-17 06:20:36 +08:00
|
|
|
rkp = XFS_ALLOC_KEY_ADDR(right, 1, cur);
|
|
|
|
rpp = XFS_ALLOC_PTR_ADDR(right, 1, cur);
|
|
|
|
#ifdef DEBUG
|
2005-11-02 12:11:25 +08:00
|
|
|
for (i = be16_to_cpu(right->bb_numrecs) - 1; i >= 0; i--) {
|
|
|
|
if ((error = xfs_btree_check_sptr(cur, be32_to_cpu(rpp[i]), level)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
#endif
|
2005-11-02 12:11:25 +08:00
|
|
|
memmove(rkp + 1, rkp, be16_to_cpu(right->bb_numrecs) * sizeof(*rkp));
|
|
|
|
memmove(rpp + 1, rpp, be16_to_cpu(right->bb_numrecs) * sizeof(*rpp));
|
2005-04-17 06:20:36 +08:00
|
|
|
#ifdef DEBUG
|
2005-11-02 12:11:25 +08:00
|
|
|
if ((error = xfs_btree_check_sptr(cur, be32_to_cpu(*lpp), level)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
#endif
|
2006-09-28 08:57:17 +08:00
|
|
|
*rkp = *lkp;
|
|
|
|
*rpp = *lpp;
|
2005-11-02 12:11:25 +08:00
|
|
|
xfs_alloc_log_keys(cur, rbp, 1, be16_to_cpu(right->bb_numrecs) + 1);
|
|
|
|
xfs_alloc_log_ptrs(cur, rbp, 1, be16_to_cpu(right->bb_numrecs) + 1);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_btree_check_key(cur->bc_btnum, rkp, rkp + 1);
|
|
|
|
} else {
|
|
|
|
xfs_alloc_rec_t *lrp; /* record pointer for left block */
|
|
|
|
xfs_alloc_rec_t *rrp; /* record pointer for right block */
|
|
|
|
|
2005-11-02 12:11:25 +08:00
|
|
|
lrp = XFS_ALLOC_REC_ADDR(left, be16_to_cpu(left->bb_numrecs), cur);
|
2005-04-17 06:20:36 +08:00
|
|
|
rrp = XFS_ALLOC_REC_ADDR(right, 1, cur);
|
2005-11-02 12:11:25 +08:00
|
|
|
memmove(rrp + 1, rrp, be16_to_cpu(right->bb_numrecs) * sizeof(*rrp));
|
2005-04-17 06:20:36 +08:00
|
|
|
*rrp = *lrp;
|
2005-11-02 12:11:25 +08:00
|
|
|
xfs_alloc_log_recs(cur, rbp, 1, be16_to_cpu(right->bb_numrecs) + 1);
|
|
|
|
key.ar_startblock = rrp->ar_startblock;
|
|
|
|
key.ar_blockcount = rrp->ar_blockcount;
|
2005-04-17 06:20:36 +08:00
|
|
|
rkp = &key;
|
|
|
|
xfs_btree_check_rec(cur->bc_btnum, rrp, rrp + 1);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Decrement and log left's numrecs, bump and log right's numrecs.
|
|
|
|
*/
|
2008-02-14 07:03:29 +08:00
|
|
|
be16_add_cpu(&left->bb_numrecs, -1);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_block(cur->bc_tp, lbp, XFS_BB_NUMRECS);
|
2008-02-14 07:03:29 +08:00
|
|
|
be16_add_cpu(&right->bb_numrecs, 1);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_block(cur->bc_tp, rbp, XFS_BB_NUMRECS);
|
|
|
|
/*
|
|
|
|
* Using a temporary cursor, update the parent key values of the
|
|
|
|
* block on the right.
|
|
|
|
*/
|
|
|
|
if ((error = xfs_btree_dup_cursor(cur, &tcur)))
|
|
|
|
return error;
|
|
|
|
i = xfs_btree_lastrec(tcur, level);
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(i == 1, error0);
|
2008-10-30 13:55:45 +08:00
|
|
|
if ((error = xfs_btree_increment(tcur, level, &i)) ||
|
2008-10-30 13:56:22 +08:00
|
|
|
(error = xfs_btree_updkey(tcur, (union xfs_btree_key *)rkp, level + 1)))
|
2005-04-17 06:20:36 +08:00
|
|
|
goto error0;
|
|
|
|
xfs_btree_del_cursor(tcur, XFS_BTREE_NOERROR);
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
error0:
|
|
|
|
xfs_btree_del_cursor(tcur, XFS_BTREE_ERROR);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Split cur/level block in half.
|
|
|
|
* Return new block number and its first record (to be inserted into parent).
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_alloc_split(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
int level, /* level to split */
|
|
|
|
xfs_agblock_t *bnop, /* output: block number allocated */
|
|
|
|
xfs_alloc_key_t *keyp, /* output: first key of new block */
|
|
|
|
xfs_btree_cur_t **curp, /* output: new cursor */
|
|
|
|
int *stat) /* success/failure */
|
|
|
|
{
|
|
|
|
int error; /* error return value */
|
|
|
|
int i; /* loop index/record number */
|
|
|
|
xfs_agblock_t lbno; /* left (current) block number */
|
|
|
|
xfs_buf_t *lbp; /* buffer for left block */
|
|
|
|
xfs_alloc_block_t *left; /* left (current) btree block */
|
|
|
|
xfs_agblock_t rbno; /* right (new) block number */
|
|
|
|
xfs_buf_t *rbp; /* buffer for right block */
|
|
|
|
xfs_alloc_block_t *right; /* right (new) btree block */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate the new block from the freelist.
|
|
|
|
* If we can't do it, we're toast. Give up.
|
|
|
|
*/
|
[XFS] Lazy Superblock Counters
When we have a couple of hundred transactions on the fly at once, they all
typically modify the on disk superblock in some way.
create/unclink/mkdir/rmdir modify inode counts, allocation/freeing modify
free block counts.
When these counts are modified in a transaction, they must eventually lock
the superblock buffer and apply the mods. The buffer then remains locked
until the transaction is committed into the incore log buffer. The result
of this is that with enough transactions on the fly the incore superblock
buffer becomes a bottleneck.
The result of contention on the incore superblock buffer is that
transaction rates fall - the more pressure that is put on the superblock
buffer, the slower things go.
The key to removing the contention is to not require the superblock fields
in question to be locked. We do that by not marking the superblock dirty
in the transaction. IOWs, we modify the incore superblock but do not
modify the cached superblock buffer. In short, we do not log superblock
modifications to critical fields in the superblock on every transaction.
In fact we only do it just before we write the superblock to disk every
sync period or just before unmount.
This creates an interesting problem - if we don't log or write out the
fields in every transaction, then how do the values get recovered after a
crash? the answer is simple - we keep enough duplicate, logged information
in other structures that we can reconstruct the correct count after log
recovery has been performed.
It is the AGF and AGI structures that contain the duplicate information;
after recovery, we walk every AGI and AGF and sum their individual
counters to get the correct value, and we do a transaction into the log to
correct them. An optimisation of this is that if we have a clean unmount
record, we know the value in the superblock is correct, so we can avoid
the summation walk under normal conditions and so mount/recovery times do
not change under normal operation.
One wrinkle that was discovered during development was that the blocks
used in the freespace btrees are never accounted for in the AGF counters.
This was once a valid optimisation to make; when the filesystem is full,
the free space btrees are empty and consume no space. Hence when it
matters, the "accounting" is correct. But that means the when we do the
AGF summations, we would not have a correct count and xfs_check would
complain. Hence a new counter was added to track the number of blocks used
by the free space btrees. This is an *on-disk format change*.
As a result of this, lazy superblock counters are a mkfs option and at the
moment on linux there is no way to convert an old filesystem. This is
possible - xfs_db can be used to twiddle the right bits and then
xfs_repair will do the format conversion for you. Similarly, you can
convert backwards as well. At some point we'll add functionality to
xfs_admin to do the bit twiddling easily....
SGI-PV: 964999
SGI-Modid: xfs-linux-melb:xfs-kern:28652a
Signed-off-by: David Chinner <dgc@sgi.com>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Tim Shimmin <tes@sgi.com>
2007-05-24 13:26:31 +08:00
|
|
|
error = xfs_alloc_get_freelist(cur->bc_tp,
|
|
|
|
cur->bc_private.a.agbp, &rbno, 1);
|
|
|
|
if (error)
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
if (rbno == NULLAGBLOCK) {
|
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
xfs_trans_agbtree_delta(cur->bc_tp, 1);
|
|
|
|
rbp = xfs_btree_get_bufs(cur->bc_mp, cur->bc_tp, cur->bc_private.a.agno,
|
|
|
|
rbno, 0);
|
|
|
|
/*
|
|
|
|
* Set up the new block as "right".
|
|
|
|
*/
|
|
|
|
right = XFS_BUF_TO_ALLOC_BLOCK(rbp);
|
|
|
|
/*
|
|
|
|
* "Left" is the current (according to the cursor) block.
|
|
|
|
*/
|
|
|
|
lbp = cur->bc_bufs[level];
|
|
|
|
left = XFS_BUF_TO_ALLOC_BLOCK(lbp);
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, left, level, lbp)))
|
|
|
|
return error;
|
|
|
|
#endif
|
|
|
|
/*
|
|
|
|
* Fill in the btree header for the new block.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
right->bb_magic = cpu_to_be32(xfs_magics[cur->bc_btnum]);
|
|
|
|
right->bb_level = left->bb_level;
|
|
|
|
right->bb_numrecs = cpu_to_be16(be16_to_cpu(left->bb_numrecs) / 2);
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Make sure that if there's an odd number of entries now, that
|
|
|
|
* each new block will have the same number of entries.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if ((be16_to_cpu(left->bb_numrecs) & 1) &&
|
|
|
|
cur->bc_ptrs[level] <= be16_to_cpu(right->bb_numrecs) + 1)
|
2008-02-14 07:03:29 +08:00
|
|
|
be16_add_cpu(&right->bb_numrecs, 1);
|
2005-11-02 12:11:25 +08:00
|
|
|
i = be16_to_cpu(left->bb_numrecs) - be16_to_cpu(right->bb_numrecs) + 1;
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* For non-leaf blocks, copy keys and addresses over to the new block.
|
|
|
|
*/
|
|
|
|
if (level > 0) {
|
|
|
|
xfs_alloc_key_t *lkp; /* left btree key pointer */
|
|
|
|
xfs_alloc_ptr_t *lpp; /* left btree address pointer */
|
|
|
|
xfs_alloc_key_t *rkp; /* right btree key pointer */
|
|
|
|
xfs_alloc_ptr_t *rpp; /* right btree address pointer */
|
|
|
|
|
|
|
|
lkp = XFS_ALLOC_KEY_ADDR(left, i, cur);
|
|
|
|
lpp = XFS_ALLOC_PTR_ADDR(left, i, cur);
|
|
|
|
rkp = XFS_ALLOC_KEY_ADDR(right, 1, cur);
|
|
|
|
rpp = XFS_ALLOC_PTR_ADDR(right, 1, cur);
|
|
|
|
#ifdef DEBUG
|
2005-11-02 12:11:25 +08:00
|
|
|
for (i = 0; i < be16_to_cpu(right->bb_numrecs); i++) {
|
|
|
|
if ((error = xfs_btree_check_sptr(cur, be32_to_cpu(lpp[i]), level)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
#endif
|
2005-11-02 12:11:25 +08:00
|
|
|
memcpy(rkp, lkp, be16_to_cpu(right->bb_numrecs) * sizeof(*rkp));
|
|
|
|
memcpy(rpp, lpp, be16_to_cpu(right->bb_numrecs) * sizeof(*rpp));
|
|
|
|
xfs_alloc_log_keys(cur, rbp, 1, be16_to_cpu(right->bb_numrecs));
|
|
|
|
xfs_alloc_log_ptrs(cur, rbp, 1, be16_to_cpu(right->bb_numrecs));
|
2005-04-17 06:20:36 +08:00
|
|
|
*keyp = *rkp;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* For leaf blocks, copy records over to the new block.
|
|
|
|
*/
|
|
|
|
else {
|
|
|
|
xfs_alloc_rec_t *lrp; /* left btree record pointer */
|
|
|
|
xfs_alloc_rec_t *rrp; /* right btree record pointer */
|
|
|
|
|
|
|
|
lrp = XFS_ALLOC_REC_ADDR(left, i, cur);
|
|
|
|
rrp = XFS_ALLOC_REC_ADDR(right, 1, cur);
|
2005-11-02 12:11:25 +08:00
|
|
|
memcpy(rrp, lrp, be16_to_cpu(right->bb_numrecs) * sizeof(*rrp));
|
|
|
|
xfs_alloc_log_recs(cur, rbp, 1, be16_to_cpu(right->bb_numrecs));
|
|
|
|
keyp->ar_startblock = rrp->ar_startblock;
|
|
|
|
keyp->ar_blockcount = rrp->ar_blockcount;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Find the left block number by looking in the buffer.
|
|
|
|
* Adjust numrecs, sibling pointers.
|
|
|
|
*/
|
|
|
|
lbno = XFS_DADDR_TO_AGBNO(cur->bc_mp, XFS_BUF_ADDR(lbp));
|
2008-02-14 07:03:29 +08:00
|
|
|
be16_add_cpu(&left->bb_numrecs, -(be16_to_cpu(right->bb_numrecs)));
|
2005-11-02 12:11:25 +08:00
|
|
|
right->bb_rightsib = left->bb_rightsib;
|
|
|
|
left->bb_rightsib = cpu_to_be32(rbno);
|
|
|
|
right->bb_leftsib = cpu_to_be32(lbno);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_block(cur->bc_tp, rbp, XFS_BB_ALL_BITS);
|
|
|
|
xfs_alloc_log_block(cur->bc_tp, lbp, XFS_BB_NUMRECS | XFS_BB_RIGHTSIB);
|
|
|
|
/*
|
|
|
|
* If there's a block to the new block's right, make that block
|
|
|
|
* point back to right instead of to left.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (be32_to_cpu(right->bb_rightsib) != NULLAGBLOCK) {
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_block_t *rrblock; /* rr btree block */
|
|
|
|
xfs_buf_t *rrbp; /* buffer for rrblock */
|
|
|
|
|
|
|
|
if ((error = xfs_btree_read_bufs(cur->bc_mp, cur->bc_tp,
|
2005-11-02 12:11:25 +08:00
|
|
|
cur->bc_private.a.agno, be32_to_cpu(right->bb_rightsib), 0,
|
2005-04-17 06:20:36 +08:00
|
|
|
&rrbp, XFS_ALLOC_BTREE_REF)))
|
|
|
|
return error;
|
|
|
|
rrblock = XFS_BUF_TO_ALLOC_BLOCK(rrbp);
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, rrblock, level, rrbp)))
|
|
|
|
return error;
|
2005-11-02 12:11:25 +08:00
|
|
|
rrblock->bb_leftsib = cpu_to_be32(rbno);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_block(cur->bc_tp, rrbp, XFS_BB_LEFTSIB);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If the cursor is really in the right block, move it there.
|
|
|
|
* If it's just pointing past the last entry in left, then we'll
|
|
|
|
* insert there, so don't change anything in that case.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (cur->bc_ptrs[level] > be16_to_cpu(left->bb_numrecs) + 1) {
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_btree_setbuf(cur, level, rbp);
|
2005-11-02 12:11:25 +08:00
|
|
|
cur->bc_ptrs[level] -= be16_to_cpu(left->bb_numrecs);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If there are more levels, we'll need another cursor which refers to
|
|
|
|
* the right block, no matter where this cursor was.
|
|
|
|
*/
|
|
|
|
if (level + 1 < cur->bc_nlevels) {
|
|
|
|
if ((error = xfs_btree_dup_cursor(cur, curp)))
|
|
|
|
return error;
|
|
|
|
(*curp)->bc_ptrs[level + 1]++;
|
|
|
|
}
|
|
|
|
*bnop = rbno;
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Externally visible routines.
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Delete the record pointed to by cur.
|
|
|
|
* The cursor refers to the place where the record was (could be inserted)
|
|
|
|
* when the operation returns.
|
|
|
|
*/
|
|
|
|
int /* error */
|
|
|
|
xfs_alloc_delete(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
int *stat) /* success/failure */
|
|
|
|
{
|
|
|
|
int error; /* error return value */
|
|
|
|
int i; /* result code */
|
|
|
|
int level; /* btree level */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Go up the tree, starting at leaf level.
|
|
|
|
* If 2 is returned then a join was done; go to the next level.
|
|
|
|
* Otherwise we are done.
|
|
|
|
*/
|
|
|
|
for (level = 0, i = 2; i == 2; level++) {
|
|
|
|
if ((error = xfs_alloc_delrec(cur, level, &i)))
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
if (i == 0) {
|
|
|
|
for (level = 1; level < cur->bc_nlevels; level++) {
|
|
|
|
if (cur->bc_ptrs[level] == 0) {
|
2008-10-30 13:55:58 +08:00
|
|
|
if ((error = xfs_btree_decrement(cur, level, &i)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
*stat = i;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Get the data from the pointed-to record.
|
|
|
|
*/
|
|
|
|
int /* error */
|
|
|
|
xfs_alloc_get_rec(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
xfs_agblock_t *bno, /* output: starting block of extent */
|
|
|
|
xfs_extlen_t *len, /* output: length of extent */
|
|
|
|
int *stat) /* output: success/failure */
|
|
|
|
{
|
|
|
|
xfs_alloc_block_t *block; /* btree block */
|
|
|
|
#ifdef DEBUG
|
|
|
|
int error; /* error return value */
|
|
|
|
#endif
|
|
|
|
int ptr; /* record number */
|
|
|
|
|
|
|
|
ptr = cur->bc_ptrs[0];
|
|
|
|
block = XFS_BUF_TO_ALLOC_BLOCK(cur->bc_bufs[0]);
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, block, 0, cur->bc_bufs[0])))
|
|
|
|
return error;
|
|
|
|
#endif
|
|
|
|
/*
|
|
|
|
* Off the right end or left end, return failure.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
if (ptr > be16_to_cpu(block->bb_numrecs) || ptr <= 0) {
|
2005-04-17 06:20:36 +08:00
|
|
|
*stat = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Point to the record and extract its data.
|
|
|
|
*/
|
|
|
|
{
|
|
|
|
xfs_alloc_rec_t *rec; /* record data */
|
|
|
|
|
|
|
|
rec = XFS_ALLOC_REC_ADDR(block, ptr, cur);
|
2005-11-02 12:11:25 +08:00
|
|
|
*bno = be32_to_cpu(rec->ar_startblock);
|
|
|
|
*len = be32_to_cpu(rec->ar_blockcount);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
*stat = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Insert the current record at the point referenced by cur.
|
|
|
|
* The cursor may be inconsistent on return if splits have been done.
|
|
|
|
*/
|
|
|
|
int /* error */
|
|
|
|
xfs_alloc_insert(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
int *stat) /* success/failure */
|
|
|
|
{
|
|
|
|
int error; /* error return value */
|
|
|
|
int i; /* result value, 0 for failure */
|
|
|
|
int level; /* current level number in btree */
|
|
|
|
xfs_agblock_t nbno; /* new block number (split result) */
|
|
|
|
xfs_btree_cur_t *ncur; /* new cursor (split result) */
|
|
|
|
xfs_alloc_rec_t nrec; /* record being inserted this level */
|
|
|
|
xfs_btree_cur_t *pcur; /* previous level's cursor */
|
|
|
|
|
|
|
|
level = 0;
|
|
|
|
nbno = NULLAGBLOCK;
|
2005-11-02 12:11:25 +08:00
|
|
|
nrec.ar_startblock = cpu_to_be32(cur->bc_rec.a.ar_startblock);
|
|
|
|
nrec.ar_blockcount = cpu_to_be32(cur->bc_rec.a.ar_blockcount);
|
2006-09-28 08:58:40 +08:00
|
|
|
ncur = NULL;
|
2005-04-17 06:20:36 +08:00
|
|
|
pcur = cur;
|
|
|
|
/*
|
|
|
|
* Loop going up the tree, starting at the leaf level.
|
|
|
|
* Stop when we don't get a split block, that must mean that
|
|
|
|
* the insert is finished with this level.
|
|
|
|
*/
|
|
|
|
do {
|
|
|
|
/*
|
|
|
|
* Insert nrec/nbno into this level of the tree.
|
|
|
|
* Note if we fail, nbno will be null.
|
|
|
|
*/
|
|
|
|
if ((error = xfs_alloc_insrec(pcur, level++, &nbno, &nrec, &ncur,
|
|
|
|
&i))) {
|
|
|
|
if (pcur != cur)
|
|
|
|
xfs_btree_del_cursor(pcur, XFS_BTREE_ERROR);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* See if the cursor we just used is trash.
|
|
|
|
* Can't trash the caller's cursor, but otherwise we should
|
|
|
|
* if ncur is a new cursor or we're about to be done.
|
|
|
|
*/
|
|
|
|
if (pcur != cur && (ncur || nbno == NULLAGBLOCK)) {
|
|
|
|
cur->bc_nlevels = pcur->bc_nlevels;
|
|
|
|
xfs_btree_del_cursor(pcur, XFS_BTREE_NOERROR);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If we got a new cursor, switch to it.
|
|
|
|
*/
|
|
|
|
if (ncur) {
|
|
|
|
pcur = ncur;
|
2006-09-28 08:58:40 +08:00
|
|
|
ncur = NULL;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
} while (nbno != NULLAGBLOCK);
|
|
|
|
*stat = i;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Update the record referred to by cur, to the value given by [bno, len].
|
|
|
|
* This either works (return 0) or gets an EFSCORRUPTED error.
|
|
|
|
*/
|
|
|
|
int /* error */
|
|
|
|
xfs_alloc_update(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
xfs_agblock_t bno, /* starting block of extent */
|
|
|
|
xfs_extlen_t len) /* length of extent */
|
|
|
|
{
|
|
|
|
xfs_alloc_block_t *block; /* btree block to update */
|
|
|
|
int error; /* error return value */
|
|
|
|
int ptr; /* current record number (updating) */
|
|
|
|
|
|
|
|
ASSERT(len > 0);
|
|
|
|
/*
|
|
|
|
* Pick up the a.g. freelist struct and the current block.
|
|
|
|
*/
|
|
|
|
block = XFS_BUF_TO_ALLOC_BLOCK(cur->bc_bufs[0]);
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_sblock(cur, block, 0, cur->bc_bufs[0])))
|
|
|
|
return error;
|
|
|
|
#endif
|
|
|
|
/*
|
|
|
|
* Get the address of the rec to be updated.
|
|
|
|
*/
|
|
|
|
ptr = cur->bc_ptrs[0];
|
|
|
|
{
|
|
|
|
xfs_alloc_rec_t *rp; /* pointer to updated record */
|
|
|
|
|
|
|
|
rp = XFS_ALLOC_REC_ADDR(block, ptr, cur);
|
|
|
|
/*
|
|
|
|
* Fill in the new contents and log them.
|
|
|
|
*/
|
2005-11-02 12:11:25 +08:00
|
|
|
rp->ar_startblock = cpu_to_be32(bno);
|
|
|
|
rp->ar_blockcount = cpu_to_be32(len);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_recs(cur, cur->bc_bufs[0], ptr, ptr);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If it's the by-size btree and it's the last leaf block and
|
|
|
|
* it's the last record... then update the size of the longest
|
|
|
|
* extent in the a.g., which we cache in the a.g. freelist header.
|
|
|
|
*/
|
|
|
|
if (cur->bc_btnum == XFS_BTNUM_CNT &&
|
2005-11-02 12:11:25 +08:00
|
|
|
be32_to_cpu(block->bb_rightsib) == NULLAGBLOCK &&
|
|
|
|
ptr == be16_to_cpu(block->bb_numrecs)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_agf_t *agf; /* a.g. freespace header */
|
|
|
|
xfs_agnumber_t seqno;
|
|
|
|
|
|
|
|
agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp);
|
2005-11-02 12:11:25 +08:00
|
|
|
seqno = be32_to_cpu(agf->agf_seqno);
|
2005-04-17 06:20:36 +08:00
|
|
|
cur->bc_mp->m_perag[seqno].pagf_longest = len;
|
2005-11-02 12:11:25 +08:00
|
|
|
agf->agf_longest = cpu_to_be32(len);
|
2005-04-17 06:20:36 +08:00
|
|
|
xfs_alloc_log_agf(cur->bc_tp, cur->bc_private.a.agbp,
|
|
|
|
XFS_AGF_LONGEST);
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Updating first record in leaf. Pass new key value up to our parent.
|
|
|
|
*/
|
|
|
|
if (ptr == 1) {
|
|
|
|
xfs_alloc_key_t key; /* key containing [bno, len] */
|
|
|
|
|
2005-11-02 12:11:25 +08:00
|
|
|
key.ar_startblock = cpu_to_be32(bno);
|
|
|
|
key.ar_blockcount = cpu_to_be32(len);
|
2008-10-30 13:56:22 +08:00
|
|
|
if ((error = xfs_btree_updkey(cur, (union xfs_btree_key *)&key, 1)))
|
2005-04-17 06:20:36 +08:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2008-10-30 13:53:59 +08:00
|
|
|
|
|
|
|
STATIC struct xfs_btree_cur *
|
|
|
|
xfs_allocbt_dup_cursor(
|
|
|
|
struct xfs_btree_cur *cur)
|
|
|
|
{
|
|
|
|
return xfs_allocbt_init_cursor(cur->bc_mp, cur->bc_tp,
|
|
|
|
cur->bc_private.a.agbp, cur->bc_private.a.agno,
|
|
|
|
cur->bc_btnum);
|
|
|
|
}
|
|
|
|
|
2008-10-30 13:55:23 +08:00
|
|
|
STATIC int
|
|
|
|
xfs_allocbt_get_maxrecs(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
int level)
|
|
|
|
{
|
|
|
|
return cur->bc_mp->m_alloc_mxr[level != 0];
|
|
|
|
}
|
|
|
|
|
2008-10-30 13:56:09 +08:00
|
|
|
STATIC void
|
|
|
|
xfs_allocbt_init_key_from_rec(
|
|
|
|
union xfs_btree_key *key,
|
|
|
|
union xfs_btree_rec *rec)
|
|
|
|
{
|
|
|
|
ASSERT(rec->alloc.ar_startblock != 0);
|
|
|
|
|
|
|
|
key->alloc.ar_startblock = rec->alloc.ar_startblock;
|
|
|
|
key->alloc.ar_blockcount = rec->alloc.ar_blockcount;
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC void
|
|
|
|
xfs_allocbt_init_ptr_from_cur(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
union xfs_btree_ptr *ptr)
|
|
|
|
{
|
|
|
|
struct xfs_agf *agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp);
|
|
|
|
|
|
|
|
ASSERT(cur->bc_private.a.agno == be32_to_cpu(agf->agf_seqno));
|
|
|
|
ASSERT(agf->agf_roots[cur->bc_btnum] != 0);
|
|
|
|
|
|
|
|
ptr->s = agf->agf_roots[cur->bc_btnum];
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC __int64_t
|
|
|
|
xfs_allocbt_key_diff(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
union xfs_btree_key *key)
|
|
|
|
{
|
|
|
|
xfs_alloc_rec_incore_t *rec = &cur->bc_rec.a;
|
|
|
|
xfs_alloc_key_t *kp = &key->alloc;
|
|
|
|
__int64_t diff;
|
|
|
|
|
|
|
|
if (cur->bc_btnum == XFS_BTNUM_BNO) {
|
|
|
|
return (__int64_t)be32_to_cpu(kp->ar_startblock) -
|
|
|
|
rec->ar_startblock;
|
|
|
|
}
|
|
|
|
|
|
|
|
diff = (__int64_t)be32_to_cpu(kp->ar_blockcount) - rec->ar_blockcount;
|
|
|
|
if (diff)
|
|
|
|
return diff;
|
|
|
|
|
|
|
|
return (__int64_t)be32_to_cpu(kp->ar_startblock) - rec->ar_startblock;
|
|
|
|
}
|
|
|
|
|
2008-10-30 13:55:13 +08:00
|
|
|
#ifdef XFS_BTREE_TRACE
|
|
|
|
ktrace_t *xfs_allocbt_trace_buf;
|
|
|
|
|
|
|
|
STATIC void
|
|
|
|
xfs_allocbt_trace_enter(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
const char *func,
|
|
|
|
char *s,
|
|
|
|
int type,
|
|
|
|
int line,
|
|
|
|
__psunsigned_t a0,
|
|
|
|
__psunsigned_t a1,
|
|
|
|
__psunsigned_t a2,
|
|
|
|
__psunsigned_t a3,
|
|
|
|
__psunsigned_t a4,
|
|
|
|
__psunsigned_t a5,
|
|
|
|
__psunsigned_t a6,
|
|
|
|
__psunsigned_t a7,
|
|
|
|
__psunsigned_t a8,
|
|
|
|
__psunsigned_t a9,
|
|
|
|
__psunsigned_t a10)
|
|
|
|
{
|
|
|
|
ktrace_enter(xfs_allocbt_trace_buf, (void *)(__psint_t)type,
|
|
|
|
(void *)func, (void *)s, NULL, (void *)cur,
|
|
|
|
(void *)a0, (void *)a1, (void *)a2, (void *)a3,
|
|
|
|
(void *)a4, (void *)a5, (void *)a6, (void *)a7,
|
|
|
|
(void *)a8, (void *)a9, (void *)a10);
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC void
|
|
|
|
xfs_allocbt_trace_cursor(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
__uint32_t *s0,
|
|
|
|
__uint64_t *l0,
|
|
|
|
__uint64_t *l1)
|
|
|
|
{
|
|
|
|
*s0 = cur->bc_private.a.agno;
|
|
|
|
*l0 = cur->bc_rec.a.ar_startblock;
|
|
|
|
*l1 = cur->bc_rec.a.ar_blockcount;
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC void
|
|
|
|
xfs_allocbt_trace_key(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
union xfs_btree_key *key,
|
|
|
|
__uint64_t *l0,
|
|
|
|
__uint64_t *l1)
|
|
|
|
{
|
|
|
|
*l0 = be32_to_cpu(key->alloc.ar_startblock);
|
|
|
|
*l1 = be32_to_cpu(key->alloc.ar_blockcount);
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC void
|
|
|
|
xfs_allocbt_trace_record(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
union xfs_btree_rec *rec,
|
|
|
|
__uint64_t *l0,
|
|
|
|
__uint64_t *l1,
|
|
|
|
__uint64_t *l2)
|
|
|
|
{
|
|
|
|
*l0 = be32_to_cpu(rec->alloc.ar_startblock);
|
|
|
|
*l1 = be32_to_cpu(rec->alloc.ar_blockcount);
|
|
|
|
*l2 = 0;
|
|
|
|
}
|
|
|
|
#endif /* XFS_BTREE_TRACE */
|
|
|
|
|
2008-10-30 13:53:59 +08:00
|
|
|
static const struct xfs_btree_ops xfs_allocbt_ops = {
|
2008-10-30 13:55:34 +08:00
|
|
|
.rec_len = sizeof(xfs_alloc_rec_t),
|
|
|
|
.key_len = sizeof(xfs_alloc_key_t),
|
|
|
|
|
2008-10-30 13:53:59 +08:00
|
|
|
.dup_cursor = xfs_allocbt_dup_cursor,
|
2008-10-30 13:55:23 +08:00
|
|
|
.get_maxrecs = xfs_allocbt_get_maxrecs,
|
2008-10-30 13:56:09 +08:00
|
|
|
.init_key_from_rec = xfs_allocbt_init_key_from_rec,
|
|
|
|
.init_ptr_from_cur = xfs_allocbt_init_ptr_from_cur,
|
|
|
|
.key_diff = xfs_allocbt_key_diff,
|
2008-10-30 13:55:13 +08:00
|
|
|
|
|
|
|
#ifdef XFS_BTREE_TRACE
|
|
|
|
.trace_enter = xfs_allocbt_trace_enter,
|
|
|
|
.trace_cursor = xfs_allocbt_trace_cursor,
|
|
|
|
.trace_key = xfs_allocbt_trace_key,
|
|
|
|
.trace_record = xfs_allocbt_trace_record,
|
|
|
|
#endif
|
2008-10-30 13:53:59 +08:00
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate a new allocation btree cursor.
|
|
|
|
*/
|
|
|
|
struct xfs_btree_cur * /* new alloc btree cursor */
|
|
|
|
xfs_allocbt_init_cursor(
|
|
|
|
struct xfs_mount *mp, /* file system mount point */
|
|
|
|
struct xfs_trans *tp, /* transaction pointer */
|
|
|
|
struct xfs_buf *agbp, /* buffer for agf structure */
|
|
|
|
xfs_agnumber_t agno, /* allocation group number */
|
|
|
|
xfs_btnum_t btnum) /* btree identifier */
|
|
|
|
{
|
|
|
|
struct xfs_agf *agf = XFS_BUF_TO_AGF(agbp);
|
|
|
|
struct xfs_btree_cur *cur;
|
|
|
|
|
|
|
|
ASSERT(btnum == XFS_BTNUM_BNO || btnum == XFS_BTNUM_CNT);
|
|
|
|
|
|
|
|
cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_SLEEP);
|
|
|
|
|
|
|
|
cur->bc_tp = tp;
|
|
|
|
cur->bc_mp = mp;
|
|
|
|
cur->bc_nlevels = be32_to_cpu(agf->agf_levels[btnum]);
|
|
|
|
cur->bc_btnum = btnum;
|
|
|
|
cur->bc_blocklog = mp->m_sb.sb_blocklog;
|
|
|
|
|
|
|
|
cur->bc_ops = &xfs_allocbt_ops;
|
|
|
|
|
|
|
|
cur->bc_private.a.agbp = agbp;
|
|
|
|
cur->bc_private.a.agno = agno;
|
|
|
|
|
|
|
|
return cur;
|
|
|
|
}
|