37 */ 38 39#include <sys/param.h> 40#include <sys/systm.h> 41#include <sys/mount.h> 42#include <sys/bio.h> 43#include <sys/buf.h> 44#include <sys/vnode.h> 45#include <sys/malloc.h> 46 47#include <vm/vm.h> 48#include <vm/vm_extern.h> 49 50#include <gnu/ext2fs/inode.h> 51#include <gnu/ext2fs/ext2_mount.h> 52#include <gnu/ext2fs/ext2_fs.h> 53#include <gnu/ext2fs/ext2_fs_sb.h> 54#include <gnu/ext2fs/fs.h> 55#include <gnu/ext2fs/ext2_extern.h> 56 57static int ext2_indirtrunc(struct inode *, int32_t, int32_t, int32_t, int, 58 long *); 59 60/* 61 * Update the access, modified, and inode change times as specified by the 62 * IN_ACCESS, IN_UPDATE, and IN_CHANGE flags respectively. Write the inode 63 * to disk if the IN_MODIFIED flag is set (it may be set initially, or by 64 * the timestamp update). The IN_LAZYMOD flag is set to force a write 65 * later if not now. If we write now, then clear both IN_MODIFIED and 66 * IN_LAZYMOD to reflect the presumably successful write, and if waitfor is 67 * set, then wait for the write to complete. 68 */ 69int 70ext2_update(vp, waitfor) 71 struct vnode *vp; 72 int waitfor; 73{ 74 struct ext2_sb_info *fs; 75 struct buf *bp; 76 struct inode *ip; 77 int error; 78 79 ext2_itimes(vp); 80 ip = VTOI(vp); 81 if ((ip->i_flag & IN_MODIFIED) == 0) 82 return (0); 83 ip->i_flag &= ~(IN_LAZYMOD | IN_MODIFIED); 84 if (vp->v_mount->mnt_flag & MNT_RDONLY) 85 return (0); 86 fs = ip->i_e2fs; 87 if ((error = bread(ip->i_devvp, 88 fsbtodb(fs, ino_to_fsba(fs, ip->i_number)), 89 (int)fs->s_blocksize, NOCRED, &bp)) != 0) { 90 brelse(bp); 91 return (error); 92 } 93 ext2_i2ei(ip, (struct ext2_inode *)((char *)bp->b_data + 94 EXT2_INODE_SIZE * ino_to_fsbo(fs, ip->i_number))); 95 if (waitfor && (vp->v_mount->mnt_flag & MNT_ASYNC) == 0) 96 return (bwrite(bp)); 97 else { 98 bdwrite(bp); 99 return (0); 100 } 101} 102 103#define SINGLE 0 /* index of single indirect block */ 104#define DOUBLE 1 /* index of double indirect block */ 105#define TRIPLE 2 /* index of triple indirect block */ 106/* 107 * Truncate the inode oip to at most length size, freeing the 108 * disk blocks. 109 */ 110int 111ext2_truncate(vp, length, flags, cred, td) 112 struct vnode *vp; 113 off_t length; 114 int flags; 115 struct ucred *cred; 116 struct thread *td; 117{ 118 struct vnode *ovp = vp; 119 int32_t lastblock; 120 struct inode *oip; 121 int32_t bn, lbn, lastiblock[NIADDR], indir_lbn[NIADDR]; 122 int32_t oldblks[NDADDR + NIADDR], newblks[NDADDR + NIADDR]; 123 struct ext2_sb_info *fs; 124 struct buf *bp; 125 int offset, size, level; 126 long count, nblocks, blocksreleased = 0; 127 int aflags, error, i, allerror; 128 off_t osize; 129/* 130printf("ext2_truncate called %d to %d\n", VTOI(ovp)->i_number, length); 131*/ /* 132 * negative file sizes will totally break the code below and 133 * are not meaningful anyways. 134 */ 135 if (length < 0) 136 return EFBIG; 137 138 oip = VTOI(ovp); 139 if (ovp->v_type == VLNK && 140 oip->i_size < ovp->v_mount->mnt_maxsymlinklen) { 141#if DIAGNOSTIC 142 if (length != 0) 143 panic("ext2_truncate: partial truncate of symlink"); 144#endif 145 bzero((char *)&oip->i_shortlink, (u_int)oip->i_size); 146 oip->i_size = 0; 147 oip->i_flag |= IN_CHANGE | IN_UPDATE; 148 return (ext2_update(ovp, 1)); 149 } 150 if (oip->i_size == length) { 151 oip->i_flag |= IN_CHANGE | IN_UPDATE; 152 return (ext2_update(ovp, 0)); 153 } 154 fs = oip->i_e2fs; 155 osize = oip->i_size; 156 ext2_discard_prealloc(oip); 157 /* 158 * Lengthen the size of the file. We must ensure that the 159 * last byte of the file is allocated. Since the smallest 160 * value of oszie is 0, length will be at least 1. 161 */ 162 if (osize < length) { 163 if (length > oip->i_e2fs->fs_maxfilesize) 164 return (EFBIG); 165 offset = blkoff(fs, length - 1); 166 lbn = lblkno(fs, length - 1); 167 aflags = B_CLRBUF; 168 if (flags & IO_SYNC) 169 aflags |= B_SYNC; 170 vnode_pager_setsize(ovp, length); 171 if ((error = ext2_balloc(oip, lbn, offset + 1, cred, &bp, 172 aflags)) != 0) 173 return (error); 174 oip->i_size = length; 175 if (aflags & IO_SYNC) 176 bwrite(bp); 177 else 178 bawrite(bp); 179 oip->i_flag |= IN_CHANGE | IN_UPDATE; 180 return (ext2_update(ovp, 1)); 181 } 182 /* 183 * Shorten the size of the file. If the file is not being 184 * truncated to a block boundry, the contents of the 185 * partial block following the end of the file must be 186 * zero'ed in case it ever become accessible again because 187 * of subsequent file growth. 188 */ 189 /* I don't understand the comment above */ 190 offset = blkoff(fs, length); 191 if (offset == 0) { 192 oip->i_size = length; 193 } else { 194 lbn = lblkno(fs, length); 195 aflags = B_CLRBUF; 196 if (flags & IO_SYNC) 197 aflags |= B_SYNC; 198 if ((error = ext2_balloc(oip, lbn, offset, cred, &bp, 199 aflags)) != 0) 200 return (error); 201 oip->i_size = length; 202 size = blksize(fs, oip, lbn); 203 bzero((char *)bp->b_data + offset, (u_int)(size - offset)); 204 allocbuf(bp, size); 205 if (aflags & IO_SYNC) 206 bwrite(bp); 207 else 208 bawrite(bp); 209 } 210 /* 211 * Calculate index into inode's block list of 212 * last direct and indirect blocks (if any) 213 * which we want to keep. Lastblock is -1 when 214 * the file is truncated to 0. 215 */ 216 lastblock = lblkno(fs, length + fs->s_blocksize - 1) - 1; 217 lastiblock[SINGLE] = lastblock - NDADDR; 218 lastiblock[DOUBLE] = lastiblock[SINGLE] - NINDIR(fs); 219 lastiblock[TRIPLE] = lastiblock[DOUBLE] - NINDIR(fs) * NINDIR(fs); 220 nblocks = btodb(fs->s_blocksize); 221 /* 222 * Update file and block pointers on disk before we start freeing 223 * blocks. If we crash before free'ing blocks below, the blocks 224 * will be returned to the free list. lastiblock values are also 225 * normalized to -1 for calls to ext2_indirtrunc below. 226 */ 227 bcopy((caddr_t)&oip->i_db[0], (caddr_t)oldblks, sizeof oldblks); 228 for (level = TRIPLE; level >= SINGLE; level--) 229 if (lastiblock[level] < 0) { 230 oip->i_ib[level] = 0; 231 lastiblock[level] = -1; 232 } 233 for (i = NDADDR - 1; i > lastblock; i--) 234 oip->i_db[i] = 0; 235 oip->i_flag |= IN_CHANGE | IN_UPDATE; 236 allerror = ext2_update(ovp, 1); 237 238 /* 239 * Having written the new inode to disk, save its new configuration 240 * and put back the old block pointers long enough to process them. 241 * Note that we save the new block configuration so we can check it 242 * when we are done. 243 */ 244 bcopy((caddr_t)&oip->i_db[0], (caddr_t)newblks, sizeof newblks); 245 bcopy((caddr_t)oldblks, (caddr_t)&oip->i_db[0], sizeof oldblks); 246 oip->i_size = osize; 247 error = vtruncbuf(ovp, cred, td, length, (int)fs->s_blocksize); 248 if (error && (allerror == 0)) 249 allerror = error; 250 251 /* 252 * Indirect blocks first. 253 */ 254 indir_lbn[SINGLE] = -NDADDR; 255 indir_lbn[DOUBLE] = indir_lbn[SINGLE] - NINDIR(fs) - 1; 256 indir_lbn[TRIPLE] = indir_lbn[DOUBLE] - NINDIR(fs) * NINDIR(fs) - 1; 257 for (level = TRIPLE; level >= SINGLE; level--) { 258 bn = oip->i_ib[level]; 259 if (bn != 0) { 260 error = ext2_indirtrunc(oip, indir_lbn[level], 261 fsbtodb(fs, bn), lastiblock[level], level, &count); 262 if (error) 263 allerror = error; 264 blocksreleased += count; 265 if (lastiblock[level] < 0) { 266 oip->i_ib[level] = 0; 267 ext2_blkfree(oip, bn, fs->s_frag_size); 268 blocksreleased += nblocks; 269 } 270 } 271 if (lastiblock[level] >= 0) 272 goto done; 273 } 274 275 /* 276 * All whole direct blocks or frags. 277 */ 278 for (i = NDADDR - 1; i > lastblock; i--) { 279 long bsize; 280 281 bn = oip->i_db[i]; 282 if (bn == 0) 283 continue; 284 oip->i_db[i] = 0; 285 bsize = blksize(fs, oip, i); 286 ext2_blkfree(oip, bn, bsize); 287 blocksreleased += btodb(bsize); 288 } 289 if (lastblock < 0) 290 goto done; 291 292 /* 293 * Finally, look for a change in size of the 294 * last direct block; release any frags. 295 */ 296 bn = oip->i_db[lastblock]; 297 if (bn != 0) { 298 long oldspace, newspace; 299 300 /* 301 * Calculate amount of space we're giving 302 * back as old block size minus new block size. 303 */ 304 oldspace = blksize(fs, oip, lastblock); 305 oip->i_size = length; 306 newspace = blksize(fs, oip, lastblock); 307 if (newspace == 0) 308 panic("itrunc: newspace"); 309 if (oldspace - newspace > 0) { 310 /* 311 * Block number of space to be free'd is 312 * the old block # plus the number of frags 313 * required for the storage we're keeping. 314 */ 315 bn += numfrags(fs, newspace); 316 ext2_blkfree(oip, bn, oldspace - newspace); 317 blocksreleased += btodb(oldspace - newspace); 318 } 319 } 320done: 321#if DIAGNOSTIC 322 for (level = SINGLE; level <= TRIPLE; level++) 323 if (newblks[NDADDR + level] != oip->i_ib[level]) 324 panic("itrunc1"); 325 for (i = 0; i < NDADDR; i++) 326 if (newblks[i] != oip->i_db[i]) 327 panic("itrunc2"); 328 VI_LOCK(ovp); 329 if (length == 0 && (ovp->v_bufobj.bo_dirty.bv_cnt != 0 || 330 ovp->v_bufobj.bo_clean.bv_cnt != 0)) 331 panic("itrunc3"); 332 VI_UNLOCK(ovp); 333#endif /* DIAGNOSTIC */ 334 /* 335 * Put back the real size. 336 */ 337 oip->i_size = length; 338 oip->i_blocks -= blocksreleased; 339 if (oip->i_blocks < 0) /* sanity */ 340 oip->i_blocks = 0; 341 oip->i_flag |= IN_CHANGE; 342 vnode_pager_setsize(ovp, length); 343 return (allerror); 344} 345 346/* 347 * Release blocks associated with the inode ip and stored in the indirect 348 * block bn. Blocks are free'd in LIFO order up to (but not including) 349 * lastbn. If level is greater than SINGLE, the block is an indirect block 350 * and recursive calls to indirtrunc must be used to cleanse other indirect 351 * blocks. 352 * 353 * NB: triple indirect blocks are untested. 354 */ 355 356static int 357ext2_indirtrunc(ip, lbn, dbn, lastbn, level, countp) 358 struct inode *ip; 359 int32_t lbn, lastbn; 360 int32_t dbn; 361 int level; 362 long *countp; 363{ 364 struct buf *bp; 365 struct ext2_sb_info *fs = ip->i_e2fs; 366 struct vnode *vp; 367 int32_t *bap, *copy, nb, nlbn, last; 368 long blkcount, factor; 369 int i, nblocks, blocksreleased = 0; 370 int error = 0, allerror = 0; 371 372 /* 373 * Calculate index in current block of last 374 * block to be kept. -1 indicates the entire 375 * block so we need not calculate the index. 376 */ 377 factor = 1; 378 for (i = SINGLE; i < level; i++) 379 factor *= NINDIR(fs); 380 last = lastbn; 381 if (lastbn > 0) 382 last /= factor; 383 nblocks = btodb(fs->s_blocksize); 384 /* 385 * Get buffer of block pointers, zero those entries corresponding 386 * to blocks to be free'd, and update on disk copy first. Since 387 * double(triple) indirect before single(double) indirect, calls 388 * to bmap on these blocks will fail. However, we already have 389 * the on disk address, so we have to set the b_blkno field 390 * explicitly instead of letting bread do everything for us. 391 */ 392 vp = ITOV(ip); 393 bp = getblk(vp, lbn, (int)fs->s_blocksize, 0, 0, 0); 394 if (bp->b_flags & (B_DONE | B_DELWRI)) { 395 } else { 396 bp->b_iocmd = BIO_READ; 397 if (bp->b_bcount > bp->b_bufsize) 398 panic("ext2_indirtrunc: bad buffer size"); 399 bp->b_blkno = dbn; 400 vfs_busy_pages(bp, 0); 401 bp->b_iooffset = dbtob(bp->b_blkno); 402 bstrategy(bp); 403 error = bufwait(bp); 404 } 405 if (error) { 406 brelse(bp); 407 *countp = 0; 408 return (error); 409 } 410 411 bap = (int32_t *)bp->b_data; 412 MALLOC(copy, int32_t *, fs->s_blocksize, M_TEMP, M_WAITOK); 413 bcopy((caddr_t)bap, (caddr_t)copy, (u_int)fs->s_blocksize); 414 bzero((caddr_t)&bap[last + 1], 415 (u_int)(NINDIR(fs) - (last + 1)) * sizeof (int32_t)); 416 if (last == -1) 417 bp->b_flags |= B_INVAL; 418 error = bwrite(bp); 419 if (error) 420 allerror = error; 421 bap = copy; 422 423 /* 424 * Recursively free totally unused blocks. 425 */ 426 for (i = NINDIR(fs) - 1, nlbn = lbn + 1 - i * factor; i > last; 427 i--, nlbn += factor) { 428 nb = bap[i]; 429 if (nb == 0) 430 continue; 431 if (level > SINGLE) { 432 if ((error = ext2_indirtrunc(ip, nlbn, 433 fsbtodb(fs, nb), (int32_t)-1, level - 1, &blkcount)) != 0) 434 allerror = error; 435 blocksreleased += blkcount; 436 } 437 ext2_blkfree(ip, nb, fs->s_blocksize); 438 blocksreleased += nblocks; 439 } 440 441 /* 442 * Recursively free last partial block. 443 */ 444 if (level > SINGLE && lastbn >= 0) { 445 last = lastbn % factor; 446 nb = bap[i]; 447 if (nb != 0) { 448 if ((error = ext2_indirtrunc(ip, nlbn, fsbtodb(fs, nb), 449 last, level - 1, &blkcount)) != 0) 450 allerror = error; 451 blocksreleased += blkcount; 452 } 453 } 454 FREE(copy, M_TEMP); 455 *countp = blocksreleased; 456 return (allerror); 457} 458 459/* 460 * discard preallocated blocks 461 */ 462int 463ext2_inactive(ap) 464 struct vop_inactive_args /* { 465 struct vnode *a_vp; 466 struct thread *a_td; 467 } */ *ap; 468{ 469 struct vnode *vp = ap->a_vp; 470 struct inode *ip = VTOI(vp); 471 struct thread *td = ap->a_td; 472 int mode, error = 0; 473 474 ext2_discard_prealloc(ip); 475 if (prtactive && vrefcnt(vp) != 0) 476 vprint("ext2_inactive: pushing active", vp); 477 478 /* 479 * Ignore inodes related to stale file handles. 480 */ 481 if (ip->i_mode == 0) 482 goto out; 483 if (ip->i_nlink <= 0) { 484 (void) vn_write_suspend_wait(vp, NULL, V_WAIT); 485 error = ext2_truncate(vp, (off_t)0, 0, NOCRED, td); 486 ip->i_rdev = 0; 487 mode = ip->i_mode; 488 ip->i_mode = 0; 489 ip->i_flag |= IN_CHANGE | IN_UPDATE; 490 ext2_vfree(vp, ip->i_number, mode); 491 } 492 if (ip->i_flag & (IN_ACCESS | IN_CHANGE | IN_MODIFIED | IN_UPDATE)) { 493 if ((ip->i_flag & (IN_CHANGE | IN_UPDATE | IN_MODIFIED)) == 0 && 494 vn_write_suspend_wait(vp, NULL, V_NOWAIT)) { 495 ip->i_flag &= ~IN_ACCESS; 496 } else { 497 (void) vn_write_suspend_wait(vp, NULL, V_WAIT); 498 ext2_update(vp, 0); 499 } 500 } 501out: 502 VOP_UNLOCK(vp, 0, td); 503 /* 504 * If we are done with the inode, reclaim it 505 * so that it can be reused immediately. 506 */ 507 if (ip->i_mode == 0) 508 vrecycle(vp, NULL, td); 509 return (error); 510} 511 512/* 513 * Reclaim an inode so that it can be used for other purposes. 514 */ 515int 516ext2_reclaim(ap) 517 struct vop_reclaim_args /* { 518 struct vnode *a_vp; 519 struct thread *a_td; 520 } */ *ap; 521{ 522 struct inode *ip; 523 struct vnode *vp = ap->a_vp; 524 525 if (prtactive && vrefcnt(vp) != 0) 526 vprint("ufs_reclaim: pushing active", vp); 527 ip = VTOI(vp); 528 if (ip->i_flag & IN_LAZYMOD) { 529 ip->i_flag |= IN_MODIFIED; 530 ext2_update(vp, 0); 531 } 532 /* 533 * Remove the inode from its hash chain. 534 */ 535 ext2_ihashrem(ip); 536 /* 537 * Purge old data structures associated with the inode. 538 */ 539 if (ip->i_devvp) { 540 vrele(ip->i_devvp); 541 ip->i_devvp = 0; 542 } 543 FREE(vp->v_data, M_EXT2NODE); 544 vp->v_data = 0; 545 return (0); 546}
| 37 */ 38 39#include <sys/param.h> 40#include <sys/systm.h> 41#include <sys/mount.h> 42#include <sys/bio.h> 43#include <sys/buf.h> 44#include <sys/vnode.h> 45#include <sys/malloc.h> 46 47#include <vm/vm.h> 48#include <vm/vm_extern.h> 49 50#include <gnu/ext2fs/inode.h> 51#include <gnu/ext2fs/ext2_mount.h> 52#include <gnu/ext2fs/ext2_fs.h> 53#include <gnu/ext2fs/ext2_fs_sb.h> 54#include <gnu/ext2fs/fs.h> 55#include <gnu/ext2fs/ext2_extern.h> 56 57static int ext2_indirtrunc(struct inode *, int32_t, int32_t, int32_t, int, 58 long *); 59 60/* 61 * Update the access, modified, and inode change times as specified by the 62 * IN_ACCESS, IN_UPDATE, and IN_CHANGE flags respectively. Write the inode 63 * to disk if the IN_MODIFIED flag is set (it may be set initially, or by 64 * the timestamp update). The IN_LAZYMOD flag is set to force a write 65 * later if not now. If we write now, then clear both IN_MODIFIED and 66 * IN_LAZYMOD to reflect the presumably successful write, and if waitfor is 67 * set, then wait for the write to complete. 68 */ 69int 70ext2_update(vp, waitfor) 71 struct vnode *vp; 72 int waitfor; 73{ 74 struct ext2_sb_info *fs; 75 struct buf *bp; 76 struct inode *ip; 77 int error; 78 79 ext2_itimes(vp); 80 ip = VTOI(vp); 81 if ((ip->i_flag & IN_MODIFIED) == 0) 82 return (0); 83 ip->i_flag &= ~(IN_LAZYMOD | IN_MODIFIED); 84 if (vp->v_mount->mnt_flag & MNT_RDONLY) 85 return (0); 86 fs = ip->i_e2fs; 87 if ((error = bread(ip->i_devvp, 88 fsbtodb(fs, ino_to_fsba(fs, ip->i_number)), 89 (int)fs->s_blocksize, NOCRED, &bp)) != 0) { 90 brelse(bp); 91 return (error); 92 } 93 ext2_i2ei(ip, (struct ext2_inode *)((char *)bp->b_data + 94 EXT2_INODE_SIZE * ino_to_fsbo(fs, ip->i_number))); 95 if (waitfor && (vp->v_mount->mnt_flag & MNT_ASYNC) == 0) 96 return (bwrite(bp)); 97 else { 98 bdwrite(bp); 99 return (0); 100 } 101} 102 103#define SINGLE 0 /* index of single indirect block */ 104#define DOUBLE 1 /* index of double indirect block */ 105#define TRIPLE 2 /* index of triple indirect block */ 106/* 107 * Truncate the inode oip to at most length size, freeing the 108 * disk blocks. 109 */ 110int 111ext2_truncate(vp, length, flags, cred, td) 112 struct vnode *vp; 113 off_t length; 114 int flags; 115 struct ucred *cred; 116 struct thread *td; 117{ 118 struct vnode *ovp = vp; 119 int32_t lastblock; 120 struct inode *oip; 121 int32_t bn, lbn, lastiblock[NIADDR], indir_lbn[NIADDR]; 122 int32_t oldblks[NDADDR + NIADDR], newblks[NDADDR + NIADDR]; 123 struct ext2_sb_info *fs; 124 struct buf *bp; 125 int offset, size, level; 126 long count, nblocks, blocksreleased = 0; 127 int aflags, error, i, allerror; 128 off_t osize; 129/* 130printf("ext2_truncate called %d to %d\n", VTOI(ovp)->i_number, length); 131*/ /* 132 * negative file sizes will totally break the code below and 133 * are not meaningful anyways. 134 */ 135 if (length < 0) 136 return EFBIG; 137 138 oip = VTOI(ovp); 139 if (ovp->v_type == VLNK && 140 oip->i_size < ovp->v_mount->mnt_maxsymlinklen) { 141#if DIAGNOSTIC 142 if (length != 0) 143 panic("ext2_truncate: partial truncate of symlink"); 144#endif 145 bzero((char *)&oip->i_shortlink, (u_int)oip->i_size); 146 oip->i_size = 0; 147 oip->i_flag |= IN_CHANGE | IN_UPDATE; 148 return (ext2_update(ovp, 1)); 149 } 150 if (oip->i_size == length) { 151 oip->i_flag |= IN_CHANGE | IN_UPDATE; 152 return (ext2_update(ovp, 0)); 153 } 154 fs = oip->i_e2fs; 155 osize = oip->i_size; 156 ext2_discard_prealloc(oip); 157 /* 158 * Lengthen the size of the file. We must ensure that the 159 * last byte of the file is allocated. Since the smallest 160 * value of oszie is 0, length will be at least 1. 161 */ 162 if (osize < length) { 163 if (length > oip->i_e2fs->fs_maxfilesize) 164 return (EFBIG); 165 offset = blkoff(fs, length - 1); 166 lbn = lblkno(fs, length - 1); 167 aflags = B_CLRBUF; 168 if (flags & IO_SYNC) 169 aflags |= B_SYNC; 170 vnode_pager_setsize(ovp, length); 171 if ((error = ext2_balloc(oip, lbn, offset + 1, cred, &bp, 172 aflags)) != 0) 173 return (error); 174 oip->i_size = length; 175 if (aflags & IO_SYNC) 176 bwrite(bp); 177 else 178 bawrite(bp); 179 oip->i_flag |= IN_CHANGE | IN_UPDATE; 180 return (ext2_update(ovp, 1)); 181 } 182 /* 183 * Shorten the size of the file. If the file is not being 184 * truncated to a block boundry, the contents of the 185 * partial block following the end of the file must be 186 * zero'ed in case it ever become accessible again because 187 * of subsequent file growth. 188 */ 189 /* I don't understand the comment above */ 190 offset = blkoff(fs, length); 191 if (offset == 0) { 192 oip->i_size = length; 193 } else { 194 lbn = lblkno(fs, length); 195 aflags = B_CLRBUF; 196 if (flags & IO_SYNC) 197 aflags |= B_SYNC; 198 if ((error = ext2_balloc(oip, lbn, offset, cred, &bp, 199 aflags)) != 0) 200 return (error); 201 oip->i_size = length; 202 size = blksize(fs, oip, lbn); 203 bzero((char *)bp->b_data + offset, (u_int)(size - offset)); 204 allocbuf(bp, size); 205 if (aflags & IO_SYNC) 206 bwrite(bp); 207 else 208 bawrite(bp); 209 } 210 /* 211 * Calculate index into inode's block list of 212 * last direct and indirect blocks (if any) 213 * which we want to keep. Lastblock is -1 when 214 * the file is truncated to 0. 215 */ 216 lastblock = lblkno(fs, length + fs->s_blocksize - 1) - 1; 217 lastiblock[SINGLE] = lastblock - NDADDR; 218 lastiblock[DOUBLE] = lastiblock[SINGLE] - NINDIR(fs); 219 lastiblock[TRIPLE] = lastiblock[DOUBLE] - NINDIR(fs) * NINDIR(fs); 220 nblocks = btodb(fs->s_blocksize); 221 /* 222 * Update file and block pointers on disk before we start freeing 223 * blocks. If we crash before free'ing blocks below, the blocks 224 * will be returned to the free list. lastiblock values are also 225 * normalized to -1 for calls to ext2_indirtrunc below. 226 */ 227 bcopy((caddr_t)&oip->i_db[0], (caddr_t)oldblks, sizeof oldblks); 228 for (level = TRIPLE; level >= SINGLE; level--) 229 if (lastiblock[level] < 0) { 230 oip->i_ib[level] = 0; 231 lastiblock[level] = -1; 232 } 233 for (i = NDADDR - 1; i > lastblock; i--) 234 oip->i_db[i] = 0; 235 oip->i_flag |= IN_CHANGE | IN_UPDATE; 236 allerror = ext2_update(ovp, 1); 237 238 /* 239 * Having written the new inode to disk, save its new configuration 240 * and put back the old block pointers long enough to process them. 241 * Note that we save the new block configuration so we can check it 242 * when we are done. 243 */ 244 bcopy((caddr_t)&oip->i_db[0], (caddr_t)newblks, sizeof newblks); 245 bcopy((caddr_t)oldblks, (caddr_t)&oip->i_db[0], sizeof oldblks); 246 oip->i_size = osize; 247 error = vtruncbuf(ovp, cred, td, length, (int)fs->s_blocksize); 248 if (error && (allerror == 0)) 249 allerror = error; 250 251 /* 252 * Indirect blocks first. 253 */ 254 indir_lbn[SINGLE] = -NDADDR; 255 indir_lbn[DOUBLE] = indir_lbn[SINGLE] - NINDIR(fs) - 1; 256 indir_lbn[TRIPLE] = indir_lbn[DOUBLE] - NINDIR(fs) * NINDIR(fs) - 1; 257 for (level = TRIPLE; level >= SINGLE; level--) { 258 bn = oip->i_ib[level]; 259 if (bn != 0) { 260 error = ext2_indirtrunc(oip, indir_lbn[level], 261 fsbtodb(fs, bn), lastiblock[level], level, &count); 262 if (error) 263 allerror = error; 264 blocksreleased += count; 265 if (lastiblock[level] < 0) { 266 oip->i_ib[level] = 0; 267 ext2_blkfree(oip, bn, fs->s_frag_size); 268 blocksreleased += nblocks; 269 } 270 } 271 if (lastiblock[level] >= 0) 272 goto done; 273 } 274 275 /* 276 * All whole direct blocks or frags. 277 */ 278 for (i = NDADDR - 1; i > lastblock; i--) { 279 long bsize; 280 281 bn = oip->i_db[i]; 282 if (bn == 0) 283 continue; 284 oip->i_db[i] = 0; 285 bsize = blksize(fs, oip, i); 286 ext2_blkfree(oip, bn, bsize); 287 blocksreleased += btodb(bsize); 288 } 289 if (lastblock < 0) 290 goto done; 291 292 /* 293 * Finally, look for a change in size of the 294 * last direct block; release any frags. 295 */ 296 bn = oip->i_db[lastblock]; 297 if (bn != 0) { 298 long oldspace, newspace; 299 300 /* 301 * Calculate amount of space we're giving 302 * back as old block size minus new block size. 303 */ 304 oldspace = blksize(fs, oip, lastblock); 305 oip->i_size = length; 306 newspace = blksize(fs, oip, lastblock); 307 if (newspace == 0) 308 panic("itrunc: newspace"); 309 if (oldspace - newspace > 0) { 310 /* 311 * Block number of space to be free'd is 312 * the old block # plus the number of frags 313 * required for the storage we're keeping. 314 */ 315 bn += numfrags(fs, newspace); 316 ext2_blkfree(oip, bn, oldspace - newspace); 317 blocksreleased += btodb(oldspace - newspace); 318 } 319 } 320done: 321#if DIAGNOSTIC 322 for (level = SINGLE; level <= TRIPLE; level++) 323 if (newblks[NDADDR + level] != oip->i_ib[level]) 324 panic("itrunc1"); 325 for (i = 0; i < NDADDR; i++) 326 if (newblks[i] != oip->i_db[i]) 327 panic("itrunc2"); 328 VI_LOCK(ovp); 329 if (length == 0 && (ovp->v_bufobj.bo_dirty.bv_cnt != 0 || 330 ovp->v_bufobj.bo_clean.bv_cnt != 0)) 331 panic("itrunc3"); 332 VI_UNLOCK(ovp); 333#endif /* DIAGNOSTIC */ 334 /* 335 * Put back the real size. 336 */ 337 oip->i_size = length; 338 oip->i_blocks -= blocksreleased; 339 if (oip->i_blocks < 0) /* sanity */ 340 oip->i_blocks = 0; 341 oip->i_flag |= IN_CHANGE; 342 vnode_pager_setsize(ovp, length); 343 return (allerror); 344} 345 346/* 347 * Release blocks associated with the inode ip and stored in the indirect 348 * block bn. Blocks are free'd in LIFO order up to (but not including) 349 * lastbn. If level is greater than SINGLE, the block is an indirect block 350 * and recursive calls to indirtrunc must be used to cleanse other indirect 351 * blocks. 352 * 353 * NB: triple indirect blocks are untested. 354 */ 355 356static int 357ext2_indirtrunc(ip, lbn, dbn, lastbn, level, countp) 358 struct inode *ip; 359 int32_t lbn, lastbn; 360 int32_t dbn; 361 int level; 362 long *countp; 363{ 364 struct buf *bp; 365 struct ext2_sb_info *fs = ip->i_e2fs; 366 struct vnode *vp; 367 int32_t *bap, *copy, nb, nlbn, last; 368 long blkcount, factor; 369 int i, nblocks, blocksreleased = 0; 370 int error = 0, allerror = 0; 371 372 /* 373 * Calculate index in current block of last 374 * block to be kept. -1 indicates the entire 375 * block so we need not calculate the index. 376 */ 377 factor = 1; 378 for (i = SINGLE; i < level; i++) 379 factor *= NINDIR(fs); 380 last = lastbn; 381 if (lastbn > 0) 382 last /= factor; 383 nblocks = btodb(fs->s_blocksize); 384 /* 385 * Get buffer of block pointers, zero those entries corresponding 386 * to blocks to be free'd, and update on disk copy first. Since 387 * double(triple) indirect before single(double) indirect, calls 388 * to bmap on these blocks will fail. However, we already have 389 * the on disk address, so we have to set the b_blkno field 390 * explicitly instead of letting bread do everything for us. 391 */ 392 vp = ITOV(ip); 393 bp = getblk(vp, lbn, (int)fs->s_blocksize, 0, 0, 0); 394 if (bp->b_flags & (B_DONE | B_DELWRI)) { 395 } else { 396 bp->b_iocmd = BIO_READ; 397 if (bp->b_bcount > bp->b_bufsize) 398 panic("ext2_indirtrunc: bad buffer size"); 399 bp->b_blkno = dbn; 400 vfs_busy_pages(bp, 0); 401 bp->b_iooffset = dbtob(bp->b_blkno); 402 bstrategy(bp); 403 error = bufwait(bp); 404 } 405 if (error) { 406 brelse(bp); 407 *countp = 0; 408 return (error); 409 } 410 411 bap = (int32_t *)bp->b_data; 412 MALLOC(copy, int32_t *, fs->s_blocksize, M_TEMP, M_WAITOK); 413 bcopy((caddr_t)bap, (caddr_t)copy, (u_int)fs->s_blocksize); 414 bzero((caddr_t)&bap[last + 1], 415 (u_int)(NINDIR(fs) - (last + 1)) * sizeof (int32_t)); 416 if (last == -1) 417 bp->b_flags |= B_INVAL; 418 error = bwrite(bp); 419 if (error) 420 allerror = error; 421 bap = copy; 422 423 /* 424 * Recursively free totally unused blocks. 425 */ 426 for (i = NINDIR(fs) - 1, nlbn = lbn + 1 - i * factor; i > last; 427 i--, nlbn += factor) { 428 nb = bap[i]; 429 if (nb == 0) 430 continue; 431 if (level > SINGLE) { 432 if ((error = ext2_indirtrunc(ip, nlbn, 433 fsbtodb(fs, nb), (int32_t)-1, level - 1, &blkcount)) != 0) 434 allerror = error; 435 blocksreleased += blkcount; 436 } 437 ext2_blkfree(ip, nb, fs->s_blocksize); 438 blocksreleased += nblocks; 439 } 440 441 /* 442 * Recursively free last partial block. 443 */ 444 if (level > SINGLE && lastbn >= 0) { 445 last = lastbn % factor; 446 nb = bap[i]; 447 if (nb != 0) { 448 if ((error = ext2_indirtrunc(ip, nlbn, fsbtodb(fs, nb), 449 last, level - 1, &blkcount)) != 0) 450 allerror = error; 451 blocksreleased += blkcount; 452 } 453 } 454 FREE(copy, M_TEMP); 455 *countp = blocksreleased; 456 return (allerror); 457} 458 459/* 460 * discard preallocated blocks 461 */ 462int 463ext2_inactive(ap) 464 struct vop_inactive_args /* { 465 struct vnode *a_vp; 466 struct thread *a_td; 467 } */ *ap; 468{ 469 struct vnode *vp = ap->a_vp; 470 struct inode *ip = VTOI(vp); 471 struct thread *td = ap->a_td; 472 int mode, error = 0; 473 474 ext2_discard_prealloc(ip); 475 if (prtactive && vrefcnt(vp) != 0) 476 vprint("ext2_inactive: pushing active", vp); 477 478 /* 479 * Ignore inodes related to stale file handles. 480 */ 481 if (ip->i_mode == 0) 482 goto out; 483 if (ip->i_nlink <= 0) { 484 (void) vn_write_suspend_wait(vp, NULL, V_WAIT); 485 error = ext2_truncate(vp, (off_t)0, 0, NOCRED, td); 486 ip->i_rdev = 0; 487 mode = ip->i_mode; 488 ip->i_mode = 0; 489 ip->i_flag |= IN_CHANGE | IN_UPDATE; 490 ext2_vfree(vp, ip->i_number, mode); 491 } 492 if (ip->i_flag & (IN_ACCESS | IN_CHANGE | IN_MODIFIED | IN_UPDATE)) { 493 if ((ip->i_flag & (IN_CHANGE | IN_UPDATE | IN_MODIFIED)) == 0 && 494 vn_write_suspend_wait(vp, NULL, V_NOWAIT)) { 495 ip->i_flag &= ~IN_ACCESS; 496 } else { 497 (void) vn_write_suspend_wait(vp, NULL, V_WAIT); 498 ext2_update(vp, 0); 499 } 500 } 501out: 502 VOP_UNLOCK(vp, 0, td); 503 /* 504 * If we are done with the inode, reclaim it 505 * so that it can be reused immediately. 506 */ 507 if (ip->i_mode == 0) 508 vrecycle(vp, NULL, td); 509 return (error); 510} 511 512/* 513 * Reclaim an inode so that it can be used for other purposes. 514 */ 515int 516ext2_reclaim(ap) 517 struct vop_reclaim_args /* { 518 struct vnode *a_vp; 519 struct thread *a_td; 520 } */ *ap; 521{ 522 struct inode *ip; 523 struct vnode *vp = ap->a_vp; 524 525 if (prtactive && vrefcnt(vp) != 0) 526 vprint("ufs_reclaim: pushing active", vp); 527 ip = VTOI(vp); 528 if (ip->i_flag & IN_LAZYMOD) { 529 ip->i_flag |= IN_MODIFIED; 530 ext2_update(vp, 0); 531 } 532 /* 533 * Remove the inode from its hash chain. 534 */ 535 ext2_ihashrem(ip); 536 /* 537 * Purge old data structures associated with the inode. 538 */ 539 if (ip->i_devvp) { 540 vrele(ip->i_devvp); 541 ip->i_devvp = 0; 542 } 543 FREE(vp->v_data, M_EXT2NODE); 544 vp->v_data = 0; 545 return (0); 546}
|