• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/fs/jfs/

Lines Matching defs:log

165 static int diLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
167 static int dataLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
169 static void dtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
171 static void mapLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
176 static int txLog(struct jfs_log * log, struct tblock * tblk,
180 static void xtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
366 struct jfs_log *log;
369 log = JFS_SBI(sb)->log;
380 if (test_bit(log_SYNCBARRIER, &log->flag) ||
381 test_bit(log_QUIESCE, &log->flag)) {
383 TXN_SLEEP(&log->syncwait);
435 ++log->logtid;
436 tblk->logtid = log->logtid;
438 ++log->active;
463 struct jfs_log *log;
465 log = JFS_SBI(sb)->log;
474 if (test_bit(log_SYNCBARRIER, &log->flag) ||
475 test_bit(log_QUIESCE, &log->flag)) {
477 TXN_SLEEP(&log->syncwait);
504 struct jfs_log *log;
515 log = JFS_SBI(tblk->sb)->log;
529 spin_lock_irq(&log->gclock); // LOGGC_LOCK
531 spin_unlock_irq(&log->gclock); // LOGGC_UNLOCK
548 if (--log->active == 0) {
549 clear_bit(log_FLUSH, &log->flag);
554 if (test_bit(log_SYNCBARRIER, &log->flag)) {
557 /* write dirty metadata & forward log syncpt */
558 jfs_syncpt(log, 1);
560 jfs_info("log barrier off: 0x%x", log->lsn);
563 clear_bit(log_SYNCBARRIER, &log->flag);
566 TXN_WAKEUP(&log->syncwait);
915 struct jfs_log *log;
920 log = JFS_SBI(tblk->sb)->log;
923 * mark page under tlock homeok (its log has been written):
944 LOGSYNC_LOCK(log, flags);
946 logdiff(difft, tblk->clsn, log);
947 logdiff(diffp, mp->clsn, log);
952 LOGSYNC_UNLOCK(log, flags);
983 LOGSYNC_LOCK(log, flags);
984 log->count--;
986 LOGSYNC_UNLOCK(log, flags);
1067 * function: allocate a transaction lock for log vector list
1143 struct jfs_log *log;
1170 log = JFS_SBI(sb)->log;
1171 cd.log = log;
1173 /* initialize log record descriptor in commit */
1270 * write log records from transaction locks
1274 if ((rc = txLog(log, tblk, &cd)))
1309 * write COMMIT log record
1313 lsn = lmLog(log, tblk, lrd, NULL);
1315 lmGroupCommit(log, tblk);
1375 * FUNCTION: Writes AFTER log records for all lines modified
1383 static int txLog(struct jfs_log * log, struct tblock * tblk, struct commit * cd)
1392 * write log record(s) for each tlock of transaction,
1402 lrd->log.redopage.fileset = cpu_to_le32(JFS_IP(ip)->fileset);
1403 lrd->log.redopage.inode = cpu_to_le32(ip->i_ino);
1405 /* write log record of page from the tlock */
1408 xtLog(log, tblk, lrd, tlck);
1412 dtLog(log, tblk, lrd, tlck);
1416 diLog(log, tblk, lrd, tlck, cd);
1420 mapLog(log, tblk, lrd, tlck);
1424 dataLog(log, tblk, lrd, tlck);
1438 * function: log inode tlock and format maplock to update bmap;
1440 static int diLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
1451 lrd->log.redopage.type = cpu_to_le16(LOG_INODE);
1452 lrd->log.redopage.l2linesize = cpu_to_le16(L2INODESLOTSIZE);
1454 pxd = &lrd->log.redopage.pxd;
1460 /* log after-image for logredo(): */
1465 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
1480 * page before the log;
1483 /* log LOG_NOREDOINOEXT of the freed inode extent for
1495 lrd->log.noredoinoext.iagnum =
1497 lrd->log.noredoinoext.inoext_idx =
1502 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL));
1521 /* log LOG_UPDATEMAP for logredo() to update bmap for
1529 lrd->log.updatemap.type =
1532 lrd->log.updatemap.type =
1534 lrd->log.updatemap.nxd = cpu_to_le16(1);
1535 lrd->log.updatemap.pxd = pxdlock->pxd;
1537 cpu_to_le32(lmLog(log, tblk, lrd, NULL));
1551 * function: log data tlock
1553 static int dataLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
1562 lrd->log.redopage.type = cpu_to_le16(LOG_DATA);
1563 lrd->log.redopage.l2linesize = cpu_to_le16(L2DATASLOTSIZE);
1565 pxd = &lrd->log.redopage.pxd;
1567 /* log after-image for logredo(): */
1586 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
1597 * function: log dtree tlock and format maplock to update bmap;
1599 static void dtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
1609 lrd->log.redopage.type = cpu_to_le16(LOG_DTREE);
1610 lrd->log.redopage.l2linesize = cpu_to_le16(L2DTSLOTSIZE);
1612 pxd = &lrd->log.redopage.pxd;
1615 lrd->log.redopage.type |= cpu_to_le16(LOG_BTROOT);
1624 /* log after-image of the new page for logredo():
1625 * mark log (LOG_NEW) for logredo() to initialize
1630 lrd->log.redopage.type |= cpu_to_le16(LOG_EXTEND);
1632 lrd->log.redopage.type |= cpu_to_le16(LOG_NEW);
1636 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
1660 /* log after-image for logredo(): */
1665 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
1680 /* log LOG_NOREDOPAGE of the deleted page for logredo()
1687 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL));
1700 * function: log xtree tlock and format maplock to update bmap;
1702 static void xtLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
1719 lrd->log.redopage.type = cpu_to_le16(LOG_XTREE);
1720 lrd->log.redopage.l2linesize = cpu_to_le16(L2XTSLOTSIZE);
1722 page_pxd = &lrd->log.redopage.pxd;
1725 lrd->log.redopage.type |= cpu_to_le16(LOG_BTROOT);
1728 lrd->log.redopage.type |=
1744 /* log after-image for logredo():
1755 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
1818 * (page will be invalidated after log is written and bmap
1822 /* LOG_NOREDOPAGE log for NoRedoPage filter:
1831 * log for the page itself is generated from processing
1834 /* if page free from file truncation, log LOG_NOREDOPAGE
1846 cpu_to_le32(lmLog(log, tblk, lrd, NULL));
1852 cpu_to_le32(lmLog(log, tblk, lrd, tlck));
1861 lrd->log.updatemap.type = cpu_to_le16(LOG_FREEXADLIST);
1864 lrd->log.updatemap.nxd =
1870 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
1952 * write log records
1954 /* log after-image for logredo():
1966 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, tlck));
1980 lrd->log.updatemap.type = cpu_to_le16(LOG_FREEPXD);
1981 lrd->log.updatemap.nxd = cpu_to_le16(1);
1982 lrd->log.updatemap.pxd = pxdlock->pxd;
1985 cpu_to_le32(lmLog(log, tblk, lrd, NULL));
1997 lrd->log.updatemap.type =
2001 lrd->log.updatemap.nxd =
2008 cpu_to_le32(lmLog(log, tblk, lrd, tlck));
2086 * function: log from maplock of freed data extents;
2088 static void mapLog(struct jfs_log * log, struct tblock * tblk, struct lrd * lrd,
2103 /* log LOG_NOREDOPAGE of the old relocated page
2108 pxd = &lrd->log.redopage.pxd;
2110 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL));
2114 * if page free from relocation, LOG_UPDATEMAP log is
2120 * avoiding an extra log write);
2123 lrd->log.updatemap.type = cpu_to_le16(LOG_FREEPXD);
2124 lrd->log.updatemap.nxd = cpu_to_le16(1);
2125 lrd->log.updatemap.pxd = pxdlock->pxd;
2126 lrd->backchain = cpu_to_le32(lmLog(log, tblk, lrd, NULL));
2140 /* log LOG_UPDATEMAP for logredo() to update bmap for
2150 lrd->log.updatemap.type =
2153 lrd->log.updatemap.type =
2155 lrd->log.updatemap.nxd = cpu_to_le16(1);
2156 lrd->log.updatemap.pxd = pxdlock->pxd;
2158 cpu_to_le32(lmLog(log, tblk, lrd, NULL));
2423 * lsn - log sequence number;
2625 * log age of page-frames in memory for which caller has
2687 struct jfs_log *log;
2701 log = (struct jfs_log *) JFS_SBI(tblk->sb)->log;
2703 spin_lock_irq(&log->gclock); // LOGGC_LOCK
2708 log->gcrtc--;
2713 * Can't release log->gclock until we've tested tblk->flag
2716 spin_unlock_irq(&log->gclock); // LOGGC_UNLOCK
2721 spin_unlock_irq(&log->gclock); // LOGGC_UNLOCK
2829 struct jfs_log *log = mp->log;
2832 assert(log);
2850 struct jfs_log *log = JFS_SBI(sb)->log;
2853 set_bit(log_QUIESCE, &log->flag);
2895 jfs_flush_journal(log, 0);
2905 struct jfs_log *log = JFS_SBI(sb)->log;
2907 clear_bit(log_QUIESCE, &log->flag);
2908 TXN_WAKEUP(&log->syncwait);