Lines Matching refs:tx

51 static boolean_t dbuf_undirty(dmu_buf_impl_t *db, dmu_tx_t *tx);
52 static void dbuf_write(dbuf_dirty_record_t *dr, arc_buf_t *data, dmu_tx_t *tx);
815 dbuf_free_range(dnode_t *dn, uint64_t start, uint64_t end, dmu_tx_t *tx)
818 uint64_t txg = tx->tx_txg;
856 dbuf_will_dirty(db, tx);
871 if (dbuf_undirty(db, tx)) {
963 dbuf_new_size(dmu_buf_impl_t *db, int size, dmu_tx_t *tx)
988 dbuf_will_dirty(db, tx);
1006 ASSERT3U(db->db_last_dirty->dr_txg, ==, tx->tx_txg);
1011 dnode_willuse_space(dn, size-osize, tx);
1030 dbuf_dirty(dmu_buf_impl_t *db, dmu_tx_t *tx)
1037 int txgoff = tx->tx_txg & TXG_MASK;
1039 ASSERT(tx->tx_txg != 0);
1041 DMU_TX_DIRTY_BUF(tx, db);
1050 ASSERT(!dmu_tx_is_syncing(tx) ||
1061 (dmu_tx_is_syncing(tx) ? DN_DIRTY_SYNC : DN_DIRTY_OPEN));
1081 (dmu_tx_is_syncing(tx) ? DN_DIRTY_SYNC : DN_DIRTY_OPEN);
1094 ASSERT(*drp == NULL || (*drp)->dr_txg <= tx->tx_txg ||
1096 while ((dr = *drp) != NULL && dr->dr_txg > tx->tx_txg)
1098 if (dr && dr->dr_txg == tx->tx_txg) {
1120 (dmu_tx_is_syncing(tx) ? DN_DIRTY_SYNC : DN_DIRTY_OPEN));
1126 dn->dn_next_nlevels[(tx->tx_txg-1) & TXG_MASK] > db->db_level ||
1127 dn->dn_next_nlevels[(tx->tx_txg-2) & TXG_MASK] > db->db_level);
1137 ASSERT(!dmu_tx_is_syncing(tx) || DMU_OBJECT_IS_SPECIAL(dn->dn_object) ||
1151 dnode_willuse_space(dn, db->db.db_size, tx);
1166 dbuf_fix_old_data(db, tx->tx_txg);
1179 dbuf_fix_old_data(db, tx->tx_txg);
1192 dr->dr_txg = tx->tx_txg;
1204 dnode_clear_range(dn, db->db_blkid, 1, tx);
1212 dbuf_add_ref(db, (void *)(uintptr_t)tx->tx_txg);
1224 dnode_setdirty(dn, tx);
1240 dnode_willuse_space(dn, -willfree, tx);
1249 dnode_new_blkid(dn, db->db_blkid, tx, drop_struct_lock);
1269 di = dbuf_dirty(parent, tx);
1278 ASSERT3U(di->dr_txg, ==, tx->tx_txg);
1297 dnode_setdirty(dn, tx);
1307 dbuf_undirty(dmu_buf_impl_t *db, dmu_tx_t *tx)
1310 uint64_t txg = tx->tx_txg;
1395 dbuf_will_dirty(dmu_buf_impl_t *db, dmu_tx_t *tx)
1399 ASSERT(tx->tx_txg != 0);
1407 (void) dbuf_dirty(db, tx);
1411 dmu_buf_will_not_fill(dmu_buf_t *db_fake, dmu_tx_t *tx)
1417 dmu_buf_will_fill(db_fake, tx);
1421 dmu_buf_will_fill(dmu_buf_t *db_fake, dmu_tx_t *tx)
1426 ASSERT(tx->tx_txg != 0);
1431 dmu_tx_private_ok(tx));
1434 (void) dbuf_dirty(db, tx);
1440 dbuf_fill_done(dmu_buf_impl_t *db, dmu_tx_t *tx)
1464 dbuf_assign_arcbuf(dmu_buf_impl_t *db, arc_buf_t *buf, dmu_tx_t *tx)
1472 ASSERT(tx->tx_txg != 0);
1487 (void) dbuf_dirty(db, tx);
1499 if (dr != NULL && dr->dr_txg == tx->tx_txg) {
1518 (void) dbuf_dirty(db, tx);
1519 dbuf_fill_done(db, tx);
1981 dbuf_spill_set_blksz(dmu_buf_t *db_fake, uint64_t blksz, dmu_tx_t *tx)
1998 dbuf_new_size(db, blksz, tx);
2006 dbuf_rm_spill(dnode_t *dn, dmu_tx_t *tx)
2008 dbuf_free_range(dn, DMU_SPILL_BLKID, DMU_SPILL_BLKID, tx);
2210 /* ASSERT(dmu_tx_is_syncing(tx) */
2253 dbuf_sync_indirect(dbuf_dirty_record_t *dr, dmu_tx_t *tx)
2259 ASSERT(dmu_tx_is_syncing(tx));
2288 dbuf_write(dr, db->db_buf, tx);
2292 dbuf_sync_list(&dr->dt.di.dr_children, tx);
2299 dbuf_sync_leaf(dbuf_dirty_record_t *dr, dmu_tx_t *tx)
2305 uint64_t txg = tx->tx_txg;
2307 ASSERT(dmu_tx_is_syncing(tx));
2418 dbuf_write(dr, *datap, tx);
2438 dbuf_sync_list(list_t *list, dmu_tx_t *tx)
2457 dbuf_sync_indirect(dr, tx);
2459 dbuf_sync_leaf(dr, tx);
2561 dmu_tx_t *tx;
2565 tx = os->os_synctx;
2567 (void) dsl_dataset_block_kill(ds, bp_orig, tx, B_TRUE);
2568 dsl_dataset_block_born(ds, bp, tx);
2677 dbuf_write(dbuf_dirty_record_t *dr, arc_buf_t *data, dmu_tx_t *tx)
2683 uint64_t txg = tx->tx_txg;