/freebsd-13-stable/sys/contrib/openzfs/include/sys/ |
H A D | txg.h | 41 #define TXG_MASK (TXG_SIZE - 1) /* mask for size */ macro 43 #define TXG_IDX (txg & TXG_MASK)
|
/freebsd-13-stable/sys/contrib/openzfs/module/zfs/ |
H A D | txg.c | 335 tc->tc_count[txg & TXG_MASK]++; 357 int g = th->th_txg & TXG_MASK; 368 int g = th->th_txg & TXG_MASK; 390 int g = txg & TXG_MASK; 461 int g = txg & TXG_MASK; 864 return (tl->tl_head[txg & TXG_MASK] == NULL); 917 int t = txg & TXG_MASK; 942 int t = txg & TXG_MASK; 970 int t = txg & TXG_MASK; 995 int t = txg & TXG_MASK; [all...] |
H A D | vdev_initialize.c | 77 uint64_t last_offset = vd->vdev_initialize_offset[txg & TXG_MASK]; 78 vd->vdev_initialize_offset[txg & TXG_MASK] = 0; 174 &vd->vdev_initialize_offset[zio->io_txg & TXG_MASK]; 217 if (vd->vdev_initialize_offset[txg & TXG_MASK] == 0) { 242 vd->vdev_initialize_offset[txg & TXG_MASK] = start + size; 243 zio_nowait(zio_write_phys(spa->spa_txg_zio[txg & TXG_MASK], vd, start,
|
H A D | dnode.c | 404 dn->dn_next_bonuslen[tx->tx_txg & TXG_MASK] = DN_ZERO_BONUSLEN; 406 dn->dn_next_bonuslen[tx->tx_txg & TXG_MASK] = dn->dn_bonuslen; 417 dn->dn_next_bonustype[tx->tx_txg & TXG_MASK] = dn->dn_bonustype; 427 dn->dn_rm_spillblk[tx->tx_txg & TXG_MASK] = DN_KILL_SPILLBLK; 657 dn->dn_next_indblkshift[tx->tx_txg & TXG_MASK] = ibs; 658 dn->dn_next_bonuslen[tx->tx_txg & TXG_MASK] = dn->dn_bonuslen; 659 dn->dn_next_bonustype[tx->tx_txg & TXG_MASK] = dn->dn_bonustype; 660 dn->dn_next_blksz[tx->tx_txg & TXG_MASK] = dn->dn_datablksz; 701 dn->dn_next_blksz[tx->tx_txg & TXG_MASK] = blocksize; 704 dn->dn_next_bonuslen[tx->tx_txg & TXG_MASK] [all...] |
H A D | vdev_removal.c | 560 int txgoff = (txg + i) & TXG_MASK; 619 svr->svr_bytes_done[txg & TXG_MASK] += inflight_size; 622 ASSERT0(svr->svr_max_offset_to_sync[TXG_CLEAN(txg) & TXG_MASK]); 643 svr->svr_bytes_done[txg & TXG_MASK] += size; 736 &svr->svr_new_segments[txg & TXG_MASK], tx); 745 range_tree_vacate(svr->svr_frees[txg & TXG_MASK], 747 ASSERT3U(svr->svr_max_offset_to_sync[txg & TXG_MASK], >=, 749 svr->svr_max_offset_to_sync[txg & TXG_MASK] = 0; 1059 zio_t *nzio = zio_null(spa->spa_txg_zio[txg & TXG_MASK], spa, NULL, 1074 list_insert_tail(&svr->svr_new_segments[txg & TXG_MASK], entr [all...] |
H A D | dmu_redact.c | 520 list_t *list = &md->md_blocks[txg & TXG_MASK]; 522 &md->md_furthest[txg & TXG_MASK]; 556 md->md_synctask_txg[txg & TXG_MASK] = B_FALSE; 569 if (!md->md_synctask_txg[txg & TXG_MASK]) { 572 md->md_synctask_txg[txg & TXG_MASK] = B_TRUE; 575 md->md_furthest[txg & TXG_MASK].rbp_object = object; 576 md->md_furthest[txg & TXG_MASK].rbp_blkid = blkid; 577 list_move_tail(&md->md_blocks[txg & TXG_MASK],
|
H A D | vdev_rebuild.c | 198 if (vr->vr_scan_offset[txg & TXG_MASK] > 0) { 199 vrp->vrp_last_offset = vr->vr_scan_offset[txg & TXG_MASK]; 200 vr->vr_scan_offset[txg & TXG_MASK] = 0; 483 uint64_t *off = &vr->vr_scan_offset[zio->io_txg & TXG_MASK]; 581 if (vr->vr_scan_offset[txg & TXG_MASK] == 0) { 582 vr->vr_scan_offset[txg & TXG_MASK] = start; 601 vr->vr_scan_offset[txg & TXG_MASK] = start + size; 605 zio_nowait(zio_read(spa->spa_txg_zio[txg & TXG_MASK], spa, &blk,
|
H A D | dsl_pool.c | 672 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK] == 0); 673 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK] = 0; 709 !os->os_next_write_raw[txg & TXG_MASK]) { 729 !os->os_next_write_raw[txg & TXG_MASK]) { 774 dsl_pool_undirty_space(dp, dp->dp_dirty_pertxg[txg & TXG_MASK], txg); 918 dp->dp_dirty_pertxg[tx->tx_txg & TXG_MASK] += space; 932 if (dp->dp_dirty_pertxg[txg & TXG_MASK] < space) { 934 space = dp->dp_dirty_pertxg[txg & TXG_MASK]; 936 ASSERT3U(dp->dp_dirty_pertxg[txg & TXG_MASK], >=, space); 937 dp->dp_dirty_pertxg[txg & TXG_MASK] [all...] |
H A D | vdev_trim.c | 208 uint64_t last_offset = vd->vdev_trim_offset[txg & TXG_MASK]; 209 vd->vdev_trim_offset[txg & TXG_MASK] = 0; 377 &vd->vdev_trim_offset[zio->io_txg & TXG_MASK]; 512 vd->vdev_trim_offset[txg & TXG_MASK] == 0) { 540 vd->vdev_trim_offset[txg & TXG_MASK] = start + size; 550 zio_nowait(zio_trim(spa->spa_txg_zio[txg & TXG_MASK], vd,
|
H A D | zil.c | 642 if (txg_list_member(&dp->dp_dirty_zilogs, zilog, txg & TXG_MASK)) 868 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE; 898 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE; 1908 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; 1952 itxg = &zilog->zl_itxg[txg & TXG_MASK]; 2023 itxg_t *itxg = &zilog->zl_itxg[synced_txg & TXG_MASK]; 2076 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; 2121 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; 3019 uint64_t *replayed_seq = &zilog->zl_replayed_seq[txg & TXG_MASK]; 3635 zilog->zl_replayed_seq[dmu_tx_get_txg(tx) & TXG_MASK] [all...] |
H A D | dsl_dir.c | 1158 ASSERT0(dd->dd_tempreserved[tx->tx_txg & TXG_MASK]); 1160 dd->dd_space_towrite[tx->tx_txg & TXG_MASK] / 1024); 1161 dd->dd_space_towrite[tx->tx_txg & TXG_MASK] = 0; 1176 space += dd->dd_space_towrite[i & TXG_MASK]; 1177 ASSERT3U(dd->dd_space_towrite[i & TXG_MASK], >=, 0); 1355 dd->dd_tempreserved[txg & TXG_MASK] += asize; 1452 int txgidx = tx->tx_txg & TXG_MASK; 1497 dd->dd_space_towrite[tx->tx_txg & TXG_MASK] += space;
|
H A D | dnode_sync.c | 45 int txgoff = tx->tx_txg & TXG_MASK; 562 int txgoff = tx->tx_txg & TXG_MASK; 628 int txgoff = tx->tx_txg & TXG_MASK;
|
H A D | metaslab.c | 1936 range_tree_space(msp->ms_allocating[(txg + t) & TXG_MASK]); 3659 range_tree_walk(msp->ms_allocating[(txg + t) & TXG_MASK], 3923 range_tree_t *alloctree = msp->ms_allocating[txg & TXG_MASK]; 4187 ASSERT0(range_tree_space(msp->ms_allocating[txg & TXG_MASK])); 4189 & TXG_MASK])); 4216 msp->ms_allocating[(txg + t) & TXG_MASK])); 4350 ASSERT0(range_tree_space(msp->ms_allocating[txg & TXG_MASK])); 4579 if (range_tree_is_empty(msp->ms_allocating[txg & TXG_MASK])) 4582 range_tree_add(msp->ms_allocating[txg & TXG_MASK], start, size); 5552 range_tree_remove(msp->ms_allocating[txg & TXG_MASK], [all...] |
H A D | vdev_indirect.c | 552 &sci->sci_new_mapping_entries[txg & TXG_MASK], tx); 553 ASSERT(list_is_empty(&sci->sci_new_mapping_entries[txg & TXG_MASK])); 571 int txgoff = dmu_tx_get_txg(tx) & TXG_MASK;
|
H A D | dmu_object.c | 478 dn->dn_next_type[tx->tx_txg & TXG_MASK] = dn->dn_type =
|
H A D | dsl_dataset.c | 1261 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE; 1446 !os->os_next_write_raw[tx->tx_txg & TXG_MASK]) { 2080 if (ds->ds_resume_bytes[tx->tx_txg & TXG_MASK] != 0) { 2083 &ds->ds_resume_object[tx->tx_txg & TXG_MASK], tx)); 2086 &ds->ds_resume_offset[tx->tx_txg & TXG_MASK], tx)); 2089 &ds->ds_resume_bytes[tx->tx_txg & TXG_MASK], tx)); 2090 ds->ds_resume_object[tx->tx_txg & TXG_MASK] = 0; 2091 ds->ds_resume_offset[tx->tx_txg & TXG_MASK] = 0; 2092 ds->ds_resume_bytes[tx->tx_txg & TXG_MASK] = 0; 2273 os->os_next_write_raw[tx->tx_txg & TXG_MASK] [all...] |
H A D | dmu_objset.c | 1066 mdn->dn_next_nlevels[tx->tx_txg & TXG_MASK] = 1650 os->os_next_write_raw[tx->tx_txg & TXG_MASK]) { 1684 txgoff = tx->tx_txg & TXG_MASK; 1739 return (!multilist_is_empty(&os->os_dirty_dnodes[txg & TXG_MASK]));
|
H A D | dbuf.c | 2001 int txgoff = tx->tx_txg & TXG_MASK; 2051 int txgoff = tx->tx_txg & TXG_MASK; 2259 dn->dn_next_nlevels[(tx->tx_txg-1) & TXG_MASK] > db->db_level || 2260 dn->dn_next_nlevels[(tx->tx_txg-2) & TXG_MASK] > db->db_level); 2403 list_remove(&dn->dn_dirty_records[txg & TXG_MASK], dr); 4317 list_insert_tail(&dn->dn_dirty_records[txg & TXG_MASK], dr);
|
H A D | spa_stats.c | 414 ts->ndirty = dp->dp_dirty_pertxg[txg & TXG_MASK];
|
H A D | dmu.c | 888 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK]; 915 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK] +=
|
H A D | dsl_crypt.c | 2109 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE;
|
H A D | spa.c | 9014 bplist_t *free_bpl = &spa->spa_free_bplist[txg & TXG_MASK]; 9166 (void) zio_wait(spa->spa_txg_zio[txg & TXG_MASK]); 9167 spa->spa_txg_zio[txg & TXG_MASK] = zio_root(spa, NULL, NULL, 9263 ASSERT0(spa->spa_vdev_removal->svr_bytes_done[txg & TXG_MASK]);
|
H A D | dmu_recv.c | 1292 int txgoff = dmu_tx_get_txg(tx) & TXG_MASK;
|
H A D | zio.c | 1227 bplist_append(&spa->spa_free_bplist[txg & TXG_MASK], bp);
|