Searched refs:TXG_MASK (Results 1 - 24 of 24) sorted by relevance

/freebsd-13-stable/sys/contrib/openzfs/include/sys/
H A Dtxg.h41 #define TXG_MASK (TXG_SIZE - 1) /* mask for size */ macro
43 #define TXG_IDX (txg & TXG_MASK)
/freebsd-13-stable/sys/contrib/openzfs/module/zfs/
H A Dtxg.c335 tc->tc_count[txg & TXG_MASK]++;
357 int g = th->th_txg & TXG_MASK;
368 int g = th->th_txg & TXG_MASK;
390 int g = txg & TXG_MASK;
461 int g = txg & TXG_MASK;
864 return (tl->tl_head[txg & TXG_MASK] == NULL);
917 int t = txg & TXG_MASK;
942 int t = txg & TXG_MASK;
970 int t = txg & TXG_MASK;
995 int t = txg & TXG_MASK;
[all...]
H A Dvdev_initialize.c77 uint64_t last_offset = vd->vdev_initialize_offset[txg & TXG_MASK];
78 vd->vdev_initialize_offset[txg & TXG_MASK] = 0;
174 &vd->vdev_initialize_offset[zio->io_txg & TXG_MASK];
217 if (vd->vdev_initialize_offset[txg & TXG_MASK] == 0) {
242 vd->vdev_initialize_offset[txg & TXG_MASK] = start + size;
243 zio_nowait(zio_write_phys(spa->spa_txg_zio[txg & TXG_MASK], vd, start,
H A Ddnode.c404 dn->dn_next_bonuslen[tx->tx_txg & TXG_MASK] = DN_ZERO_BONUSLEN;
406 dn->dn_next_bonuslen[tx->tx_txg & TXG_MASK] = dn->dn_bonuslen;
417 dn->dn_next_bonustype[tx->tx_txg & TXG_MASK] = dn->dn_bonustype;
427 dn->dn_rm_spillblk[tx->tx_txg & TXG_MASK] = DN_KILL_SPILLBLK;
657 dn->dn_next_indblkshift[tx->tx_txg & TXG_MASK] = ibs;
658 dn->dn_next_bonuslen[tx->tx_txg & TXG_MASK] = dn->dn_bonuslen;
659 dn->dn_next_bonustype[tx->tx_txg & TXG_MASK] = dn->dn_bonustype;
660 dn->dn_next_blksz[tx->tx_txg & TXG_MASK] = dn->dn_datablksz;
701 dn->dn_next_blksz[tx->tx_txg & TXG_MASK] = blocksize;
704 dn->dn_next_bonuslen[tx->tx_txg & TXG_MASK]
[all...]
H A Dvdev_removal.c560 int txgoff = (txg + i) & TXG_MASK;
619 svr->svr_bytes_done[txg & TXG_MASK] += inflight_size;
622 ASSERT0(svr->svr_max_offset_to_sync[TXG_CLEAN(txg) & TXG_MASK]);
643 svr->svr_bytes_done[txg & TXG_MASK] += size;
736 &svr->svr_new_segments[txg & TXG_MASK], tx);
745 range_tree_vacate(svr->svr_frees[txg & TXG_MASK],
747 ASSERT3U(svr->svr_max_offset_to_sync[txg & TXG_MASK], >=,
749 svr->svr_max_offset_to_sync[txg & TXG_MASK] = 0;
1059 zio_t *nzio = zio_null(spa->spa_txg_zio[txg & TXG_MASK], spa, NULL,
1074 list_insert_tail(&svr->svr_new_segments[txg & TXG_MASK], entr
[all...]
H A Ddmu_redact.c520 list_t *list = &md->md_blocks[txg & TXG_MASK];
522 &md->md_furthest[txg & TXG_MASK];
556 md->md_synctask_txg[txg & TXG_MASK] = B_FALSE;
569 if (!md->md_synctask_txg[txg & TXG_MASK]) {
572 md->md_synctask_txg[txg & TXG_MASK] = B_TRUE;
575 md->md_furthest[txg & TXG_MASK].rbp_object = object;
576 md->md_furthest[txg & TXG_MASK].rbp_blkid = blkid;
577 list_move_tail(&md->md_blocks[txg & TXG_MASK],
H A Dvdev_rebuild.c198 if (vr->vr_scan_offset[txg & TXG_MASK] > 0) {
199 vrp->vrp_last_offset = vr->vr_scan_offset[txg & TXG_MASK];
200 vr->vr_scan_offset[txg & TXG_MASK] = 0;
483 uint64_t *off = &vr->vr_scan_offset[zio->io_txg & TXG_MASK];
581 if (vr->vr_scan_offset[txg & TXG_MASK] == 0) {
582 vr->vr_scan_offset[txg & TXG_MASK] = start;
601 vr->vr_scan_offset[txg & TXG_MASK] = start + size;
605 zio_nowait(zio_read(spa->spa_txg_zio[txg & TXG_MASK], spa, &blk,
H A Ddsl_pool.c672 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK] == 0);
673 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK] = 0;
709 !os->os_next_write_raw[txg & TXG_MASK]) {
729 !os->os_next_write_raw[txg & TXG_MASK]) {
774 dsl_pool_undirty_space(dp, dp->dp_dirty_pertxg[txg & TXG_MASK], txg);
918 dp->dp_dirty_pertxg[tx->tx_txg & TXG_MASK] += space;
932 if (dp->dp_dirty_pertxg[txg & TXG_MASK] < space) {
934 space = dp->dp_dirty_pertxg[txg & TXG_MASK];
936 ASSERT3U(dp->dp_dirty_pertxg[txg & TXG_MASK], >=, space);
937 dp->dp_dirty_pertxg[txg & TXG_MASK]
[all...]
H A Dvdev_trim.c208 uint64_t last_offset = vd->vdev_trim_offset[txg & TXG_MASK];
209 vd->vdev_trim_offset[txg & TXG_MASK] = 0;
377 &vd->vdev_trim_offset[zio->io_txg & TXG_MASK];
512 vd->vdev_trim_offset[txg & TXG_MASK] == 0) {
540 vd->vdev_trim_offset[txg & TXG_MASK] = start + size;
550 zio_nowait(zio_trim(spa->spa_txg_zio[txg & TXG_MASK], vd,
H A Dzil.c642 if (txg_list_member(&dp->dp_dirty_zilogs, zilog, txg & TXG_MASK))
868 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE;
898 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE;
1908 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK];
1952 itxg = &zilog->zl_itxg[txg & TXG_MASK];
2023 itxg_t *itxg = &zilog->zl_itxg[synced_txg & TXG_MASK];
2076 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK];
2121 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK];
3019 uint64_t *replayed_seq = &zilog->zl_replayed_seq[txg & TXG_MASK];
3635 zilog->zl_replayed_seq[dmu_tx_get_txg(tx) & TXG_MASK]
[all...]
H A Ddsl_dir.c1158 ASSERT0(dd->dd_tempreserved[tx->tx_txg & TXG_MASK]);
1160 dd->dd_space_towrite[tx->tx_txg & TXG_MASK] / 1024);
1161 dd->dd_space_towrite[tx->tx_txg & TXG_MASK] = 0;
1176 space += dd->dd_space_towrite[i & TXG_MASK];
1177 ASSERT3U(dd->dd_space_towrite[i & TXG_MASK], >=, 0);
1355 dd->dd_tempreserved[txg & TXG_MASK] += asize;
1452 int txgidx = tx->tx_txg & TXG_MASK;
1497 dd->dd_space_towrite[tx->tx_txg & TXG_MASK] += space;
H A Ddnode_sync.c45 int txgoff = tx->tx_txg & TXG_MASK;
562 int txgoff = tx->tx_txg & TXG_MASK;
628 int txgoff = tx->tx_txg & TXG_MASK;
H A Dmetaslab.c1936 range_tree_space(msp->ms_allocating[(txg + t) & TXG_MASK]);
3659 range_tree_walk(msp->ms_allocating[(txg + t) & TXG_MASK],
3923 range_tree_t *alloctree = msp->ms_allocating[txg & TXG_MASK];
4187 ASSERT0(range_tree_space(msp->ms_allocating[txg & TXG_MASK]));
4189 & TXG_MASK]));
4216 msp->ms_allocating[(txg + t) & TXG_MASK]));
4350 ASSERT0(range_tree_space(msp->ms_allocating[txg & TXG_MASK]));
4579 if (range_tree_is_empty(msp->ms_allocating[txg & TXG_MASK]))
4582 range_tree_add(msp->ms_allocating[txg & TXG_MASK], start, size);
5552 range_tree_remove(msp->ms_allocating[txg & TXG_MASK],
[all...]
H A Dvdev_indirect.c552 &sci->sci_new_mapping_entries[txg & TXG_MASK], tx);
553 ASSERT(list_is_empty(&sci->sci_new_mapping_entries[txg & TXG_MASK]));
571 int txgoff = dmu_tx_get_txg(tx) & TXG_MASK;
H A Ddmu_object.c478 dn->dn_next_type[tx->tx_txg & TXG_MASK] = dn->dn_type =
H A Ddsl_dataset.c1261 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE;
1446 !os->os_next_write_raw[tx->tx_txg & TXG_MASK]) {
2080 if (ds->ds_resume_bytes[tx->tx_txg & TXG_MASK] != 0) {
2083 &ds->ds_resume_object[tx->tx_txg & TXG_MASK], tx));
2086 &ds->ds_resume_offset[tx->tx_txg & TXG_MASK], tx));
2089 &ds->ds_resume_bytes[tx->tx_txg & TXG_MASK], tx));
2090 ds->ds_resume_object[tx->tx_txg & TXG_MASK] = 0;
2091 ds->ds_resume_offset[tx->tx_txg & TXG_MASK] = 0;
2092 ds->ds_resume_bytes[tx->tx_txg & TXG_MASK] = 0;
2273 os->os_next_write_raw[tx->tx_txg & TXG_MASK]
[all...]
H A Ddmu_objset.c1066 mdn->dn_next_nlevels[tx->tx_txg & TXG_MASK] =
1650 os->os_next_write_raw[tx->tx_txg & TXG_MASK]) {
1684 txgoff = tx->tx_txg & TXG_MASK;
1739 return (!multilist_is_empty(&os->os_dirty_dnodes[txg & TXG_MASK]));
H A Ddbuf.c2001 int txgoff = tx->tx_txg & TXG_MASK;
2051 int txgoff = tx->tx_txg & TXG_MASK;
2259 dn->dn_next_nlevels[(tx->tx_txg-1) & TXG_MASK] > db->db_level ||
2260 dn->dn_next_nlevels[(tx->tx_txg-2) & TXG_MASK] > db->db_level);
2403 list_remove(&dn->dn_dirty_records[txg & TXG_MASK], dr);
4317 list_insert_tail(&dn->dn_dirty_records[txg & TXG_MASK], dr);
H A Dspa_stats.c414 ts->ndirty = dp->dp_dirty_pertxg[txg & TXG_MASK];
H A Ddmu.c888 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK];
915 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK] +=
H A Ddsl_crypt.c2109 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE;
H A Dspa.c9014 bplist_t *free_bpl = &spa->spa_free_bplist[txg & TXG_MASK];
9166 (void) zio_wait(spa->spa_txg_zio[txg & TXG_MASK]);
9167 spa->spa_txg_zio[txg & TXG_MASK] = zio_root(spa, NULL, NULL,
9263 ASSERT0(spa->spa_vdev_removal->svr_bytes_done[txg & TXG_MASK]);
H A Ddmu_recv.c1292 int txgoff = dmu_tx_get_txg(tx) & TXG_MASK;
H A Dzio.c1227 bplist_append(&spa->spa_free_bplist[txg & TXG_MASK], bp);

Completed in 267 milliseconds