| /f-stack/freebsd/contrib/openzfs/module/zfs/ |
| H A D | txg.c | 336 int g = th->th_txg & TXG_MASK; in txg_register_callbacks() 347 int g = th->th_txg & TXG_MASK; in txg_rele_to_sync() 369 int g = txg & TXG_MASK; in txg_quiesce() 439 int g = txg & TXG_MASK; in txg_dispatch_callbacks() 895 int t = txg & TXG_MASK; in txg_list_add() 920 int t = txg & TXG_MASK; in txg_list_add_tail() 948 int t = txg & TXG_MASK; in txg_list_remove() 973 int t = txg & TXG_MASK; in txg_list_remove_this() 997 int t = txg & TXG_MASK; in txg_list_member() 1010 int t = txg & TXG_MASK; in txg_list_head() [all …]
|
| H A D | dnode.c | 657 dn->dn_next_indblkshift[tx->tx_txg & TXG_MASK] = ibs; in dnode_allocate() 701 dn->dn_next_blksz[tx->tx_txg & TXG_MASK] = blocksize; in dnode_reallocate() 715 dn->dn_next_nblkptr[tx->tx_txg & TXG_MASK] = nblkptr; in dnode_reallocate() 1694 ASSERT0(dn->dn_next_bonuslen[txg & TXG_MASK]); in dnode_setdirty() 1695 ASSERT0(dn->dn_next_blksz[txg & TXG_MASK]); in dnode_setdirty() 1696 ASSERT0(dn->dn_next_bonustype[txg & TXG_MASK]); in dnode_setdirty() 1787 dn->dn_next_blksz[tx->tx_txg&TXG_MASK] = size; in dnode_set_blksz() 1790 dn->dn_next_indblkshift[tx->tx_txg&TXG_MASK] = ibs; in dnode_set_blksz() 1807 uint64_t txgoff = tx->tx_txg & TXG_MASK; in dnode_set_nlevels_impl() 1911 dn->dn_next_maxblkid[tx->tx_txg & TXG_MASK] = in dnode_new_blkid() [all …]
|
| H A D | dsl_pool.c | 673 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK] == 0); in dsl_pool_sync() 674 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK] = 0; in dsl_pool_sync() 710 !os->os_next_write_raw[txg & TXG_MASK]) { in dsl_pool_sync() 730 !os->os_next_write_raw[txg & TXG_MASK]) { in dsl_pool_sync() 775 dsl_pool_undirty_space(dp, dp->dp_dirty_pertxg[txg & TXG_MASK], txg); in dsl_pool_sync() 919 dp->dp_dirty_pertxg[tx->tx_txg & TXG_MASK] += space; in dsl_pool_dirty_space() 933 if (dp->dp_dirty_pertxg[txg & TXG_MASK] < space) { in dsl_pool_undirty_space() 935 space = dp->dp_dirty_pertxg[txg & TXG_MASK]; in dsl_pool_undirty_space() 937 ASSERT3U(dp->dp_dirty_pertxg[txg & TXG_MASK], >=, space); in dsl_pool_undirty_space() 938 dp->dp_dirty_pertxg[txg & TXG_MASK] -= space; in dsl_pool_undirty_space()
|
| H A D | vdev_initialize.c | 77 uint64_t last_offset = vd->vdev_initialize_offset[txg & TXG_MASK]; in vdev_initialize_zap_update_sync() 78 vd->vdev_initialize_offset[txg & TXG_MASK] = 0; in vdev_initialize_zap_update_sync() 174 &vd->vdev_initialize_offset[zio->io_txg & TXG_MASK]; in vdev_initialize_cb() 217 if (vd->vdev_initialize_offset[txg & TXG_MASK] == 0) { in vdev_initialize_write() 242 vd->vdev_initialize_offset[txg & TXG_MASK] = start + size; in vdev_initialize_write() 243 zio_nowait(zio_write_phys(spa->spa_txg_zio[txg & TXG_MASK], vd, start, in vdev_initialize_write()
|
| H A D | vdev_rebuild.c | 198 if (vr->vr_scan_offset[txg & TXG_MASK] > 0) { in vdev_rebuild_update_sync() 199 vrp->vrp_last_offset = vr->vr_scan_offset[txg & TXG_MASK]; in vdev_rebuild_update_sync() 200 vr->vr_scan_offset[txg & TXG_MASK] = 0; in vdev_rebuild_update_sync() 480 uint64_t *off = &vr->vr_scan_offset[zio->io_txg & TXG_MASK]; in vdev_rebuild_cb() 578 if (vr->vr_scan_offset[txg & TXG_MASK] == 0) { in vdev_rebuild_range() 579 vr->vr_scan_offset[txg & TXG_MASK] = start; in vdev_rebuild_range() 598 vr->vr_scan_offset[txg & TXG_MASK] = start + size; in vdev_rebuild_range() 602 zio_nowait(zio_read(spa->spa_txg_zio[txg & TXG_MASK], spa, &blk, in vdev_rebuild_range()
|
| H A D | vdev_removal.c | 560 int txgoff = (txg + i) & TXG_MASK; in free_from_removing_vdev() 619 svr->svr_bytes_done[txg & TXG_MASK] += inflight_size; in free_from_removing_vdev() 643 svr->svr_bytes_done[txg & TXG_MASK] += size; in free_from_removing_vdev() 736 &svr->svr_new_segments[txg & TXG_MASK], tx); in vdev_mapping_sync() 745 range_tree_vacate(svr->svr_frees[txg & TXG_MASK], in vdev_mapping_sync() 747 ASSERT3U(svr->svr_max_offset_to_sync[txg & TXG_MASK], >=, in vdev_mapping_sync() 749 svr->svr_max_offset_to_sync[txg & TXG_MASK] = 0; in vdev_mapping_sync() 1318 if (svr->svr_max_offset_to_sync[txg & TXG_MASK] == 0) { in spa_vdev_copy_impl() 1329 svr->svr_bytes_done[txg & TXG_MASK] += range_tree_space(segs); in spa_vdev_copy_impl() 1756 svr->svr_bytes_done[dmu_tx_get_txg(tx) & TXG_MASK] = 0; in spa_vdev_remove_cancel_sync() [all …]
|
| H A D | dmu_redact.c | 520 list_t *list = &md->md_blocks[txg & TXG_MASK]; in redaction_list_update_sync() 522 &md->md_furthest[txg & TXG_MASK]; in redaction_list_update_sync() 556 md->md_synctask_txg[txg & TXG_MASK] = B_FALSE; in redaction_list_update_sync() 569 if (!md->md_synctask_txg[txg & TXG_MASK]) { in commit_rl_updates() 572 md->md_synctask_txg[txg & TXG_MASK] = B_TRUE; in commit_rl_updates() 575 md->md_furthest[txg & TXG_MASK].rbp_object = object; in commit_rl_updates() 576 md->md_furthest[txg & TXG_MASK].rbp_blkid = blkid; in commit_rl_updates() 577 list_move_tail(&md->md_blocks[txg & TXG_MASK], in commit_rl_updates()
|
| H A D | vdev_trim.c | 208 uint64_t last_offset = vd->vdev_trim_offset[txg & TXG_MASK]; in vdev_trim_zap_update_sync() 209 vd->vdev_trim_offset[txg & TXG_MASK] = 0; in vdev_trim_zap_update_sync() 377 &vd->vdev_trim_offset[zio->io_txg & TXG_MASK]; in vdev_trim_cb() 512 vd->vdev_trim_offset[txg & TXG_MASK] == 0) { in vdev_trim_range() 540 vd->vdev_trim_offset[txg & TXG_MASK] = start + size; in vdev_trim_range() 550 zio_nowait(zio_trim(spa->spa_txg_zio[txg & TXG_MASK], vd, in vdev_trim_range()
|
| H A D | dsl_dir.c | 1158 ASSERT0(dd->dd_tempreserved[tx->tx_txg & TXG_MASK]); in dsl_dir_sync() 1160 dd->dd_space_towrite[tx->tx_txg & TXG_MASK] / 1024); in dsl_dir_sync() 1161 dd->dd_space_towrite[tx->tx_txg & TXG_MASK] = 0; in dsl_dir_sync() 1176 space += dd->dd_space_towrite[i & TXG_MASK]; in dsl_dir_space_towrite() 1177 ASSERT3U(dd->dd_space_towrite[i & TXG_MASK], >=, 0); in dsl_dir_space_towrite() 1355 dd->dd_tempreserved[txg & TXG_MASK] += asize; in dsl_dir_tempreserve_impl() 1452 int txgidx = tx->tx_txg & TXG_MASK; in dsl_dir_tempreserve_clear() 1497 dd->dd_space_towrite[tx->tx_txg & TXG_MASK] += space; in dsl_dir_willuse_space()
|
| H A D | zil.c | 642 if (txg_list_member(&dp->dp_dirty_zilogs, zilog, txg & TXG_MASK)) in zilog_is_dirty_in_txg() 868 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE; in zil_claim() 898 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE; in zil_claim() 1907 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_remove_async() 1951 itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_itx_assign() 2022 itxg_t *itxg = &zilog->zl_itxg[synced_txg & TXG_MASK]; in zil_clean() 2075 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_get_commit_list() 2120 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_async_to_sync() 3018 uint64_t *replayed_seq = &zilog->zl_replayed_seq[txg & TXG_MASK]; in zil_sync() 3634 zilog->zl_replayed_seq[dmu_tx_get_txg(tx) & TXG_MASK] = in zil_replaying()
|
| H A D | dnode_sync.c | 45 int txgoff = tx->tx_txg & TXG_MASK; in dnode_increase_indirection() 562 int txgoff = tx->tx_txg & TXG_MASK; in dnode_sync_free() 628 int txgoff = tx->tx_txg & TXG_MASK; in dnode_sync()
|
| H A D | metaslab.c | 1912 range_tree_space(msp->ms_allocating[(txg + t) & TXG_MASK]); in metaslab_verify_space() 3623 range_tree_walk(msp->ms_allocating[(txg + t) & TXG_MASK], in metaslab_condense() 3887 range_tree_t *alloctree = msp->ms_allocating[txg & TXG_MASK]; in metaslab_sync() 4153 ASSERT0(range_tree_space(msp->ms_allocating[txg & TXG_MASK])); in metaslab_sync() 4155 & TXG_MASK])); in metaslab_sync() 4182 msp->ms_allocating[(txg + t) & TXG_MASK])); in metaslab_evict() 4355 ASSERT0(range_tree_space(msp->ms_allocating[txg & TXG_MASK])); in metaslab_sync_done() 4584 if (range_tree_is_empty(msp->ms_allocating[txg & TXG_MASK])) in metaslab_block_alloc() 5557 range_tree_remove(msp->ms_allocating[txg & TXG_MASK], in metaslab_unalloc_dva() 5697 if (range_tree_is_empty(msp->ms_allocating[txg & TXG_MASK])) in metaslab_claim_concrete() [all …]
|
| H A D | dsl_dataset.c | 1261 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE; in dsl_dataset_zero_zil() 1446 !os->os_next_write_raw[tx->tx_txg & TXG_MASK]) { in dsl_dataset_dirty() 2080 if (ds->ds_resume_bytes[tx->tx_txg & TXG_MASK] != 0) { in dsl_dataset_sync() 2083 &ds->ds_resume_object[tx->tx_txg & TXG_MASK], tx)); in dsl_dataset_sync() 2086 &ds->ds_resume_offset[tx->tx_txg & TXG_MASK], tx)); in dsl_dataset_sync() 2089 &ds->ds_resume_bytes[tx->tx_txg & TXG_MASK], tx)); in dsl_dataset_sync() 2090 ds->ds_resume_object[tx->tx_txg & TXG_MASK] = 0; in dsl_dataset_sync() 2091 ds->ds_resume_offset[tx->tx_txg & TXG_MASK] = 0; in dsl_dataset_sync() 2092 ds->ds_resume_bytes[tx->tx_txg & TXG_MASK] = 0; in dsl_dataset_sync() 2274 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_FALSE; in dsl_dataset_sync_done() [all …]
|
| H A D | vdev_indirect.c | 553 &sci->sci_new_mapping_entries[txg & TXG_MASK], tx); in spa_condense_indirect_commit_sync() 554 ASSERT(list_is_empty(&sci->sci_new_mapping_entries[txg & TXG_MASK])); in spa_condense_indirect_commit_sync() 572 int txgoff = dmu_tx_get_txg(tx) & TXG_MASK; in spa_condense_indirect_commit_entry()
|
| H A D | dmu_object.c | 478 dn->dn_next_type[tx->tx_txg & TXG_MASK] = dn->dn_type = in dmu_object_zapify()
|
| H A D | dmu_objset.c | 1067 mdn->dn_next_nlevels[tx->tx_txg & TXG_MASK] = in dmu_objset_create_impl_dnstats() 1651 os->os_next_write_raw[tx->tx_txg & TXG_MASK]) { in dmu_objset_sync() 1685 txgoff = tx->tx_txg & TXG_MASK; in dmu_objset_sync() 1741 return (!multilist_is_empty(os->os_dirty_dnodes[txg & TXG_MASK])); in dmu_objset_is_dirty()
|
| H A D | dbuf.c | 2000 int txgoff = tx->tx_txg & TXG_MASK; in dbuf_dirty_lightweight() 2050 int txgoff = tx->tx_txg & TXG_MASK; in dbuf_dirty() 2258 dn->dn_next_nlevels[(tx->tx_txg-1) & TXG_MASK] > db->db_level || in dbuf_dirty() 2259 dn->dn_next_nlevels[(tx->tx_txg-2) & TXG_MASK] > db->db_level); in dbuf_dirty() 2402 list_remove(&dn->dn_dirty_records[txg & TXG_MASK], dr); in dbuf_undirty() 4316 list_insert_tail(&dn->dn_dirty_records[txg & TXG_MASK], dr); in dbuf_sync_leaf()
|
| H A D | spa_stats.c | 414 ts->ndirty = dp->dp_dirty_pertxg[txg & TXG_MASK]; in spa_txg_history_init_io()
|
| H A D | dmu.c | 867 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK]; in dmu_free_long_range_impl() 894 dp->dp_long_free_dirty_pertxg[txg & TXG_MASK] += in dmu_free_long_range_impl()
|
| H A D | spa.c | 8924 bplist_t *free_bpl = &spa->spa_free_bplist[txg & TXG_MASK]; 9076 (void) zio_wait(spa->spa_txg_zio[txg & TXG_MASK]); 9077 spa->spa_txg_zio[txg & TXG_MASK] = zio_root(spa, NULL, NULL, 9173 ASSERT0(spa->spa_vdev_removal->svr_bytes_done[txg & TXG_MASK]);
|
| H A D | dsl_crypt.c | 2104 os->os_next_write_raw[tx->tx_txg & TXG_MASK] = B_TRUE; in dsl_crypto_recv_raw_objset_sync()
|
| H A D | dmu_recv.c | 1293 int txgoff = dmu_tx_get_txg(tx) & TXG_MASK; in save_resume_state()
|
| H A D | zio.c | 1233 bplist_append(&spa->spa_free_bplist[txg & TXG_MASK], bp); in zio_free()
|
| /f-stack/freebsd/contrib/openzfs/include/sys/ |
| H A D | txg.h | 41 #define TXG_MASK (TXG_SIZE - 1) /* mask for size */ macro 43 #define TXG_IDX (txg & TXG_MASK)
|