/freebsd-current/crypto/heimdal/base/ |
H A D | heimbase.c | 72 HEIM_TAILQ_HEAD(, heim_base) pool; 127 /* remove from autorel pool list */ 131 HEIM_TAILQ_REMOVE(&ar->pool, p, autorel); 444 heim_abort("autorelease pool released on thread w/o autorelease inited"); 448 if (!HEIM_TAILQ_EMPTY(&ar->pool)) 449 heim_abort("pool not empty after draining"); 453 heim_abort("autorelease not releaseing top pool"); 475 "autorelease-pool", 523 /* drop from old pool */ 526 HEIM_TAILQ_REMOVE(&ar->pool, [all...] |
/freebsd-current/sys/dev/ocs_fc/ |
H A D | ocs_utils.h | 124 extern void ocs_pool_reset(ocs_pool_t *pool); 125 extern void ocs_pool_free(ocs_pool_t *pool); 126 extern void *ocs_pool_get(ocs_pool_t *pool); 127 extern void ocs_pool_put(ocs_pool_t *pool, void *item); 128 extern uint32_t ocs_pool_get_count(ocs_pool_t *pool); 129 extern void *ocs_pool_get_instance(ocs_pool_t *pool, uint32_t idx); 130 extern uint32_t ocs_pool_get_freelist_count(ocs_pool_t *pool);
|
H A D | ocs_io.c | 57 * @brief IO pool. 59 * Structure encapsulating a pool of IO objects. 65 ocs_lock_t lock; /* IO pool lock */ 67 ocs_pool_t *pool; member in struct:ocs_io_pool_s 71 * @brief Create a pool of IO objects. 107 /* Allocate the IO pool */ 110 ocs_log_err(ocs, "allocate of IO pool failed\n"); 117 /* initialize IO pool lock */ 120 io_pool->pool = ocs_pool_alloc(ocs, sizeof(ocs_io_t), io_pool->io_num_ios, FALSE); 123 ocs_io_t *io = ocs_pool_get_instance(io_pool->pool, [all...] |
/freebsd-current/sys/dev/isci/scil/ |
H A D | sci_abstract_list.c | 89 * @pre The supplied free pool should be constructed prior to utilization 90 * of this abstract list. It isn't mandatory for the free pool to be 95 * @param[in] free_pool This parameter specifies the free pool to be 110 * Initialize the abstract list with its free pool 112 * @param[in] pool 113 * the free pool from which the elements will be extracted 123 SCI_ABSTRACT_ELEMENT_POOL_T * pool, 130 memset(pool, 0, sizeof(SCI_ABSTRACT_ELEMENT_POOL_T)); 133 pool->elements = list_elements; 134 pool 122 sci_abstract_element_pool_construct( SCI_ABSTRACT_ELEMENT_POOL_T * pool, SCI_ABSTRACT_ELEMENT_T * list_elements, int element_count ) argument [all...] |
H A D | scif_sas_internal_io_request.c | 168 if ( !sci_pool_full(fw_controller->hprq.pool) ) 171 fw_controller->hprq.pool, (POINTER_UINT) internal_io_memory
|
/freebsd-current/sys/vm/ |
H A D | vm_phys.c | 193 vm_phys_pool_valid(int pool) argument 196 if (pool == VM_FREEPOOL_LAZYINIT) 199 return (pool >= 0 && pool < VM_NFREEPOOL); 712 if (__predict_false(m->pool == VM_FREEPOOL_LAZYINIT)) { 799 * Set the pool for a contiguous, power of two-sized set of physical pages. 801 * If the pages currently belong to the lazy init pool, then the corresponding 806 vm_phys_set_pool(int pool, vm_page_t m, int order) argument 809 if (__predict_false(m->pool == VM_FREEPOOL_LAZYINIT)) { 813 m->pool 841 vm_phys_alloc_npages(int domain, int pool, int npages, vm_page_t ma[]) argument 850 ("vm_phys_alloc_npages: pool %d is out of range", pool)); local 913 vm_phys_alloc_pages(int domain, int pool, int order) argument 934 vm_phys_alloc_freelist_pages(int domain, int freelist, int pool, int order) argument 947 ("vm_phys_alloc_freelist_pages: pool %d is out of range", pool)); local [all...] |
/freebsd-current/sys/dev/mthca/ |
H A D | mthca_av.c | 189 ah->av = pci_pool_alloc(dev->av_table.pool, 250 pci_pool_free(dev->av_table.pool, ah->av, ah->avdma); 338 dev->av_table.pool = pci_pool_create("mthca_av", dev->pdev, 341 if (!dev->av_table.pool) 358 pci_pool_destroy(dev->av_table.pool); 372 pci_pool_destroy(dev->av_table.pool);
|
/freebsd-current/sys/contrib/openzfs/tests/zfs-tests/tests/functional/cli_user/misc/ |
H A D | setup.ksh | 128 # and create a pool we can perform attach remove replace, 139 # create an exported pool to test import 155 # copy a v1 pool from cli_root 156 cp $STF_SUITE/tests/functional/cli_root/zpool_upgrade/blockfiles/zfs-pool-v1.dat.bz2 \ 158 log_must bunzip2 $TEST_BASE_DIR/zfs-pool-v1.dat.bz2 159 log_must zpool import -d $TEST_BASE_DIR/ v1-pool
|
/freebsd-current/tests/sys/cddl/zfs/tests/cli_user/misc/ |
H A D | setup.ksh | 132 # and create a pool we can perform attach remove replace, 141 # create an exported pool to test import 145 # Now setup pool properties if they're supported 162 # copy a v1 pool from cli_root 163 $CP $STF_SUITE/tests/cli_root/zpool_upgrade/blockfiles/zfs-pool-v1.dat.Z \ 165 log_must $UNCOMPRESS /$TESTDIR/zfs-pool-v1.dat.Z 166 log_must $ZPOOL import -d /$TESTDIR v1-pool
|
/freebsd-current/contrib/ofed/opensm/include/complib/ |
H A D | cl_map.h | 112 cl_qpool_t pool; member in struct:_cl_map 119 * pool 275 * The map's pool of map items is the last thing initialized. 278 return (cl_is_qpool_inited(&p_map->pool));
|
/freebsd-current/sys/contrib/openzfs/tests/zfs-tests/tests/functional/replacement/ |
H A D | rebuild_multiple.ksh | 34 # 1. Create a mirrored pool. 47 pool=$1 51 count=$(zpool history -i $pool | grep "rebuild" | grep -c "$msg")
|
/freebsd-current/sys/contrib/openzfs/tests/zfs-tests/tests/functional/channel_program/ |
H A D | channel_common.kshlib | 136 pool=$1 152 echo "$args $pool $tmpin $lua_args"
|
/freebsd-current/sys/contrib/openzfs/tests/zfs-tests/tests/functional/cli_root/zpool_create/ |
H A D | zpool_create_011_neg.ksh | 38 # existent pool; device is part of an active pool; nested virtual devices; 44 # 2. For each scenario, try to create a new pool with the virtual devices 52 for pool in $TESTPOOL $TESTPOOL1 54 destroy_pool $pool 121 # now destroy the pool to be polite 125 # create/destroy a pool as a simple way to set the partitioning
|
/freebsd-current/sys/contrib/openzfs/tests/zfs-tests/tests/functional/history/ |
H A D | history_001_pos.ksh | 41 # 2. Run and verify pool commands, with special casing for destroy/export. 42 # 3. Import a pool and upgrade it, verifying 'upgrade' was logged. 80 # commands since the history will be unavailable until the pool is imported 97 log_fail "Didn't find \"$cmd1\" in pool history" 100 log_fail "Didn't find \"$cmd2\" in pool history" 104 log_fail "Didn't find \"$cmd1\" in pool history" 107 log_fail "Didn't find \"$cmd2\" in pool history" 115 log_must cp $STF_SUITE/tests/functional/history/zfs-pool-v4.dat.Z $import_dir 116 log_must uncompress $import_dir/zfs-pool-v4.dat.Z 117 upgrade_pool=$(zpool import -d $import_dir | awk '/pool [all...] |
/freebsd-current/sbin/ipf/ippool/ |
H A D | ippool.c | 254 ip_pool_t pool; local 262 bzero((char *)&pool, sizeof(pool)); 309 fprintf(stderr, "poolname not given with add/remove pool\n"); 329 strncpy(pool.ipo_name, poolname, sizeof(pool.ipo_name)); 330 pool.ipo_name[sizeof(pool.ipo_name) - 1] = '\0'; 331 pool.ipo_unit = role; 341 err = load_pool(&pool, ioct 945 ip_pool_t pool; local [all...] |
/freebsd-current/cddl/usr.sbin/zfsd/ |
H A D | vdev_iterator.cc | 62 VdevIterator::VdevIterator(zpool_handle_t *pool) argument 63 : m_poolConfig(zpool_get_config(pool, NULL)) 89 "ZPOOL_CONFIG_VDEV_TREE from pool.");
|
/freebsd-current/tests/sys/cddl/zfs/tests/cli_root/zpool_upgrade/ |
H A D | zpool_upgrade.kshlib | 31 # files make up the pool, and what the pool name is. 34 # A function to import a pool from files we have stored in the test suite 35 # We import the pool, and create some random data in the pool. 36 # $1 a version number we can use to get information about the pool 53 # Now put some random contents into the pool. 63 # A function to check the contents of a pool, upgrade it to the current version 65 # not using "zpool status -x" to see if the pool is healthy, as it's possible 67 # $1 a version number we can use to get information about the pool [all...] |
/freebsd-current/tests/sys/cddl/zfs/tests/history/ |
H A D | history_001_pos.ksh | 45 # 5. Imported specified pool and upgrade it, verify 'upgrade' was logged. 101 log_must $CP $STF_SUITE/tests/history/zfs-pool-v4.dat.Z $import_dir 102 log_must $UNCOMPRESS $import_dir/zfs-pool-v4.dat.Z 107 upgrade_pool=$($ZPOOL import -d $import_dir | $GREP "pool:" | $AWK '{print $2}')
|
/freebsd-current/contrib/llvm-project/clang/include/clang/Sema/ |
H A D | ParsedAttr.h | 615 /// individual attributes which are deallocated with the pool. 661 /// Reclaim all the attributes in the given pool chain, which is 690 "Can't take attribute from a pool that doesn't own it!"); 694 void takePool(AttributePool &pool); 697 /// Create a new pool for a factory. 707 /// Move the given pool's allocations to this pool. 708 AttributePool(AttributePool &&pool) = default; 712 AttributePool &operator=(AttributePool &&pool) = delete; 721 /// Take the given pool' [all...] |
/freebsd-current/sys/contrib/openzfs/module/os/freebsd/spl/ |
H A D | spl_kstat.c | 290 char *pool; local 350 * set of kstats belonging to a specific pool. Split on '/' to add an 351 * extra node for the pool name if needed. 355 pool = strchr(module, '/'); 356 if (pool != NULL) 357 *pool++ = '\0'; 362 * kstat.<module>[.<pool>].<class>.<name> 374 if (pool != NULL) { 376 SYSCTL_CHILDREN(root), OID_AUTO, pool, CTLFLAG_RW, 0, ""); 379 __func__, module, pool); [all...] |
/freebsd-current/contrib/libarchive/libarchive/ |
H A D | archive_read_support_format_warc.c | 114 /* string pool */ 115 warc_strbuf_t pool; member in struct:warc_s 171 if (w->pool.len > 0U) { 172 free(w->pool.str); 321 /* bang to our string pool, so we save a 323 if (fnam.len + 1U > w->pool.len) { 324 w->pool.len = ((fnam.len + 64U) / 64U) * 64U; 325 tmp = realloc(w->pool.str, w->pool.len); 332 w->pool [all...] |
/freebsd-current/sys/contrib/openzfs/contrib/zcp/ |
H A D | autosnap.lua | 3 -- Usage: zfs program <pool> autosnap.lua -- [-n] [-p <property>] <snapshot> 12 usage: zfs program <pool> autosnap.lua -- [-n] [-p <property>] <snapshot>
|
/freebsd-current/sys/contrib/openzfs/include/sys/ |
H A D | dmu_send.h | 59 int dmu_send_obj(const char *pool, uint64_t tosnap, uint64_t fromsnap,
|
/freebsd-current/sys/contrib/openzfs/tests/zfs-tests/tests/functional/cli_root/zfs_set/ |
H A D | ro_props_001_pos.ksh | 40 # 1. Create pool, fs, vol, fs@snap & vol@snap. 50 on off default pool/fs@snap $TESTDIR
|
/freebsd-current/sys/contrib/openzfs/tests/zfs-tests/tests/functional/rsend/ |
H A D | rsend_007_pos.ksh | 39 # 1. Separately rename pool clone, filesystem and volume name. 75 log_must eval "zfs send -R $POOL@final > $BACKDIR/pool-final-R" 76 log_must eval "zfs receive -d -F $POOL2 < $BACKDIR/pool-final-R" 84 "$BACKDIR/pool-init-final-IR" 91 log_must eval "zfs receive -d -F $dstds < $BACKDIR/pool-init-final-IR" 93 zfs receive -d -F $dstds < $BACKDIR/pool-init-final-IR
|