/freebsd-10-stable/sys/ofed/include/net/ |
H A D | tcp.h | 34 #include <linux/slab.h>
|
/freebsd-10-stable/sys/vm/ |
H A D | uma_dbg.c | 199 uma_slab_t slab; local 205 slab = vtoslab((vm_offset_t)mem); 208 * It is safe to return the slab here even though the 215 slab = hash_sfind(&keg->uk_hash, mem); 217 slab = (uma_slab_t)(mem + keg->uk_pgoff); 221 return (slab); 225 * Set up the slab's freei data such that uma_dbg_free can function. 229 uma_dbg_alloc(uma_zone_t zone, uma_slab_t slab, void *item) argument 236 if (slab == NULL) { 237 slab 259 uma_dbg_free(uma_zone_t zone, uma_slab_t slab, void *item) argument [all...] |
H A D | uma_dbg.h | 52 void uma_dbg_free(uma_zone_t zone, uma_slab_t slab, void *item); 53 void uma_dbg_alloc(uma_zone_t zone, uma_slab_t slab, void *item);
|
H A D | uma_core.c | 39 * The basic ideas stem from similar slab/zone based allocators whose algorithms 265 static void *slab_alloc_item(uma_keg_t keg, uma_slab_t slab); 266 static void slab_free_item(uma_keg_t keg, uma_slab_t slab, void *item); 576 uma_slab_t slab; local 593 slab = SLIST_FIRST(&oldhash->uh_slab_hash[i]); 595 hval = UMA_HASH(newhash, slab->us_data); 597 slab, us_hlink); 809 keg_free_slab(uma_keg_t keg, uma_slab_t slab, int start) argument 815 mem = slab->us_data; 816 flags = slab 842 uma_slab_t slab; local 943 uma_slab_t slab; local 1786 uma_slab_t slab; local 2329 uma_slab_t slab; local 2403 uma_slab_t slab; local 2431 uma_slab_t slab; local 2502 slab_alloc_item(uma_keg_t keg, uma_slab_t slab) argument 2528 uma_slab_t slab; local 2829 slab_free_item(uma_keg_t keg, uma_slab_t slab, void *item) argument 2858 uma_slab_t slab; local 3156 uma_slab_t slab; local 3182 uma_slab_t slab; local 3281 uma_slab_t slab; local 3301 uma_large_free(uma_slab_t slab) argument 3326 slab_print(uma_slab_t slab) argument 3345 uma_slab_t slab; local [all...] |
H A D | uma_int.h | 47 * be allocated off the page from a special slab zone. The free list within a 48 * slab is managed with a bitmask. For item sizes that would yield more than 50 * improve the number of items per slab that will fit. 58 * course, be solved with dynamic slab sizes. 77 * This is the representation for normal (Non OFFPAGE slab) 80 * s == slab pointer 85 * ||i||i||i||i||i||i||i||i||i||i||i||i||i||i||i| |slab header|| 90 * This is an OFFPAGE slab. These can be larger than UMA_SLAB_SIZE. 98 * |slab header| | 112 /* Max waste percentage before going to off page slab managemen 391 uma_slab_t slab; local 413 vsetslab(vm_offset_t va, uma_slab_t slab) argument [all...] |
/freebsd-10-stable/contrib/unbound/validator/ |
H A D | val_kcache.c | 62 kcache->slab = slabhash_create(numtables, start_size, maxmem, 65 if(!kcache->slab) { 78 slabhash_delete(kcache->slab); 95 slabhash_insert(kcache->slab, k->entry.hash, &k->entry, 121 e = slabhash_lookup(kcache->slab, lookfor.entry.hash, &lookfor, wr); 159 return sizeof(*kcache) + slabhash_get_mem(kcache->slab); 171 slabhash_remove(kcache->slab, lookfor.entry.hash, &lookfor);
|
H A D | val_kcache.h | 56 struct slabhash* slab; member in struct:key_cache
|
/freebsd-10-stable/sys/ofed/include/linux/ |
H A D | string.h | 35 #include <linux/slab.h>
|
H A D | dmapool.h | 37 #include <linux/slab.h>
|
H A D | module.h | 36 #include <linux/slab.h>
|
H A D | kthread.h | 39 #include <linux/slab.h>
|
H A D | kobject.h | 36 #include <linux/slab.h>
|
/freebsd-10-stable/contrib/unbound/util/storage/ |
H A D | slabhash.c | 223 size_t slab, cnt = 0; local 225 for(slab=0; slab<sh->size; slab++) { 226 lock_quick_lock(&sh->array[slab]->lock); 227 cnt += sh->array[slab]->num; 228 lock_quick_unlock(&sh->array[slab]->lock);
|
/freebsd-10-stable/sys/kern/ |
H A D | kern_malloc.c | 553 uma_slab_t slab; local 574 slab = vtoslab((vm_offset_t)addr & (~UMA_SLAB_MASK)); 576 if (slab == NULL) 580 if (!(slab->us_flags & UMA_SLAB_MALLOC)) { 584 size = slab->us_keg->uk_size; 600 uma_zfree_arg(LIST_FIRST(&slab->us_keg->uk_zones), addr, slab); 602 size = slab->us_size; 603 uma_large_free(slab); 614 uma_slab_t slab; local 853 uma_slab_t slab; local [all...] |
/freebsd-10-stable/sys/ofed/drivers/infiniband/hw/mlx4/ |
H A D | doorbell.c | 33 #include <linux/slab.h>
|
/freebsd-10-stable/contrib/unbound/daemon/ |
H A D | cachedump.c | 123 size_t slab; local 125 for(slab=0; slab<r->table.size; slab++) { 126 lock_quick_lock(&r->table.array[slab]->lock); 127 if(!dump_rrset_lruhash(ssl, r->table.array[slab], 129 lock_quick_unlock(&r->table.array[slab]->lock); 132 lock_quick_unlock(&r->table.array[slab]->lock); 279 size_t slab; local 281 for(slab [all...] |
/freebsd-10-stable/contrib/netbsd-tests/sys/uvm/ |
H A D | t_uvm_physseg.c | 474 struct vm_page *pgs, *slab = malloc(sizeof(struct vm_page) * (npages1 local 495 uvm_page_init_fake(slab, npages1 + npages2 + npages3); 503 /* Scavenge plug - goes into the same slab */ 512 /* Scavenge plug should fit right in the slab */ 514 ATF_REQUIRE(pgs > slab && pgs < (slab + npages1 + npages2 + npages3)); 516 /* Hot plug - goes into a brand new slab */ 518 /* The hot plug slab should have nothing to do with the original slab */ 520 ATF_REQUIRE(pgs < slab || pg 541 struct vm_page *slab = malloc(sizeof(struct vm_page) * (npages1 + npages2 + npages3)); local 681 struct vm_page *slab = malloc(sizeof(struct vm_page) * (npages1 + npages2)); local 775 struct vm_page *slab, *pgs; local 785 uvm_physseg_seg_chomp_slab(PHYSSEG_NODE_TO_HANDLE(seg), slab, npages * 2); local 810 uvm_physseg_seg_chomp_slab(PHYSSEG_NODE_TO_HANDLE(seg), slab, npages * 2); local 830 struct vm_page *slab, *pgs; local 840 uvm_physseg_seg_chomp_slab(PHYSSEG_NODE_TO_HANDLE(seg), slab, npages * 2); local 868 struct vm_page *slab, *pgs; local 878 uvm_physseg_seg_chomp_slab(PHYSSEG_NODE_TO_HANDLE(seg), slab, npages * 2); local [all...] |
H A D | t_uvm_physseg_load.c | 543 struct vm_page *slab = malloc(sizeof(struct vm_page) * local 553 uvm_page_init_fake(slab, npages1 + npages2); 592 struct vm_page *slab = malloc(sizeof(struct vm_page) * local 602 uvm_page_init_fake(slab, npages1 + npages2); 641 struct vm_page *slab = malloc(sizeof(struct vm_page) local 651 uvm_page_init_fake(slab, npages1 + npages2); 690 struct vm_page *slab = malloc(sizeof(struct vm_page) * (npages1 + npages2)); local 699 uvm_page_init_fake(slab, npages1 + npages2);
|
/freebsd-10-stable/sys/contrib/xz-embedded/linux/lib/xz/ |
H A D | xz_private.h | 19 # include <linux/slab.h>
|
/freebsd-10-stable/sys/dev/mlx5/mlx5_ib/ |
H A D | mlx5_ib_doorbell.c | 30 #include <linux/slab.h>
|
/freebsd-10-stable/sys/ofed/drivers/infiniband/debug/ |
H A D | mtrack.h | 6 #include <linux/slab.h>
|
/freebsd-10-stable/sys/ofed/drivers/net/mlx4/ |
H A D | en_resources.c | 34 #include <linux/slab.h>
|
H A D | reset.c | 37 #include <linux/slab.h>
|
/freebsd-10-stable/sys/contrib/xz-embedded/linux/lib/ |
H A D | decompress_unxz.c | 112 # include <linux/slab.h>
|
/freebsd-10-stable/sys/ofed/drivers/infiniband/core/ |
H A D | agent.c | 39 #include <linux/slab.h>
|