/linux-master/tools/perf/scripts/perl/bin/ |
H A D | check-perf-trace-record | 2 perf record -a -e kmem:kmalloc -e irq:softirq_entry -e kmem:kfree
|
/linux-master/mm/kasan/ |
H A D | kasan_test_module.c | 20 char *kmem; local 25 kmem = kmalloc(size, GFP_KERNEL); 26 if (!kmem) 34 kfree(kmem); 41 unused = copy_from_user(kmem, usermem, size + 1); 44 unused = copy_to_user(usermem, kmem, size + 1); 47 unused = __copy_from_user(kmem, usermem, size + 1); 50 unused = __copy_to_user(usermem, kmem, size + 1); 53 unused = __copy_from_user_inatomic(kmem, usermem, size + 1); 56 unused = __copy_to_user_inatomic(usermem, kmem, siz [all...] |
/linux-master/lib/ |
H A D | test_user_copy.c | 47 static int test_check_nonzero_user(char *kmem, char __user *umem, size_t size) argument 64 kmem += start; 81 memset(kmem, 0x0, size); 83 kmem[i] = 0xff; 85 kmem[i] = 0xff; 87 ret |= test(copy_to_user(umem, kmem, size), 94 int expected = is_zeroed(kmem + start, len); 105 static int test_copy_struct_from_user(char *kmem, char __user *umem, argument 133 memset(kmem, 0x0, size); 134 ret |= test(copy_struct_from_user(kmem, ksiz 183 char *kmem; local [all...] |
/linux-master/tools/testing/selftests/ftrace/test.d/filter/ |
H A D | event-filter-function.tc | 4 # requires: set_event events/kmem/kmem_cache_free/filter 14 echo 1 > events/kmem/kmem_cache_free/enable 18 echo 0 > events/kmem/kmem_cache_free/enable 34 echo "call_site.function == $target_func" > events/kmem/kmem_cache_free/filter 51 echo "call_site.function == 0x$address" > events/kmem/kmem_cache_free/filter
|
/linux-master/tools/testing/selftests/ftrace/test.d/trigger/ |
H A D | trigger-hist.tc | 44 echo 'hist:keys=call_site.sym' > events/kmem/kmalloc/trigger 46 grep '{ call_site: \[[0-9a-f][0-9a-f]*\] [_a-zA-Z][_a-zA-Z]* *}' events/kmem/kmalloc/hist > /dev/null || \ 53 echo 'hist:keys=call_site.sym-offset' > events/kmem/kmalloc/trigger 55 grep '{ call_site: \[[0-9a-f][0-9a-f]*\] [_a-zA-Z][_a-zA-Z]*+0x[0-9a-f][0-9a-f]*' events/kmem/kmalloc/hist > /dev/null || \
|
H A D | trigger-hist-mod.tc | 45 echo 'hist:keys=bytes_req.log2' > events/kmem/kmalloc/trigger 47 grep 'bytes_req: ~ 2^[0-9]*' events/kmem/kmalloc/hist > /dev/null || \ 48 fail "log2 modifier on kmem/kmalloc did not work"
|
/linux-master/drivers/dax/ |
H A D | Makefile | 4 obj-$(CONFIG_DEV_DAX_KMEM) += kmem.o
|
/linux-master/tools/perf/scripts/perl/ |
H A D | check-perf-trace.pl | 43 sub kmem::kmalloc subroutine 59 flag_str("kmem::kmalloc", "gfp_flags", $gfp_flags));
|
/linux-master/arch/x86/kernel/cpu/resctrl/ |
H A D | pseudo_lock.c | 265 kfree(plr->kmem); 266 plr->kmem = NULL; 380 plr->kmem = kzalloc(plr->size, GFP_KERNEL); 381 if (!plr->kmem) { 485 mem_r = plr->kmem; 904 mem_r = READ_ONCE(plr->kmem); 1009 mem_r = READ_ONCE(plr->kmem); 1536 physical = __pa(plr->kmem) >> PAGE_SHIFT; 1558 memset(plr->kmem + off, 0, vsize);
|
H A D | internal.h | 226 * @kmem: the kernel memory associated with pseudo-locked region 242 void *kmem; member in struct:pseudo_lock_region
|
/linux-master/drivers/infiniband/hw/hns/ |
H A D | hns_roce_mr.c | 678 if (mtr->kmem) { 679 hns_roce_buf_free(hr_dev, mtr->kmem); 680 mtr->kmem = NULL; 694 mtr->kmem = NULL; 704 mtr->kmem = hns_roce_buf_alloc(hr_dev, total_size, 708 if (IS_ERR(mtr->kmem)) { 709 ibdev_err(ibdev, "failed to alloc kmem, ret = %ld.\n", 710 PTR_ERR(mtr->kmem)); 711 return PTR_ERR(mtr->kmem); 763 mtr->kmem, page_shif [all...] |
H A D | hns_roce_device.h | 306 struct hns_roce_buf *kmem; /* kernel space buffer */ member in struct:hns_roce_mtr
|
H A D | hns_roce_hw_v2.c | 868 return hns_roce_buf_offset(srq->buf_mtr.kmem, n << srq->wqe_shift); 873 return hns_roce_buf_offset(idx_que->mtr.kmem, 3470 return hns_roce_buf_offset(hr_cq->mtr.kmem, n * hr_cq->cqe_size); 5994 aeqe = hns_roce_buf_offset(eq->mtr.kmem, 6079 ceqe = hns_roce_buf_offset(eq->mtr.kmem,
|
H A D | hns_roce_qp.c | 1490 return hns_roce_buf_offset(hr_qp->mtr.kmem, offset);
|
/linux-master/drivers/crypto/nx/ |
H A D | nx.h | 119 void *kmem; /* unaligned, kmalloc'd buffer */ member in struct:nx_crypto_ctx 120 size_t kmem_len; /* length of kmem */ 126 struct nx_sg *in_sg; /* aligned pointer into kmem to an sg list */ 127 struct nx_sg *out_sg; /* aligned pointer into kmem to an sg list */
|
H A D | nx.c | 309 memset(nx_ctx->kmem, 0, nx_ctx->kmem_len); 670 nx_ctx->kmem = kmalloc(nx_ctx->kmem_len, GFP_KERNEL); 671 if (!nx_ctx->kmem) 675 nx_ctx->csbcpb = (struct nx_csbcpb *)(round_up((u64)nx_ctx->kmem, 750 kfree_sensitive(nx_ctx->kmem); 766 kfree_sensitive(nx_ctx->kmem);
|
/linux-master/tools/perf/ |
H A D | perf-completion.sh | 211 if [[ $prev_skip_opts == @(kvm|kmem|mem|lock|sched|
|
/linux-master/drivers/infiniband/hw/irdma/ |
H A D | verbs.h | 126 struct irdma_dma_mem kmem; member in struct:irdma_cq
|
H A D | verbs.c | 1789 dma_free_coherent(rf->sc_dev.hw->device, iwcq->kmem.size, 1790 iwcq->kmem.va, iwcq->kmem.pa); 1791 iwcq->kmem.va = NULL; 2001 cq_buf->kmem_buf = iwcq->kmem; 2006 iwcq->kmem = kmem_buf; 2163 iwcq->kmem.size = ALIGN(round_up(rsize, 256), 256); 2164 iwcq->kmem.va = dma_alloc_coherent(dev->hw->device, 2165 iwcq->kmem.size, 2166 &iwcq->kmem [all...] |
/linux-master/include/trace/events/ |
H A D | kmem.h | 3 #define TRACE_SYSTEM kmem
|
/linux-master/mm/ |
H A D | memcontrol.c | 1755 pr_info("kmem: usage %llukB, limit %llukB, failcnt %lu\n", 1756 K((u64)page_counter_read(&memcg->kmem)), 1757 K((u64)memcg->kmem.max), memcg->kmem.failcnt); 3261 page_counter_charge(&memcg->kmem, nr_pages); 3263 page_counter_uncharge(&memcg->kmem, -nr_pages); 3314 * __memcg_kmem_charge_page: charge a kmem page to the current memory cgroup 3340 * __memcg_kmem_uncharge_page: uncharge a kmem page 3907 counter = &memcg->kmem; 4072 pr_warn_once("kmem [all...] |
H A D | slab_common.c | 36 #include <trace/events/kmem.h>
|
H A D | page_alloc.c | 41 #include <trace/events/kmem.h>
|
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt_ethtool.c | 3143 u8 *kmem; local 3145 kmem = hwrm_req_dma_slice(bp, req, data_len, &dma_handle); 3146 if (!kmem) { 3153 memcpy(kmem, data, data_len); 3548 u8 *kmem = NULL; local 3572 kmem = hwrm_req_dma_slice(bp, modify, modify_len, &dma_handle); 3573 if (!kmem && modify_len > PAGE_SIZE) 3578 if (!kmem) { 3629 memcpy(kmem, fw->data + copied, len);
|
/linux-master/include/linux/ |
H A D | memcontrol.h | 215 struct page_counter kmem; /* v1 only */ member in struct:mem_cgroup 377 * __folio_memcg - Get the memory cgroup associated with a non-kmem folio 384 * kmem folios. 398 * __folio_objcg - get the object cgroup associated with a kmem folio. 427 * For a non-kmem folio any of the following ensures folio and memcg binding 436 * For a kmem folio a caller should hold an rcu read lock to protect memcg 437 * associated with a kmem folio from being released. 489 * For a non-kmem folio any of the following ensures folio and memcg binding 498 * For a kmem folio a caller should hold an rcu read lock to protect memcg 499 * associated with a kmem foli [all...] |