/linux-master/samples/bpf/ |
H A D | run_cookie_uid_helper_example.sh | 8 iptables -D OUTPUT -m bpf --object-pinned ${mnt_dir}/bpf_prog -j ACCEPT
|
/linux-master/tools/testing/selftests/powerpc/pmu/ebb/ |
H A D | event_attributes_test.c | 117 leader.attr.pinned = 0; 118 /* Expected to fail, leader isn't pinned */
|
H A D | task_event_pinned_vs_ebb_test.c | 18 * Tests a pinned per-task event vs an EBB - in that order. The pinned per-task 26 event->attr.pinned = 1;
|
/linux-master/arch/arm64/include/asm/ |
H A D | mmu.h | 25 refcount_t pinned; member in struct:__anon10
|
H A D | mmu_context.h | 177 refcount_set(&mm->context.pinned, 0);
|
/linux-master/arch/arm64/kvm/ |
H A D | pmu-emul.c | 639 attr.pinned = 1;
|
/linux-master/arch/arm64/mm/ |
H A D | context.c | 175 * If it is pinned, we can keep using it. Note that reserved 176 * takes priority, because even if it is also pinned, we need to 179 if (refcount_read(&mm->context.pinned)) 285 if (refcount_inc_not_zero(&mm->context.pinned)) 304 refcount_set(&mm->context.pinned, 1); 329 if (refcount_dec_and_test(&mm->context.pinned)) { 394 * are pinned, there still is at least one empty slot in the ASID map.
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | iommu_api.c | 61 long i, ret, locked_entries = 0, pinned = 0; local 110 pinned += n; 114 pinned += ret; 118 if (pinned != entries) { 175 unpin_user_pages(mem->hpages, pinned);
|
/linux-master/arch/powerpc/perf/ |
H A D | core-book3s.c | 601 if (!leader->attr.pinned || !leader->attr.exclusive)
|
/linux-master/arch/riscv/kvm/ |
H A D | vcpu_pmu.c | 470 .pinned = true,
|
/linux-master/arch/x86/kernel/cpu/resctrl/ |
H A D | pseudo_lock.c | 938 .pinned = 1, 946 .pinned = 1,
|
/linux-master/arch/x86/kvm/ |
H A D | pmu.c | 185 .pinned = true,
|
/linux-master/arch/x86/kvm/vmx/ |
H A D | pmu_intel.c | 212 * - set 'pinned = true' to make it task pinned so that if another 213 * cpu pinned event reclaims LBR, the event->oncpu will be set to -1; 232 .pinned = true, 660 * Higher priority host perf events (e.g. cpu pinned) could reclaim the
|
/linux-master/arch/x86/xen/ |
H A D | mmu_pv.c | 177 * During early boot all page table pages are pinned, but we do not have struct 248 /* If page is not pinned, we can just update the entry 431 /* If page is not pinned, we can just update the entry 491 * 2. It is always pinned 514 /* If page is not pinned, we can just update the entry 526 /* If it's pinned, then we can at least batch the kernel and 710 * pinned, so Xen won't do the trap'n'emulate. 744 read-only, and can be pinned. */ 804 * The init_mm pagetable is really pinned as soon as its created, but 834 * partially-pinned stat 1540 bool pinned = xen_page_pinned(mm->pgd); local 1578 bool pinned = PagePinned(page); local [all...] |
/linux-master/drivers/fpga/ |
H A D | dfl-afu-dma-region.c | 29 * @region: dma memory region to be pinned 39 int ret, pinned; local 51 pinned = pin_user_pages_fast(region->user_addr, npages, FOLL_WRITE, 53 if (pinned < 0) { 54 ret = pinned; 56 } else if (pinned != npages) { 61 dev_dbg(dev, "%d pages pinned\n", pinned); 66 unpin_user_pages(region->pages, pinned);
|
/linux-master/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem.c | 604 * Our buffers are kept pinned, so allocating them from the MOVABLE 643 int ret, pinned = 0, npages = etnaviv_obj->base.size >> PAGE_SHIFT; local 659 unsigned num_pages = npages - pinned; 660 uint64_t ptr = userptr->ptr + pinned * PAGE_SIZE; 661 struct page **pages = pvec + pinned; 665 unpin_user_pages(pvec, pinned); 670 pinned += ret; 672 } while (pinned < npages);
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_pages.c | 117 * and pinned into our object. i915_gem_object_pin_pages() may be called 362 bool pinned; local 375 pinned = !(type & I915_MAP_OVERRIDE); 389 pinned = false; 423 if (pinned) { 504 * We allow removing the mapping from underneath pinned pages!
|
H A D | i915_gem_userptr.c | 239 int pinned, ret; local 267 pinned = 0; 268 while (pinned < num_pages) { 269 ret = pin_user_pages_fast(obj->userptr.ptr + pinned * PAGE_SIZE, 270 num_pages - pinned, gup_flags, 271 &pvec[pinned]); 275 pinned += ret; 303 unpin_user_pages(pvec, pinned); 452 * restriction is that we do not allow userptr surfaces to be pinned to the
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_gt_buffer_pool.c | 107 if (node->pinned) { 112 node->pinned = false; 129 if (node->pinned) 133 /* Hide this pinned object from the shrinker until retired */ 135 node->pinned = true; 153 node->pinned = false;
|
H A D | intel_gt_buffer_pool.h | 28 GEM_WARN_ON(!node->pinned);
|
H A D | intel_gt_buffer_pool_types.h | 33 u32 pinned; member in struct:intel_gt_buffer_pool_node
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem.c | 98 u64 pinned; local 103 pinned = ggtt->vm.reserved; 106 pinned += vma->node.size; 111 args->aper_available_size = args->aper_size - pinned; 866 * in use by hardware (i.e. they are pinned), we should not be powering 963 * half the size of aperture) or hasn't been pinned
|
/linux-master/drivers/gpu/drm/i915/selftests/ |
H A D | i915_vma.c | 113 int pinned; local 116 for (pinned = 0; pinned <= 1; pinned++) { 133 if (!pinned) {
|
/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_drv.c | 243 drm_gem_lru_init(&priv->lru.pinned, &priv->lru.lock);
|
H A D | msm_drv.h | 142 * pinned (for scannout or permanently mapped GPU buffers, like 143 * ringbuffer, memptr, fw, etc) it moves to the pinned LRU. When 163 * pinned: 165 * The LRU for pinned GEM objects 167 struct drm_gem_lru pinned; member in struct:msm_drm_private::__anon471
|