/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_bo_evict.c | 18 * Evict non-pinned user BOs first (via GPU), evict pinned external BOs next 19 * (via GPU), wait for evictions, and finally evict pinned kernel BOs via CPU. 54 spin_lock(&xe->pinned.lock); 56 bo = list_first_entry_or_null(&xe->pinned.external_vram, 62 spin_unlock(&xe->pinned.lock); 69 spin_lock(&xe->pinned.lock); 71 &xe->pinned.external_vram); 72 spin_unlock(&xe->pinned.lock); 76 spin_lock(&xe->pinned [all...] |
/linux-master/samples/bpf/ |
H A D | run_cookie_uid_helper_example.sh | 8 iptables -D OUTPUT -m bpf --object-pinned ${mnt_dir}/bpf_prog -j ACCEPT
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_gt_buffer_pool_types.h | 33 u32 pinned; member in struct:intel_gt_buffer_pool_node
|
H A D | intel_gt_buffer_pool.h | 28 GEM_WARN_ON(!node->pinned);
|
H A D | intel_gt_buffer_pool.c | 107 if (node->pinned) { 112 node->pinned = false; 129 if (node->pinned) 133 /* Hide this pinned object from the shrinker until retired */ 135 node->pinned = true; 153 node->pinned = false;
|
/linux-master/tools/testing/selftests/bpf/ |
H A D | test_xdp_veth.sh | 106 bpftool map update pinned $BPF_DIR/maps/tx_port key 0 0 0 0 value 122 0 0 0 107 bpftool map update pinned $BPF_DIR/maps/tx_port key 1 0 0 0 value 133 0 0 0 108 bpftool map update pinned $BPF_DIR/maps/tx_port key 2 0 0 0 value 111 0 0 0 109 ip link set dev veth1 xdp pinned $BPF_DIR/progs/xdp_redirect_map_0 110 ip link set dev veth2 xdp pinned $BPF_DIR/progs/xdp_redirect_map_1 111 ip link set dev veth3 xdp pinned $BPF_DIR/progs/xdp_redirect_map_2
|
H A D | test_flow_dissector.sh | 30 if ! unshare --net $bpftool prog attach pinned \ 36 $bpftool prog attach pinned /sys/fs/bpf/flow/_dissect \ 39 if unshare --net $bpftool prog attach pinned \ 45 if ! $bpftool prog detach pinned \
|
/linux-master/drivers/infiniband/core/ |
H A D | umem.c | 140 * @access: IB_ACCESS_xxx flags for memory being pinned 153 int pinned, ret; local 213 pinned = pin_user_pages_fast(cur_base, 218 if (pinned < 0) { 219 ret = pinned; 223 cur_base += pinned * PAGE_SIZE; 224 npages -= pinned; 226 &umem->sgt_append, page_list, pinned, 0, 227 pinned << PAGE_SHIFT, ib_dma_max_seg_size(device), 230 unpin_user_pages_dirty_lock(page_list, pinned, [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_userptr.c | 239 int pinned, ret; local 267 pinned = 0; 268 while (pinned < num_pages) { 269 ret = pin_user_pages_fast(obj->userptr.ptr + pinned * PAGE_SIZE, 270 num_pages - pinned, gup_flags, 271 &pvec[pinned]); 275 pinned += ret; 303 unpin_user_pages(pvec, pinned); 452 * restriction is that we do not allow userptr surfaces to be pinned to the
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | pin_system.c | 119 int pinned, cleared; local 138 pinned = hfi1_acquire_user_pages(current->mm, start_address, npages, 0, 141 if (pinned < 0) { 143 SDMA_DBG(req, "pinned %d", pinned); 144 return pinned; 146 if (pinned != npages) { 147 unpin_vector_pages(current->mm, pages, node->npages, pinned); 148 SDMA_DBG(req, "npages %u pinned %d", npages, pinned); [all...] |
H A D | user_exp_rcv.c | 122 * Release pinned receive buffer pages. 161 int pinned; local 179 * pages, accept the amount pinned so far and program only that. 187 pinned = hfi1_acquire_user_pages(current->mm, vaddr, npages, true, pages); 188 if (pinned <= 0) { 190 return pinned; 193 fd->tid_n_pinned += pinned; 194 return pinned; 249 int ret = 0, need_group = 0, pinned; local 288 pinned [all...] |
/linux-master/drivers/fpga/ |
H A D | dfl-afu-dma-region.c | 29 * @region: dma memory region to be pinned 39 int ret, pinned; local 51 pinned = pin_user_pages_fast(region->user_addr, npages, FOLL_WRITE, 53 if (pinned < 0) { 54 ret = pinned; 56 } else if (pinned != npages) { 61 dev_dbg(dev, "%d pages pinned\n", pinned); 66 unpin_user_pages(region->pages, pinned);
|
/linux-master/include/trace/events/ |
H A D | xen.h | 287 TP_PROTO(struct mm_struct *mm, unsigned long pfn, unsigned level, bool pinned), 288 TP_ARGS(mm, pfn, level, pinned), 293 __field(bool, pinned) 298 __entry->pinned = pinned), 301 __entry->pinned ? "" : "un") 305 TP_PROTO(unsigned long pfn, unsigned level, bool pinned), 306 TP_ARGS(pfn, level, pinned), 310 __field(bool, pinned) 314 __entry->pinned [all...] |
/linux-master/drivers/virt/acrn/ |
H A D | mm.c | 163 int ret, pinned; local 200 pinned = pin_user_pages_fast(memmap->vma_base, 203 if (pinned < 0) { 204 ret = pinned; 206 } else if (pinned != nr_pages) { 304 for (i = 0; i < pinned; i++)
|
/linux-master/tools/testing/selftests/powerpc/pmu/ebb/ |
H A D | task_event_pinned_vs_ebb_test.c | 18 * Tests a pinned per-task event vs an EBB - in that order. The pinned per-task 26 event->attr.pinned = 1;
|
H A D | cpu_event_pinned_vs_ebb_test.c | 18 * Tests a pinned cpu event vs an EBB - in that order. The pinned cpu event 26 event->attr.pinned = 1;
|
H A D | event_attributes_test.c | 117 leader.attr.pinned = 0; 118 /* Expected to fail, leader isn't pinned */
|
/linux-master/arch/arm64/mm/ |
H A D | context.c | 175 * If it is pinned, we can keep using it. Note that reserved 176 * takes priority, because even if it is also pinned, we need to 179 if (refcount_read(&mm->context.pinned)) 285 if (refcount_inc_not_zero(&mm->context.pinned)) 304 refcount_set(&mm->context.pinned, 1); 329 if (refcount_dec_and_test(&mm->context.pinned)) { 394 * are pinned, there still is at least one empty slot in the ASID map.
|
/linux-master/arch/arm64/include/asm/ |
H A D | mmu.h | 25 refcount_t pinned; member in struct:__anon10
|
/linux-master/drivers/gpu/drm/qxl/ |
H A D | qxl_object.h | 58 bool kernel, bool pinned, u32 domain,
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | iommu_api.c | 61 long i, ret, locked_entries = 0, pinned = 0; local 110 pinned += n; 114 pinned += ret; 118 if (pinned != entries) { 175 unpin_user_pages(mem->hpages, pinned);
|
/linux-master/drivers/s390/cio/ |
H A D | vfio_ccw_cp.c | 25 /* Array that receives the pinned pages. */ 27 /* Number of pages pinned from @pa_iova. */ 45 * @len: number of pages that should be pinned from @iova 88 * Only unpin if any pages were pinned to begin with, i.e. pa_nr > 0, 121 * Returns number of pages pinned upon success. 133 int pinned = 0, npage = 1; local 136 while (pinned < pa->pa_nr) { 137 dma_addr_t *first = &pa->pa_iova[pinned]; 140 if (pinned + npage < pa->pa_nr && 149 &pa->pa_page[pinned]); [all...] |
/linux-master/drivers/gpu/drm/etnaviv/ |
H A D | etnaviv_gem.c | 604 * Our buffers are kept pinned, so allocating them from the MOVABLE 643 int ret, pinned = 0, npages = etnaviv_obj->base.size >> PAGE_SHIFT; local 659 unsigned num_pages = npages - pinned; 660 uint64_t ptr = userptr->ptr + pinned * PAGE_SIZE; 661 struct page **pages = pvec + pinned; 665 unpin_user_pages(pvec, pinned); 670 pinned += ret; 672 } while (pinned < npages);
|
/linux-master/include/drm/ttm/ |
H A D | ttm_device.h | 255 * @pinned: Buffer objects which are pinned and so not on any LRU list. 257 struct list_head pinned; member in struct:ttm_device
|
/linux-master/kernel/ |
H A D | watchdog_perf.c | 89 .pinned = 1,
|