/linux-master/block/ |
H A D | blk-settings.c | 65 bdi->ra_pages = max(lim->io_opt * 2 / PAGE_SIZE, VM_READAHEAD_PAGES); 149 if (lim->max_user_sectors < PAGE_SIZE / SECTOR_SIZE) 181 if (WARN_ON_ONCE(lim->seg_boundary_mask < PAGE_SIZE - 1)) 199 if (WARN_ON_ONCE(lim->max_segment_size < PAGE_SIZE)) 211 if (WARN_ON_ONCE(lim->dma_alignment > PAGE_SIZE)) 322 if ((max_hw_sectors << 9) < PAGE_SIZE) { 481 if (max_size < PAGE_SIZE) { 482 max_size = PAGE_SIZE; 678 max(queue_io_opt(q) * 2 / PAGE_SIZE, VM_READAHEAD_PAGES); 719 if (sectors < PAGE_SIZE >> SECTOR_SHIF [all...] |
/linux-master/arch/arm64/include/asm/ |
H A D | tlbflush.h | 68 * PAGE_SIZE. Used by TTL. 79 switch (PAGE_SIZE) { 288 (uaddr & PAGE_MASK) + PAGE_SIZE); 483 __flush_tlb_range(vma, start, end, PAGE_SIZE, false, TLBI_TTL_UNKNOWN); 490 if ((end - start) > (MAX_DVM_OPS * PAGE_SIZE)) {
|
/linux-master/net/xdp/ |
H A D | xsk.c | 652 copy = min_t(u32, PAGE_SIZE - offset, len - copied); 728 skb_add_rx_frag(skb, nr_frags, page, 0, len, PAGE_SIZE); 729 refcount_add(PAGE_SIZE, &xs->sk.sk_wmem_alloc);
|
/linux-master/kernel/trace/ |
H A D | trace_events.c | 1731 if (cnt >= PAGE_SIZE) 1880 if (cnt >= PAGE_SIZE)
|
H A D | ring_buffer.c | 1698 buffer->subbuf_size = PAGE_SIZE - BUF_PAGE_HDR_SIZE; 2335 addr &= (PAGE_SIZE << cpu_buffer->buffer->subbuf_order) - 1; 2880 addr &= ~((PAGE_SIZE << cpu_buffer->buffer->subbuf_order) - 1); 3791 addr &= ~((PAGE_SIZE << cpu_buffer->buffer->subbuf_order) - 1); 5930 psize = (1 << order) * PAGE_SIZE;
|
/linux-master/kernel/ |
H A D | kprobes.c | 104 return PAGE_SIZE/(c->insn_size * sizeof(kprobe_opcode_t)); 121 return module_alloc(PAGE_SIZE); 198 PAGE_SIZE, false, c->sym); local 222 (unsigned long)kip->insns, PAGE_SIZE, true, local 304 addr < (unsigned long)kip->insns + PAGE_SIZE) {
|
/linux-master/include/linux/ |
H A D | hyperv.h | 37 * It doesn't make much difference between BUFFER and RING if PAGE_SIZE is the 40 * If PAGE_SIZE is bigger than HV_HYP_PAGE_SIZE, the headers of ring buffers 41 * will be of PAGE_SIZE, however, only the first HV_HYP_PAGE will be put 44 * if PAGE_SIZE is 64K: 157 /* Pad it to PAGE_SIZE so that data starts on page boundary */ 158 u8 reserved2[PAGE_SIZE - 68]; 1802 #define NR_HV_HYP_PAGES_IN_PAGE (PAGE_SIZE / HV_HYP_PAGE_SIZE)
|
/linux-master/fs/bcachefs/ |
H A D | util.h | 53 ((unsigned long) p & (PAGE_SIZE - 1)), 54 PAGE_SIZE);
|
/linux-master/drivers/uio/ |
H A D | uio_hv_generic.c | 146 .size = 2 * HV_RING_SIZE * PAGE_SIZE, 156 const size_t ring_bytes = HV_RING_SIZE * PAGE_SIZE; 254 ret = vmbus_alloc_ring(channel, HV_RING_SIZE * PAGE_SIZE, 255 HV_RING_SIZE * PAGE_SIZE); 282 pdata->info.mem[INT_PAGE_MAP].size = PAGE_SIZE; 288 pdata->info.mem[MON_PAGE_MAP].size = PAGE_SIZE;
|
/linux-master/drivers/target/ |
H A D | target_core_configfs.c | 545 return snprintf(page, PAGE_SIZE, "%u\n", to_attrib(item)->_name); \ 907 return snprintf(page, PAGE_SIZE, "0\n"); 1161 return snprintf(page, PAGE_SIZE, "%d\n", 1200 return snprintf(page, PAGE_SIZE, "%d\n", 1591 return snprintf(page, PAGE_SIZE, "%#08x\n", 1721 if (len + strlen(buf) >= PAGE_SIZE) 1750 if (len + strlen(buf) >= PAGE_SIZE) \ 1756 if (len + strlen(buf) >= PAGE_SIZE) \ 1762 if (len + strlen(buf) >= PAGE_SIZE) \ 1962 if (len + strlen(buf) >= PAGE_SIZE) [all...] |
/linux-master/drivers/scsi/ |
H A D | sg.c | 1643 "size (default: max(SG_SCATTER_SZ, PAGE_SIZE))"); 1682 if (scatter_elem_sz < PAGE_SIZE) { 1683 scatter_elem_sz = PAGE_SIZE; 1896 if (num < PAGE_SIZE) { 1897 scatter_elem_sz = PAGE_SIZE; 1898 scatter_elem_sz_prev = PAGE_SIZE; 2013 if (req_size < PAGE_SIZE) 2014 req_size = PAGE_SIZE; 2020 } while (req_size > (PAGE_SIZE / 2));
|
/linux-master/drivers/scsi/hisi_sas/ |
H A D | hisi_sas_v3_hw.c | 2791 return scnprintf(buf, PAGE_SIZE, "%u\n", hisi_sas_intr_conv); 2822 return scnprintf(buf, PAGE_SIZE, "%u\n", 2861 return scnprintf(buf, PAGE_SIZE, "%u\n", 2900 return scnprintf(buf, PAGE_SIZE, "%u\n",
|
H A D | hisi_sas_main.c | 2243 s = max(lcm(max_command_entries_ru, sz_slot_buf_ru), PAGE_SIZE);
|
/linux-master/drivers/net/ |
H A D | virtio_net.c | 649 if (unlikely(len > MAX_SKB_FRAGS * PAGE_SIZE)) { 654 BUG_ON(offset >= PAGE_SIZE); 656 unsigned int frag_size = min((unsigned)PAGE_SIZE - offset, len); 1135 if (page_off + *len + tailroom > PAGE_SIZE) 1160 if ((page_off + buflen + tailroom) > PAGE_SIZE) { 1347 page_to_skb(vi, rq, page, 0, len, PAGE_SIZE, 0); 1384 * XDP core assumes that xdp frags are PAGE_SIZE in length, while in 1498 if (unlikely(len > truesize - room || cur_frag_size > PAGE_SIZE)) { 1545 /* Now XDP core assumes frag size is PAGE_SIZE, but buffers 1547 * make their length exceed PAGE_SIZE [all...] |
/linux-master/drivers/net/hyperv/ |
H A D | netvsc.c | 347 buf_size = roundup(buf_size, PAGE_SIZE); 458 buf_size = round_up(buf_size, PAGE_SIZE);
|
/linux-master/drivers/net/ethernet/amd/pds_core/ |
H A D | main.c | 139 (u64)page_num << PAGE_SHIFT, PAGE_SIZE);
|
/linux-master/drivers/hv/ |
H A D | channel.c | 31 * For RING gpadl, in each ring, the guest uses one PAGE_SIZE as the header 34 * (PAGE_SIZE - HV_HYP_PAGE_SIZE) gap. And since there are two rings in a 45 BUG_ON(size % PAGE_SIZE); 54 return size - 2 * (PAGE_SIZE - HV_HYP_PAGE_SIZE); 72 * For RING gpadl, in each ring, the guest uses one PAGE_SIZE as the 75 * therefore leaving a (PAGE_SIZE - HV_HYP_PAGE_SIZE) gap. 80 return (offset - (PAGE_SIZE - HV_HYP_PAGE_SIZE)) >> HV_HYP_PAGE_SHIFT; 107 delta = PAGE_SIZE - HV_HYP_PAGE_SIZE; 109 delta = 2 * (PAGE_SIZE - HV_HYP_PAGE_SIZE); 173 if (send_size % PAGE_SIZE || recv_siz [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_migrate.c | 541 xe_res_next(cur, min_t(u32, size, PAGE_SIZE)); 771 xe_assert(xe, IS_ALIGNED(ccs_it.start, PAGE_SIZE));
|
/linux-master/drivers/gpu/drm/xe/regs/ |
H A D | xe_engine_regs.h | 54 #define RING_CTL_SIZE(size) ((size) - PAGE_SIZE) /* in bytes -> pages */ 55 #define RING_CTL_SIZE(size) ((size) - PAGE_SIZE) /* in bytes -> pages */
|
/linux-master/drivers/gpu/drm/qxl/ |
H A D | qxl_release.c | 39 #define RELEASES_PER_BO (PAGE_SIZE / RELEASE_SIZE) 42 #define SURFACE_RELEASES_PER_BO (PAGE_SIZE / SURFACE_RELEASE_SIZE) 207 return qxl_bo_create(qdev, PAGE_SIZE, false, true,
|
/linux-master/drivers/gpu/drm/panfrost/ |
H A D | panfrost_mmu.c | 439 #define NUM_FAULT_PAGES (SZ_2M / PAGE_SIZE) 520 sgt = &bo->sgts[page_offset / (SZ_2M / PAGE_SIZE)];
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 453 return ce->parallel.guc.parent_page * PAGE_SIZE; 4311 memset(engine->status_page.addr, POISON_INUSE, PAGE_SIZE); 4321 drm_clflush_virt_range(engine->status_page.addr, PAGE_SIZE);
|
/linux-master/drivers/gpu/drm/amd/pm/swsmu/smu14/ |
H A D | smu_v14_0_0_ppt.c | 163 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM); 165 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM); 167 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
|
H A D | smu_v14_0.c | 90 ALIGN(le32_to_cpu(header->ucode_size_bytes), PAGE_SIZE); 191 ALIGN(smu->pptable_firmware.size, PAGE_SIZE);
|
/linux-master/drivers/gpu/drm/amd/pm/swsmu/smu13/ |
H A D | smu_v13_0_4_ppt.c | 156 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM); 158 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM); 160 PAGE_SIZE, AMDGPU_GEM_DOMAIN_VRAM);
|