/linux-master/fs/erofs/ |
H A D | super.c | 327 if (sbi->sb_size > PAGE_SIZE - EROFS_SUPER_OFFSET) { 594 sb->s_blocksize = PAGE_SIZE; 605 if (!sb_set_blocksize(sb, PAGE_SIZE)) {
|
H A D | fscache.c | 59 pgoff_t start_page = req->start / PAGE_SIZE; 60 pgoff_t last_page = ((req->start + req->len) / PAGE_SIZE) - 1; 289 iov_iter_xarray(&iter, ITER_DEST, &mapping->i_pages, pos, PAGE_SIZE); 294 iov_iter_zero(PAGE_SIZE - size, &iter); 296 req->submitted += PAGE_SIZE; 311 DBG_BUGON(!count || count % PAGE_SIZE);
|
/linux-master/drivers/irqchip/ |
H A D | irq-gic-v3-its.c | 142 #define PAGE_ORDER_TO_SIZE(o) (PAGE_SIZE << (o)) 2939 np = DIV_ROUND_UP(npg * psz, PAGE_SIZE); 2943 page = alloc_pages(GFP_ATOMIC | __GFP_ZERO, get_order(np * PAGE_SIZE));
|
/linux-master/drivers/i2c/ |
H A D | i2c-core-base.c | 675 len = of_device_modalias(dev, buf, PAGE_SIZE); 679 len = acpi_device_modalias(dev, buf, PAGE_SIZE - 1);
|
/linux-master/drivers/dma/xilinx/ |
H A D | xdma-regs.h | 38 #define XDMA_DESC_BLEN_MAX (BIT(XDMA_DESC_BLEN_BITS) - PAGE_SIZE)
|
/linux-master/drivers/dma/idxd/ |
H A D | idxd.h | 617 #define IDXD_PORTAL_MASK (PAGE_SIZE - 1)
|
H A D | cdev.c | 381 if ((vma->vm_end - vma->vm_start) > PAGE_SIZE) { 413 return io_remap_pfn_range(vma, vma->vm_start, pfn, PAGE_SIZE,
|
/linux-master/arch/x86/kernel/ |
H A D | sev-shared.c | 1059 if (!cc_info || !cc_info->cpuid_phys || cc_info->cpuid_len < PAGE_SIZE) 1102 for (; vaddr < vaddr_end; vaddr += PAGE_SIZE) {
|
/linux-master/arch/riscv/mm/ |
H A D | init.c | 61 unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)] 289 pgd_t early_pg_dir[PTRS_PER_PGD] __initdata __aligned(PAGE_SIZE); 361 return memblock_phys_alloc(PAGE_SIZE, PAGE_SIZE); 378 BUG_ON(sz != PAGE_SIZE); 388 static pmd_t early_pmd[PTRS_PER_PMD] __initdata __aligned(PAGE_SIZE); 398 static p4d_t early_p4d[PTRS_PER_P4D] __initdata __aligned(PAGE_SIZE); 408 static pud_t early_pud[PTRS_PER_PUD] __initdata __aligned(PAGE_SIZE); 442 return memblock_phys_alloc(PAGE_SIZE, PAGE_SIZE); [all...] |
/linux-master/arch/riscv/include/asm/ |
H A D | pgtable.h | 29 #define PTRS_PER_PGD (PAGE_SIZE / sizeof(pgd_t)) 31 #define PTRS_PER_PTE (PAGE_SIZE / sizeof(pte_t)) 439 PAGE_SIZE) 484 local_flush_tlb_page(address + nr * PAGE_SIZE); 882 #define TASK_SIZE_32 (_AC(0x80000000, UL) - PAGE_SIZE) 924 extern unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)];
|
H A D | page.h | 16 #define PAGE_SIZE (_AC(1, UL) << PAGE_SHIFT) macro 17 #define PAGE_MASK (~(PAGE_SIZE - 1)) 50 #define clear_page(pgaddr) memset((pgaddr), 0, PAGE_SIZE) 52 #define copy_page(to, from) memcpy((to), (from), PAGE_SIZE) 56 memcpy((vto), (vfrom), PAGE_SIZE)
|
/linux-master/tools/testing/selftests/mm/ |
H A D | protection_keys.c | 154 __attribute__((__aligned__(PAGE_SIZE))) 707 ret = mprotect_pkey((void *)ptr, PAGE_SIZE, prot, pkey); 1129 iov.iov_len = PAGE_SIZE; 1173 err = sys_mprotect_pkey(ptr, PAGE_SIZE, PROT_READ, i); 1185 err = sys_mprotect_pkey(ptr, PAGE_SIZE, PROT_READ, bad_pkey); 1385 int *plain_ptr = ALIGN_PTR_UP(plain_ptr_unaligned, PAGE_SIZE); 1453 p1 = ALIGN_PTR_UP(&lots_o_noops_around_write, PAGE_SIZE); 1459 p1 += PAGE_SIZE; 1465 madvise(p1, PAGE_SIZE, MADV_DONTNEED); 1482 ret = mprotect_pkey(p1, PAGE_SIZE, PROT_EXE [all...] |
/linux-master/net/tls/ |
H A D | tls_strp.c | 217 frag = &skb_shinfo(skb)->frags[skb->len / PAGE_SIZE]; 223 chunk = min_t(size_t, len, PAGE_SIZE - skb_frag_size(frag)); 261 chunk = min_t(size_t, chunk, PAGE_SIZE - skb_frag_size(frag)); 291 chunk = TLS_MAX_PAYLOAD_SIZE + PAGE_SIZE; 401 need_spc = strp->stm.full_len ?: TLS_MAX_PAYLOAD_SIZE + PAGE_SIZE; 403 for (len = need_spc; len > 0; len -= PAGE_SIZE) { 419 strp->anchor->truesize = round_up(need_spc, PAGE_SIZE);
|
/linux-master/net/bluetooth/ |
H A D | hci_conn.c | 2606 if (!req.conn_num || req.conn_num > (PAGE_SIZE * 2) / sizeof(*ci))
|
/linux-master/mm/ |
H A D | zswap.c | 271 DIV_ROUND_UP(zswap_pool_total_size, PAGE_SIZE); 278 DIV_ROUND_UP(zswap_pool_total_size, PAGE_SIZE); 943 acomp_ctx->buffer = kmalloc_node(PAGE_SIZE * 2, GFP_KERNEL, cpu_to_node(cpu)); 1005 unsigned int dlen = PAGE_SIZE; 1018 sg_set_page(&input, &folio->page, PAGE_SIZE, 0); 1021 * We need PAGE_SIZE * 2 here since there maybe over-compression case, 1025 sg_init_one(&output, dst, PAGE_SIZE * 2); 1026 acomp_request_set_params(acomp_ctx->req, &input, &output, PAGE_SIZE, dlen); 1101 sg_set_page(&output, page, PAGE_SIZE, 0); 1102 acomp_request_set_params(acomp_ctx->req, &input, &output, entry->length, PAGE_SIZE); [all...] |
H A D | hugetlb.c | 1010 return PAGE_SIZE; 3325 memblock_reserved_mark_noinit(virt_to_phys((void *)m + PAGE_SIZE), 3326 huge_page_size(h) - PAGE_SIZE); 3944 target_hstate = size_to_hstate(PAGE_SIZE << h->demote_order); 4297 unsigned long demote_size = (PAGE_SIZE << h->demote_order) / SZ_1K; 4655 if (size_to_hstate(PAGE_SIZE << order)) { 5112 /* Return the number pages of memory we physically have, in PAGE_SIZE units. */ 5879 * vm_pgoff is in PAGE_SIZE units, hence the different calculation 7816 if (hugetlb_cma_size_in_node[nid] < (PAGE_SIZE << order)) { 7818 nid, (PAGE_SIZE << orde [all...] |
/linux-master/kernel/ |
H A D | vmcore_info.c | 140 VMCOREINFO_PAGESIZE(PAGE_SIZE);
|
/linux-master/include/net/ |
H A D | sock.h | 1566 return (amt + PAGE_SIZE - 1) >> PAGE_SHIFT; 1619 if (reclaimable >= (int)PAGE_SIZE)
|
/linux-master/include/linux/ |
H A D | page-flags.h | 213 * Only addresses aligned with PAGE_SIZE of struct page may be fake head 218 if (IS_ALIGNED((unsigned long)page, PAGE_SIZE) &&
|
H A D | mm.h | 222 #define PAGE_ALIGN(addr) ALIGN(addr, PAGE_SIZE) 225 #define PAGE_ALIGN_DOWN(addr) ALIGN_DOWN(addr, PAGE_SIZE) 227 /* test whether an address (unsigned long or pointer) is aligned to PAGE_SIZE */ 228 #define PAGE_ALIGNED(addr) IS_ALIGNED((unsigned long)(addr), PAGE_SIZE) 1325 return PAGE_SIZE << compound_order(page); 1352 return PAGE_SIZE << thp_order(page); 1412 * in units of PAGE_SIZE. 2163 return PAGE_SIZE << folio_order(folio); 3490 return PAGE_SIZE; 3513 vm_end = -PAGE_SIZE; [all...] |
/linux-master/fs/ntfs3/ |
H A D | super.c | 1090 sb_set_blocksize(sb, min_t(u32, sbi->cluster_size, PAGE_SIZE)); 1132 u32 block_size = min_t(u32, sector_size, PAGE_SIZE); 1438 for (done = idx = 0; done < bytes; done += PAGE_SIZE, idx++) { 1448 min(PAGE_SIZE, tail)); 1512 for (i = 0; i < PAGE_SIZE / sizeof(u16); i++) 1515 memcpy(dst, src, PAGE_SIZE);
|
H A D | ntfs_fs.h | 126 struct buffer_head *bh[PAGE_SIZE >> SECTOR_SHIFT];
|
H A D | inode.c | 594 create && sbi->cluster_size > PAGE_SIZE); 657 off = vbo & (PAGE_SIZE - 1); 1017 len = pos + PAGE_SIZE > log_size ? (log_size - pos) : PAGE_SIZE; 1100 size_t op = bytes > PAGE_SIZE ? PAGE_SIZE : bytes; 1119 data = Add2Ptr(data, PAGE_SIZE); 2092 ret = kmalloc(PAGE_SIZE, GFP_NOFS); 2096 err = ntfs_readlink_hlp(de, inode, ret, PAGE_SIZE);
|
H A D | file.c | 138 zerofrom = pos & (PAGE_SIZE - 1); 139 len = PAGE_SIZE - zerofrom; 148 zero_user_segment(page, zerofrom, PAGE_SIZE); 184 u32 from = vbo & (PAGE_SIZE - 1); 185 pgoff_t idx_end = (vbo_to + PAGE_SIZE - 1) >> PAGE_SHIFT; 195 to = (page_off + PAGE_SIZE) > vbo_to ? (vbo_to - page_off) : 196 PAGE_SIZE; 441 sbi->cluster_size, PAGE_SIZE)); 866 if (frame_size < PAGE_SIZE) { 930 zero_user_segment(page, off, PAGE_SIZE); [all...] |
/linux-master/fs/netfs/ |
H A D | buffered_write.c | 123 pgoff_t index = pos / PAGE_SIZE; 127 fgp_flags |= fgf_set_order(pos % PAGE_SIZE + part); 169 size_t max_chunk = PAGE_SIZE << MAX_PAGECACHE_ORDER; 568 pgoff_t index = start / PAGE_SIZE; 569 pgoff_t last = (start + len - 1) / PAGE_SIZE, next; 606 pgoff_t index = start / PAGE_SIZE; 607 pgoff_t last = (start + len - 1) / PAGE_SIZE, next; 646 XA_STATE(xas, &mapping->i_pages, wreq->start / PAGE_SIZE); 652 last = (wreq->start + wreq->len - 1) / PAGE_SIZE; 783 pgoff_t index = (start + *_len) / PAGE_SIZE; [all...] |