/linux-master/drivers/misc/eeprom/ |
H A D | at24.c | 84 u16 page_size; member in struct:at24_data 390 next_page = roundup(offset + 1, at24->page_size); 590 u32 byte_len, page_size, flags, addrw; local 610 err = device_property_read_u32(dev, "pagesize", &page_size); 617 page_size = 1; 648 page_size = 1; 650 if (!page_size) { 651 dev_err(dev, "page_size must not be 0!\n"); 655 if (!is_power_of_2(page_size)) 656 dev_warn(dev, "page_size look [all...] |
/linux-master/include/linux/ |
H A D | mm.h | 1323 static inline unsigned long page_size(struct page *page) function
|
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt.c | 3391 dma_free_coherent(&pdev->dev, rmem->page_size, 3401 pg_tbl_size = rmem->page_size; 3424 pg_tbl_size = rmem->page_size; 3436 rmem->page_size, 3444 rmem->page_size); 3839 rmem->page_size = HW_CMPD_RING_SIZE; 3949 rmem->page_size = HW_CMPD_RING_SIZE; 3961 rmem->page_size = HW_RXBD_RING_SIZE; 3970 rmem->page_size = HW_RXBD_RING_SIZE; 3981 rmem->page_size [all...] |
/linux-master/drivers/mtd/nand/raw/ |
H A D | qcom_nandc.c | 454 u32 page_size; member in struct:qcom_nand_boot_partition 1838 end = start + boot_partition->page_size; 1852 end = start + boot_partition->page_size; 3237 &boot_partition->page_size); 3244 if (boot_partition->page_size % mtd->writesize) { 3251 boot_partition->page_size /= mtd->writesize;
|
/linux-master/drivers/mtd/nand/raw/brcmnand/ |
H A D | brcmnand.c | 290 unsigned int page_size; member in struct:brcmnand_cfg 1258 int sectors = cfg->page_size / (512 << cfg->sector_size_1k); 1276 int sectors = cfg->page_size / (512 << cfg->sector_size_1k); 1289 if (cfg->page_size > 512) { 1316 int sectors = cfg->page_size / (512 << cfg->sector_size_1k); 1334 int sectors = cfg->page_size / (512 << cfg->sector_size_1k); 1393 int sectors = p->page_size / (512 << p->sector_size_1k); 1411 if (p->page_size == 512) 2549 u8 block_size = 0, page_size = 0, device_size = 0; local 2580 if (ctrl->page_sizes[i] == cfg->page_size) { [all...] |
/linux-master/tools/testing/selftests/kvm/ |
H A D | max_guest_memory_test.c | 69 vcpu_args_set(vcpu, 3, info->start_gpa, info->end_gpa, vm->page_size); 104 ~((uint64_t)vm->page_size - 1); 226 for (i = 0; i < slot_size; i += vm->page_size) 245 for (i = 0; i < slot_size; i += vm->page_size)
|
/linux-master/mm/ |
H A D | madvise.c | 1121 size = page_size(compound_head(page));
|
H A D | vmalloc.c | 3236 unsigned long page_size; local 3238 page_size = PAGE_SIZE << page_order; 3240 end = max(addr + page_size, end);
|
/linux-master/arch/arm64/mm/ |
H A D | hugetlbpage.c | 153 ncontig = num_contig_ptes(page_size(pte_page(orig_pte)), &pgsize);
|
/linux-master/io_uring/ |
H A D | io_uring.c | 3466 if (sz > page_size(page))
|
/linux-master/include/asm-generic/ |
H A D | hyperv-tlfs.h | 462 u64 page_size:1; member in struct:hv_gpa_page_range::__anon116
|
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_txrx.h | 136 int page_size, pad_size; local 138 page_size = ALIGN(rx_buf_len, PAGE_SIZE / 2); 139 pad_size = SKB_WITH_OVERHEAD(page_size) - rx_buf_len;
|
/linux-master/drivers/net/ethernet/intel/e1000e/ |
H A D | hw.h | 664 u16 page_size; member in struct:e1000_nvm_info
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_pt.c | 905 u64 page_size = 1ull << xe_pt_shift(xe_pt->level); local 910 start = entry->ofs * page_size; 911 end = start + page_size * entry->qwords;
|
H A D | xe_bo.c | 1824 dma_addr_t __xe_bo_addr(struct xe_bo *bo, u64 offset, size_t page_size) argument 1830 xe_assert(xe, page_size <= PAGE_SIZE); 1838 page_size, &cur); local 1844 page_size, &cur); 1849 dma_addr_t xe_bo_addr(struct xe_bo *bo, u64 offset, size_t page_size) argument 1853 return __xe_bo_addr(bo, offset, page_size); 2276 u32 page_size = max_t(u32, PAGE_SIZE, local 2281 page_size);
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | gen8_ppgtt.c | 524 unsigned int page_size; local 537 page_size = I915_GTT_PAGE_SIZE_2M; 542 page_size = I915_GTT_PAGE_SIZE; 557 page_size = I915_GTT_PAGE_SIZE_64K; 569 page_size = I915_GTT_PAGE_SIZE_64K; 578 GEM_BUG_ON(rem < page_size); 586 start += page_size; 587 iter->dma += page_size; 588 rem -= page_size; 601 if (unlikely(!IS_ALIGNED(iter->dma, page_size))) 630 unsigned int page_size; local [all...] |
/linux-master/arch/riscv/mm/ |
H A D | tlbflush.c | 156 unsigned int page_size) 159 start, end - start, page_size); local 154 flush_tlb_mm_range(struct mm_struct *mm, unsigned long start, unsigned long end, unsigned int page_size) argument
|
/linux-master/drivers/scsi/lpfc/ |
H A D | lpfc_sli4.h | 224 uint16_t page_size; /* size of page allocated for this queue */ member in struct:lpfc_queue 1073 uint32_t page_size,
|
H A D | lpfc_sli.c | 15802 dma_free_coherent(&queue->phba->pcidev->dev, queue->page_size, 15821 * @page_size: The size of a queue page 15831 lpfc_sli4_queue_alloc(struct lpfc_hba *phba, uint32_t page_size, argument 15840 hw_page_size = page_size; 15869 queue->page_size = hw_page_size; 16297 (cq->page_size / SLI4_PAGE_SIZE)); 16344 memset(dmabuf->virt, 0, cq->page_size); 16462 hw_page_size = cq->page_size; 16901 hw_page_size = wq->page_size; 16923 (wq->page_size > SLI4_PAGE_SIZ [all...] |
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_dmem.c | 633 *dma_addr = dma_map_page(dev, spage, 0, page_size(spage), 642 if (drm->dmem->migrate.clear_func(drm, page_size(dpage),
|
/linux-master/tools/testing/selftests/mm/ |
H A D | uffd-unit-tests.c | 195 page_size = default_huge_page_size(); 197 page_size = psize(); 199 nr_pages = UFFD_TEST_MEM_SIZE / page_size; 323 if (test_pin && pin_pages(&args, area_dst, page_size)) 361 if (uffd_register(uffd, area_dst, nr_pages * page_size, 368 wp_range(uffd, (uint64_t)area_dst, page_size, true); 373 wp_range(uffd, (uint64_t)area_dst, page_size, false); 378 wp_range(uffd, (uint64_t)area_dst, page_size, true); 379 if (madvise(area_dst, page_size, MADV_DONTNEED)) 389 if (madvise(area_dst, page_size, MADV_DONTNEE 716 page_size, false); local [all...] |
H A D | uffd-common.c | 13 unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size; variable 45 if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED)) 51 *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, 67 if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED)) 70 if (madvise(rel_area, nr_pages * page_size, MADV_REMOVE)) 77 off_t size = nr_pages * page_size; 121 if (madvise(rel_area, nr_pages * page_size, MADV_REMOVE)) 128 size_t bytes = nr_pages * page_size, hpage_size = read_pmd_pagesize(); 261 if (munmap(*area, nr_pages * page_size)) 445 wp_range(uffd, msg->arg.pagefault.address, page_size, fals [all...] |
H A D | uffd-common.h | 60 ((pthread_mutex_t *) ((___area) + (___nr)*page_size)) 67 ((___area) + (___nr)*page_size + \ 100 extern unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size;
|
/linux-master/tools/lib/bpf/ |
H A D | libbpf.c | 2687 * page_size) then just return original size and let kernel reject it 12784 perf_event_read_simple(void *mmap_mem, size_t mmap_size, size_t page_size, argument 12791 void *base = ((__u8 *)header) + page_size; 12862 size_t page_size; member in struct:perf_buffer 12877 munmap(cpu_buf->base, pb->mmap_size + pb->page_size)) 12936 cpu_buf->base = mmap(NULL, pb->mmap_size + pb->page_size, 13071 pb->page_size = getpagesize(); 13072 pb->mmap_size = pb->page_size * page_cnt; 13221 pb->page_size, &cpu_buf->buf,
|
/linux-master/drivers/net/wireless/intel/iwlwifi/fw/ |
H A D | dbg.c | 1225 u32 page_size = fwrt->fw_paging_db[idx].fw_paging_size; local 1228 range->range_data_size = cpu_to_le32(page_size); 1229 dma_sync_single_for_cpu(fwrt->trans->dev, addr, page_size, 1231 memcpy(range->data, page_address(page), page_size); 1232 dma_sync_single_for_device(fwrt->trans->dev, addr, page_size, 1243 u32 page_size; local 1252 page_size = fwrt->trans->init_dram.paging[idx].size; 1255 range->range_data_size = cpu_to_le32(page_size); 1257 page_size);
|