/linux-master/fs/nilfs2/ |
H A D | dir.c | 40 #if (PAGE_SIZE >= 65536) 49 #if (PAGE_SIZE >= 65536) 76 if (last_byte > PAGE_SIZE) 77 last_byte = PAGE_SIZE; 280 ctx->pos += PAGE_SIZE - offset;
|
/linux-master/fs/fuse/ |
H A D | inode.c | 1239 ra_pages = arg->max_readahead / PAGE_SIZE; 1335 ra_pages = fc->max_read / PAGE_SIZE; 1367 ia->in.max_readahead = fm->sb->s_bdi->ra_pages * PAGE_SIZE; 1691 sb->s_blocksize = PAGE_SIZE;
|
H A D | file.c | 856 struct fuse_page_desc desc = { .length = PAGE_SIZE }; 999 fc->max_read / PAGE_SIZE); 1026 ap->descs[i].length = PAGE_SIZE; 1167 if (count >= PAGE_SIZE - offset) 1168 count -= PAGE_SIZE - offset; 1191 unsigned offset = pos & (PAGE_SIZE - 1); 1202 size_t bytes = min_t(size_t, PAGE_SIZE - offset, 1237 if (offset == PAGE_SIZE) 1241 if (tmp == PAGE_SIZE) 1510 npages = DIV_ROUND_UP(ret, PAGE_SIZE); [all...] |
H A D | fuse_i.h | 1020 descs[i].length = PAGE_SIZE - descs[i].offset;
|
H A D | dir.c | 1571 struct fuse_page_desc desc = { .length = PAGE_SIZE - 1 }; 1594 if (WARN_ON(res >= PAGE_SIZE))
|
H A D | cuse.c | 437 BUILD_BUG_ON(CUSE_INIT_INFO_MAX > PAGE_SIZE);
|
/linux-master/fs/9p/ |
H A D | vfs_file.c | 490 .range_start = (loff_t)vma->vm_pgoff * PAGE_SIZE, 492 .range_end = (loff_t)vma->vm_pgoff * PAGE_SIZE +
|
/linux-master/drivers/scsi/ |
H A D | scsi_lib.c | 3142 sg_len = PAGE_SIZE - *offset;
|
/linux-master/drivers/s390/net/ |
H A D | ism_drv.c | 184 sba = dma_alloc_coherent(&ism->pdev->dev, PAGE_SIZE, &dma_handle, 195 dma_free_coherent(&ism->pdev->dev, PAGE_SIZE, sba, dma_handle); 211 ieq = dma_alloc_coherent(&ism->pdev->dev, PAGE_SIZE, &dma_handle, 223 dma_free_coherent(&ism->pdev->dev, PAGE_SIZE, ieq, dma_handle); 245 dma_free_coherent(&ism->pdev->dev, PAGE_SIZE, 265 dma_free_coherent(&ism->pdev->dev, PAGE_SIZE, 449 bytes = max_bytes(offset, size, PAGE_SIZE);
|
/linux-master/drivers/s390/cio/ |
H A D | device.c | 227 len = snprint_alias(buf, PAGE_SIZE, id, "\n"); 229 return len > PAGE_SIZE ? PAGE_SIZE : len;
|
/linux-master/drivers/net/ |
H A D | tun.c | 1488 if (fragsz == 0 || fragsz > PAGE_SIZE) { 1520 if (prepad + len < PAGE_SIZE) 1523 if (len - linear > MAX_SKB_FRAGS * (PAGE_SIZE << PAGE_ALLOC_COSTLY_ORDER)) 1524 linear = len - MAX_SKB_FRAGS * (PAGE_SIZE << PAGE_ALLOC_COSTLY_ORDER); 1594 SKB_DATA_ALIGN(sizeof(struct skb_shared_info)) > PAGE_SIZE)
|
/linux-master/drivers/net/ethernet/stmicro/stmmac/ |
H A D | stmmac_main.c | 2023 num_pages = DIV_ROUND_UP(dma_conf->dma_buf_sz, PAGE_SIZE); 5432 buf_sz = DIV_ROUND_UP(priv->dma_conf.dma_buf_sz, PAGE_SIZE) * PAGE_SIZE;
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | main.c | 603 if (MLX5_CAP_GEN_MAX(dev, uar_4k) && PAGE_SIZE > 4096)
|
H A D | en_main.c | 597 if (next_frag.offset + frag_info[f].frag_stride > PAGE_SIZE) { 623 WARN_ON(rq->wqe.info.arr[0].frag_stride != PAGE_SIZE); 692 PAGE_SIZE, rq->buff.map_dir); 702 dma_unmap_page(rq->pdev, rq->wqe_overflow.addr, PAGE_SIZE, 908 pp_params.max_len = PAGE_SIZE; 4283 * 2. Size of SKBs allocated on XDP_PASS <= PAGE_SIZE.
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | reporter_tx.c | 372 key.size = PAGE_SIZE; 424 key.size = PAGE_SIZE;
|
/linux-master/drivers/net/ethernet/mediatek/ |
H A D | mtk_wed.c | 29 #define MTK_WED_BUF_PER_PAGE (PAGE_SIZE / 2048) 30 #define MTK_WED_RX_BUF_PER_PAGE (PAGE_SIZE / MTK_WED_PAGE_BUF_SIZE) 33 #define MTK_WED_AMSDU_BUF_SIZE (PAGE_SIZE << 4) 677 page_phys = dma_map_page(dev->hw->dev, page, 0, PAGE_SIZE, 686 dma_sync_single_for_cpu(dev->hw->dev, page_phys, PAGE_SIZE, 723 dma_sync_single_for_device(dev->hw->dev, page_phys, PAGE_SIZE, 750 dma_unmap_page(dev->hw->dev, page_phy, PAGE_SIZE, 800 page_phys = dma_map_page(dev->hw->dev, page, 0, PAGE_SIZE, 809 dma_sync_single_for_cpu(dev->hw->dev, page_phys, PAGE_SIZE, 820 dma_sync_single_for_device(dev->hw->dev, page_phys, PAGE_SIZE, [all...] |
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_kms.c | 404 if (cmd->dma.guest.ptr.offset % PAGE_SIZE ||
|
H A D | vmwgfx_drv.c | 401 .size = PAGE_SIZE, 966 dev_priv->max_mob_pages = mem_size * 1024 / PAGE_SIZE;
|
H A D | vmwgfx_blit.c | 374 copy_size = min_t(u32, copy_size, PAGE_SIZE - dst_page_offset); 375 copy_size = min_t(u32, copy_size, PAGE_SIZE - src_page_offset);
|
/linux-master/drivers/gpu/drm/ttm/ |
H A D | ttm_pool.c | 110 vaddr = dma_alloc_attrs(pool->dev, (1ULL << order) * PAGE_SIZE, 158 dma_free_attrs(pool->dev, (1UL << order) * PAGE_SIZE, vaddr, dma->addr, 197 size_t size = (1ULL << order) * PAGE_SIZE; 206 addr += PAGE_SIZE;
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_vm.c | 993 pages_addr[pfn] + PAGE_SIZE; 1001 pages_addr[idx - 1] + PAGE_SIZE))
|
H A D | amdgpu_cs.c | 140 if (size != PAGE_SIZE || data->offset > (size - 8))
|
/linux-master/drivers/char/ |
H A D | random.c | 456 BUILD_BUG_ON(PAGE_SIZE % sizeof(block) != 0); 457 if (ret % PAGE_SIZE == 0) { 1415 BUILD_BUG_ON(PAGE_SIZE % sizeof(block) != 0); 1416 if (ret % PAGE_SIZE == 0) {
|
/linux-master/arch/arm64/mm/ |
H A D | pageattr.c | 49 * This function assumes that the range is mapped with PAGE_SIZE pages. 71 unsigned long size = PAGE_SIZE * numpages; 112 PAGE_SIZE, set_mask, clear_mask); 156 return __change_memory_common(addr, PAGE_SIZE * numpages, 160 return __change_memory_common(addr, PAGE_SIZE * numpages, 177 PAGE_SIZE, change_page_range, &data); 192 PAGE_SIZE, change_page_range, &data);
|
H A D | hugetlbpage.c | 104 *pgsize = PAGE_SIZE; 136 *pgsize = PAGE_SIZE;
|