/linux-master/fs/xfs/ |
H A D | xfs_buf.c | 77 * enough to know it doesn't have to map a single page, so the check has 221 struct xfs_buf_map *map, 263 bp->b_rhash_key = map[0].bm_bn; 266 bp->b_maps[i].bm_bn = map[i].bm_bn; 267 bp->b_maps[i].bm_len = map[i].bm_len; 268 bp->b_length += map[i].bm_len; 477 const struct xfs_buf_map *map = arg->key; local 486 if (bp->b_rhash_key != map->bm_bn) 489 if (unlikely(bp->b_length != map->bm_len)) { 498 if (!(map 219 _xfs_buf_alloc( struct xfs_buftarg *target, struct xfs_buf_map *map, int nmaps, xfs_buf_flags_t flags, struct xfs_buf **bpp) argument 531 xfs_buf_map_verify( struct xfs_buftarg *btp, struct xfs_buf_map *map) argument 589 xfs_buf_lookup( struct xfs_buf_cache *bch, struct xfs_buf_map *map, xfs_buf_flags_t flags, struct xfs_buf **bpp) argument 622 xfs_buf_find_insert( struct xfs_buftarg *btp, struct xfs_buf_cache *bch, struct xfs_perag *pag, struct xfs_buf_map *cmap, struct xfs_buf_map *map, int nmaps, xfs_buf_flags_t flags, struct xfs_buf **bpp) argument 690 xfs_buftarg_get_pag( struct xfs_buftarg *btp, const struct xfs_buf_map *map) argument 717 xfs_buf_get_map( struct xfs_buftarg *btp, struct xfs_buf_map *map, int nmaps, xfs_buf_flags_t flags, struct xfs_buf **bpp) argument 845 xfs_buf_read_map( struct xfs_buftarg *target, struct xfs_buf_map *map, int nmaps, xfs_buf_flags_t flags, struct xfs_buf **bpp, const struct xfs_buf_ops *ops, xfs_failaddr_t fa) argument 930 xfs_buf_readahead_map( struct xfs_buftarg *target, struct xfs_buf_map *map, int nmaps, const struct xfs_buf_ops *ops) argument 1521 xfs_buf_ioapply_map( struct xfs_buf *bp, int map, int *buf_offset, int *count, blk_opf_t op) argument [all...] |
H A D | xfs_trans.h | 174 struct xfs_buf_map *map, int nmaps, xfs_buf_flags_t flags, 186 DEFINE_SINGLE_BUF_MAP(map, blkno, numblks); 187 return xfs_trans_get_buf_map(tp, target, &map, 1, flags, bpp); 193 struct xfs_buf_map *map, int nmaps, 209 DEFINE_SINGLE_BUF_MAP(map, blkno, numblks); 210 return xfs_trans_read_buf_map(mp, tp, target, &map, 1,
|
/linux-master/fs/ext4/ |
H A D | super.c | 5799 struct ext4_map_blocks map; local 5805 map.m_lblk = *block; 5806 map.m_len = 1; 5807 ret = ext4_map_blocks(NULL, journal->j_inode, &map, 0); 5815 *block = map.m_pblk;
|
/linux-master/drivers/ufs/core/ |
H A D | ufshcd.c | 1257 * the budget map before scsi_host_queue_ready() checks the host_self_blocked 2931 struct blk_mq_queue_map *map = &shost->tag_set.map[i]; local 2933 map->nr_queues = hba->nr_queues[i]; 2934 if (!map->nr_queues) 2936 map->queue_offset = queue_offset; 2938 map->queue_offset = 0; 2940 blk_mq_map_queues(map); 2941 queue_offset += map->nr_queues;
|
/linux-master/drivers/nvme/host/ |
H A D | fc.c | 2258 * and CQEs and dma map them - mapping their respective entries 2868 struct blk_mq_queue_map *map = &set->map[i]; local 2870 if (!map->nr_queues) { 2875 /* Call LLDD map queue functionality if defined */ 2878 map); 2880 blk_mq_map_queues(map);
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_ethtool.c | 2155 const struct ethtool_forced_speed_map *map; local 2159 map = ice_adv_lnk_speed_maps + i; 2160 if (linkmode_intersects(ks->link_modes.advertising, map->caps)) 2161 adv_link_speed |= ice_speed_to_aq_link(map->speed);
|
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_virtchnl_pf.c | 366 * @vecmap: irq map info 368 * configure irq link list from the map 2443 * i40e_validate_queue_map - check queue map is valid 2446 * @queuemap: Tx or Rx queue map 2448 * check if Tx or Rx queue map is valid 2476 * queue map 2482 struct virtchnl_vector_map *map; local 2499 map = &irqmap_info->vecmap[i]; 2501 if (!i40e_vc_isvalid_vector_id(vf, map->vector_id) || 2502 !i40e_vc_isvalid_vsi_id(vf, map [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_vm_types.h | 275 /** struct xe_vma_op_map - VMA map operation */ 277 /** @vma: VMA to map */ 349 /** @map: VMA map operation specific data */ 350 struct xe_vma_op_map map; member in union:xe_vma_op::__anon107
|
H A D | xe_pt.c | 147 struct iosys_map *map = &pt->bo->vmap; local 156 xe_map_memset(vm->xe, map, 0, 0, SZ_4K); 160 xe_pt_write(vm->xe, map, i, empty); 369 struct iosys_map *map = &parent->bo->vmap; local 374 xe_pt_write(xe_walk->vm->xe, map, offset, pte); 795 struct iosys_map *map, void *data, 804 if (map) 805 xe_map_wr(tile_to_xe(tile), map, (qword_ofs + i) * local 1490 struct xe_tile *tile, struct iosys_map *map, 1498 if (map 794 xe_vm_populate_pgtable(struct xe_migrate_pt_update *pt_update, struct xe_tile *tile, struct iosys_map *map, void *data, u32 qword_ofs, u32 num_qwords, const struct xe_vm_pgtable_update *update) argument 1489 xe_migrate_clear_pgtable_callback(struct xe_migrate_pt_update *pt_update, struct xe_tile *tile, struct iosys_map *map, void *ptr, u32 qword_ofs, u32 num_qwords, const struct xe_vm_pgtable_update *update) argument 1500 xe_map_wr(tile_to_xe(tile), map, (qword_ofs + i) * local [all...] |
H A D | xe_guc_submit.c | 537 struct iosys_map map = xe_lrc_parallel_map(lrc); local 551 xe_map_memset(xe, &map, 0, 0, PARALLEL_SCRATCH_SIZE - WQ_SIZE); 552 parallel_write(xe, map, wq_desc.wq_status, WQ_STATUS_ACTIVE); 581 struct iosys_map map = xe_lrc_parallel_map(q->lrc); local 588 q->guc->wqi_head = parallel_read(xe, map, wq_desc.head); 609 struct iosys_map map = xe_lrc_parallel_map(q->lrc); local 617 parallel_write(xe, map, wq[q->guc->wqi_tail / sizeof(u32)], 629 struct iosys_map map = xe_lrc_parallel_map(q->lrc); local 657 iosys_map_incr(&map, offsetof(struct guc_submit_parallel_scratch, 659 xe_map_memcpy_to(xe, &map, 1741 struct iosys_map map = xe_lrc_parallel_map(q->lrc); local [all...] |
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_uvmm.c | 12 * Actual map/unmap operations within the fence signalling critical path are 15 * list in order to map/unmap it's entries, can't occur concurrently. 54 struct nouveau_uvma *map; member in struct:nouveau_uvma_prealloc 468 op_map_prepare_unwind(new->map); 612 * Note: @args should not be NULL when calling for a map operation. 630 ret = op_map_prepare(uvmm, &new->map, &op->map, args); 638 op_map_prepare_unwind(new->map); 763 return op->map.gem.obj; 836 op_map(new->map); [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_engine_cs.c | 932 u8 class, const u8 *map, u8 num_instances) 943 if (intel_engines[i].instance == map[j]) { 959 const u8 map[] = { 0, 2, 4, 6, 1, 3, 5, 7 }; local 962 map, ARRAY_SIZE(map)); 965 u8 map[MAX_ENGINE_INSTANCE + 1]; local 968 map[i] = i; 970 map, ARRAY_SIZE(map)); 1085 * actually map i 931 populate_logical_ids(struct intel_gt *gt, u8 *logical_ids, u8 class, const u8 *map, u8 num_instances) argument [all...] |
/linux-master/drivers/gpu/drm/ |
H A D | drm_prime.c | 680 * @map: the virtual address of the buffer 684 * The kernel virtual address is returned in map. 688 int drm_gem_dmabuf_vmap(struct dma_buf *dma_buf, struct iosys_map *map) argument 692 return drm_gem_vmap(obj, map); 699 * @map: the virtual address of the buffer 704 void drm_gem_dmabuf_vunmap(struct dma_buf *dma_buf, struct iosys_map *map) argument 708 drm_gem_vunmap(obj, map);
|
/linux-master/drivers/block/null_blk/ |
H A D | main.c | 1446 * poll queues to map queues. 1466 struct blk_mq_queue_map *map = &set->map[i]; local 1470 map->nr_queues = submit_queues; 1473 map->nr_queues = 0; 1476 map->nr_queues = poll_queues; 1479 map->queue_offset = qoff; 1480 qoff += map->nr_queues; 1481 blk_mq_map_queues(map);
|
/linux-master/drivers/base/regmap/ |
H A D | regcache-maple.c | 16 static int regcache_maple_read(struct regmap *map, argument 19 struct maple_tree *mt = map->cache; 38 static int regcache_maple_write(struct regmap *map, unsigned int reg, argument 41 struct maple_tree *mt = map->cache; 77 map->alloc_flags); 95 ret = mas_store_gfp(&mas, entry, map->alloc_flags); 107 static int regcache_maple_drop(struct regmap *map, unsigned int min, argument 110 struct maple_tree *mt = map->cache; 137 map->alloc_flags); 151 map 189 regcache_maple_sync_block(struct regmap *map, unsigned long *entry, struct ma_state *mas, unsigned int min, unsigned int max) argument 238 regcache_maple_sync(struct regmap *map, unsigned int min, unsigned int max) argument 293 regcache_maple_exit(struct regmap *map) argument 315 regcache_maple_insert_block(struct regmap *map, int first, int last) argument 344 regcache_maple_init(struct regmap *map) argument [all...] |
/linux-master/drivers/base/ |
H A D | core.c | 33 #include <linux/dma-map-ops.h> /* for dma_default_coherent */
|
/linux-master/drivers/ata/ |
H A D | sata_gemini.c | 322 struct regmap *map; local 337 map = syscon_regmap_lookup_by_phandle(np, "syscon"); 338 if (IS_ERR(map)) { 340 return PTR_ERR(map); 373 ret = regmap_update_bits(map, GEMINI_GLOBAL_MISC_CTRL, gmask, gmode);
|
/linux-master/security/ |
H A D | security.c | 1099 * security_vm_enough_memory_mm() - Check if allocating a new mem map is allowed 5409 * security_bpf_map() - Check if access to a bpf map is allowed 5410 * @map: bpf map 5418 int security_bpf_map(struct bpf_map *map, fmode_t fmode) argument 5420 return call_int_hook(bpf_map, map, fmode); 5438 * security_bpf_map_create() - Check if BPF map creation is allowed 5439 * @map: BPF map object 5440 * @attr: BPF syscall attributes used to create BPF map 5448 security_bpf_map_create(struct bpf_map *map, union bpf_attr *attr, struct bpf_token *token) argument 5527 security_bpf_map_free(struct bpf_map *map) argument [all...] |
/linux-master/drivers/irqchip/ |
H A D | irq-armada-370-xp.c | 88 * At ->map() time, a global interrupt is unmasked at the per-CPU 90 * the current CPU, running the ->map() code. This allows to have 106 * At ->map() time, a per-CPU interrupt is unmasked at the global 368 unsigned long map = 0; local 373 map |= 1 << cpu_logical_map(cpu); 382 writel((map << 8) | d->hwirq, main_int_base + 592 .map = armada_370_xp_mpic_irq_map,
|
/linux-master/arch/x86/mm/ |
H A D | mem_encrypt_amd.c | 84 * page(s) that map this memory. It assumes that eventually the memory is 103 * There are limited number of early mapping slots, so map (at most) 157 bool map) 166 pmd = map ? (paddr & PMD_MASK) + pmd_flags : 0; 463 /* Update the protection map with memory encryption mask */ 156 __sme_early_map_unmap_mem(void *vaddr, unsigned long size, bool map) argument
|
/linux-master/include/linux/gpio/ |
H A D | driver.h | 219 * @map: 223 unsigned int *map; member in struct:gpio_irq_chip 290 * will allocate and map all IRQs during initialization.
|
/linux-master/drivers/scsi/ |
H A D | scsi_scan.c | 221 bool need_alloc = !sdev->budget_map.map; 239 * Request queue has to be frozen for reallocating budget map, 351 * Use .can_queue as budget map's depth because we have to 1095 * Set up budget map again since memory consumption of 1096 * the map depends on actual queue depth.
|
/linux-master/drivers/scsi/qla2xxx/ |
H A D | qla_os.c | 8075 struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT];
|
H A D | qla_iocb.c | 3817 int map, pos; local 3826 * index map in firmware starts with 1; decrement index 3829 map = (sp->u.iocb_cmd.u.ctrlvp.vp_index - 1) / 8; 3831 vce->vp_idx_map[map] |= 1 << pos;
|
/linux-master/drivers/md/ |
H A D | dm-integrity.c | 4750 .map = dm_integrity_map,
|