Searched refs:map (Results 126 - 150 of 2546) sorted by last modified time

1234567891011>>

/linux-master/fs/xfs/
H A Dxfs_buf.c77 * enough to know it doesn't have to map a single page, so the check has
221 struct xfs_buf_map *map,
263 bp->b_rhash_key = map[0].bm_bn;
266 bp->b_maps[i].bm_bn = map[i].bm_bn;
267 bp->b_maps[i].bm_len = map[i].bm_len;
268 bp->b_length += map[i].bm_len;
477 const struct xfs_buf_map *map = arg->key; local
486 if (bp->b_rhash_key != map->bm_bn)
489 if (unlikely(bp->b_length != map->bm_len)) {
498 if (!(map
219 _xfs_buf_alloc( struct xfs_buftarg *target, struct xfs_buf_map *map, int nmaps, xfs_buf_flags_t flags, struct xfs_buf **bpp) argument
531 xfs_buf_map_verify( struct xfs_buftarg *btp, struct xfs_buf_map *map) argument
589 xfs_buf_lookup( struct xfs_buf_cache *bch, struct xfs_buf_map *map, xfs_buf_flags_t flags, struct xfs_buf **bpp) argument
622 xfs_buf_find_insert( struct xfs_buftarg *btp, struct xfs_buf_cache *bch, struct xfs_perag *pag, struct xfs_buf_map *cmap, struct xfs_buf_map *map, int nmaps, xfs_buf_flags_t flags, struct xfs_buf **bpp) argument
690 xfs_buftarg_get_pag( struct xfs_buftarg *btp, const struct xfs_buf_map *map) argument
717 xfs_buf_get_map( struct xfs_buftarg *btp, struct xfs_buf_map *map, int nmaps, xfs_buf_flags_t flags, struct xfs_buf **bpp) argument
845 xfs_buf_read_map( struct xfs_buftarg *target, struct xfs_buf_map *map, int nmaps, xfs_buf_flags_t flags, struct xfs_buf **bpp, const struct xfs_buf_ops *ops, xfs_failaddr_t fa) argument
930 xfs_buf_readahead_map( struct xfs_buftarg *target, struct xfs_buf_map *map, int nmaps, const struct xfs_buf_ops *ops) argument
1521 xfs_buf_ioapply_map( struct xfs_buf *bp, int map, int *buf_offset, int *count, blk_opf_t op) argument
[all...]
H A Dxfs_trans.h174 struct xfs_buf_map *map, int nmaps, xfs_buf_flags_t flags,
186 DEFINE_SINGLE_BUF_MAP(map, blkno, numblks);
187 return xfs_trans_get_buf_map(tp, target, &map, 1, flags, bpp);
193 struct xfs_buf_map *map, int nmaps,
209 DEFINE_SINGLE_BUF_MAP(map, blkno, numblks);
210 return xfs_trans_read_buf_map(mp, tp, target, &map, 1,
/linux-master/fs/ext4/
H A Dsuper.c5799 struct ext4_map_blocks map; local
5805 map.m_lblk = *block;
5806 map.m_len = 1;
5807 ret = ext4_map_blocks(NULL, journal->j_inode, &map, 0);
5815 *block = map.m_pblk;
/linux-master/drivers/ufs/core/
H A Dufshcd.c1257 * the budget map before scsi_host_queue_ready() checks the host_self_blocked
2931 struct blk_mq_queue_map *map = &shost->tag_set.map[i]; local
2933 map->nr_queues = hba->nr_queues[i];
2934 if (!map->nr_queues)
2936 map->queue_offset = queue_offset;
2938 map->queue_offset = 0;
2940 blk_mq_map_queues(map);
2941 queue_offset += map->nr_queues;
/linux-master/drivers/nvme/host/
H A Dfc.c2258 * and CQEs and dma map them - mapping their respective entries
2868 struct blk_mq_queue_map *map = &set->map[i]; local
2870 if (!map->nr_queues) {
2875 /* Call LLDD map queue functionality if defined */
2878 map);
2880 blk_mq_map_queues(map);
/linux-master/drivers/net/ethernet/intel/ice/
H A Dice_ethtool.c2155 const struct ethtool_forced_speed_map *map; local
2159 map = ice_adv_lnk_speed_maps + i;
2160 if (linkmode_intersects(ks->link_modes.advertising, map->caps))
2161 adv_link_speed |= ice_speed_to_aq_link(map->speed);
/linux-master/drivers/net/ethernet/intel/i40e/
H A Di40e_virtchnl_pf.c366 * @vecmap: irq map info
368 * configure irq link list from the map
2443 * i40e_validate_queue_map - check queue map is valid
2446 * @queuemap: Tx or Rx queue map
2448 * check if Tx or Rx queue map is valid
2476 * queue map
2482 struct virtchnl_vector_map *map; local
2499 map = &irqmap_info->vecmap[i];
2501 if (!i40e_vc_isvalid_vector_id(vf, map->vector_id) ||
2502 !i40e_vc_isvalid_vsi_id(vf, map
[all...]
/linux-master/drivers/gpu/drm/xe/
H A Dxe_vm_types.h275 /** struct xe_vma_op_map - VMA map operation */
277 /** @vma: VMA to map */
349 /** @map: VMA map operation specific data */
350 struct xe_vma_op_map map; member in union:xe_vma_op::__anon107
H A Dxe_pt.c147 struct iosys_map *map = &pt->bo->vmap; local
156 xe_map_memset(vm->xe, map, 0, 0, SZ_4K);
160 xe_pt_write(vm->xe, map, i, empty);
369 struct iosys_map *map = &parent->bo->vmap; local
374 xe_pt_write(xe_walk->vm->xe, map, offset, pte);
795 struct iosys_map *map, void *data,
804 if (map)
805 xe_map_wr(tile_to_xe(tile), map, (qword_ofs + i) * local
1490 struct xe_tile *tile, struct iosys_map *map,
1498 if (map
794 xe_vm_populate_pgtable(struct xe_migrate_pt_update *pt_update, struct xe_tile *tile, struct iosys_map *map, void *data, u32 qword_ofs, u32 num_qwords, const struct xe_vm_pgtable_update *update) argument
1489 xe_migrate_clear_pgtable_callback(struct xe_migrate_pt_update *pt_update, struct xe_tile *tile, struct iosys_map *map, void *ptr, u32 qword_ofs, u32 num_qwords, const struct xe_vm_pgtable_update *update) argument
1500 xe_map_wr(tile_to_xe(tile), map, (qword_ofs + i) * local
[all...]
H A Dxe_guc_submit.c537 struct iosys_map map = xe_lrc_parallel_map(lrc); local
551 xe_map_memset(xe, &map, 0, 0, PARALLEL_SCRATCH_SIZE - WQ_SIZE);
552 parallel_write(xe, map, wq_desc.wq_status, WQ_STATUS_ACTIVE);
581 struct iosys_map map = xe_lrc_parallel_map(q->lrc); local
588 q->guc->wqi_head = parallel_read(xe, map, wq_desc.head);
609 struct iosys_map map = xe_lrc_parallel_map(q->lrc); local
617 parallel_write(xe, map, wq[q->guc->wqi_tail / sizeof(u32)],
629 struct iosys_map map = xe_lrc_parallel_map(q->lrc); local
657 iosys_map_incr(&map, offsetof(struct guc_submit_parallel_scratch,
659 xe_map_memcpy_to(xe, &map,
1741 struct iosys_map map = xe_lrc_parallel_map(q->lrc); local
[all...]
/linux-master/drivers/gpu/drm/nouveau/
H A Dnouveau_uvmm.c12 * Actual map/unmap operations within the fence signalling critical path are
15 * list in order to map/unmap it's entries, can't occur concurrently.
54 struct nouveau_uvma *map; member in struct:nouveau_uvma_prealloc
468 op_map_prepare_unwind(new->map);
612 * Note: @args should not be NULL when calling for a map operation.
630 ret = op_map_prepare(uvmm, &new->map, &op->map, args);
638 op_map_prepare_unwind(new->map);
763 return op->map.gem.obj;
836 op_map(new->map);
[all...]
/linux-master/drivers/gpu/drm/i915/gt/
H A Dintel_engine_cs.c932 u8 class, const u8 *map, u8 num_instances)
943 if (intel_engines[i].instance == map[j]) {
959 const u8 map[] = { 0, 2, 4, 6, 1, 3, 5, 7 }; local
962 map, ARRAY_SIZE(map));
965 u8 map[MAX_ENGINE_INSTANCE + 1]; local
968 map[i] = i;
970 map, ARRAY_SIZE(map));
1085 * actually map i
931 populate_logical_ids(struct intel_gt *gt, u8 *logical_ids, u8 class, const u8 *map, u8 num_instances) argument
[all...]
/linux-master/drivers/gpu/drm/
H A Ddrm_prime.c680 * @map: the virtual address of the buffer
684 * The kernel virtual address is returned in map.
688 int drm_gem_dmabuf_vmap(struct dma_buf *dma_buf, struct iosys_map *map) argument
692 return drm_gem_vmap(obj, map);
699 * @map: the virtual address of the buffer
704 void drm_gem_dmabuf_vunmap(struct dma_buf *dma_buf, struct iosys_map *map) argument
708 drm_gem_vunmap(obj, map);
/linux-master/drivers/block/null_blk/
H A Dmain.c1446 * poll queues to map queues.
1466 struct blk_mq_queue_map *map = &set->map[i]; local
1470 map->nr_queues = submit_queues;
1473 map->nr_queues = 0;
1476 map->nr_queues = poll_queues;
1479 map->queue_offset = qoff;
1480 qoff += map->nr_queues;
1481 blk_mq_map_queues(map);
/linux-master/drivers/base/regmap/
H A Dregcache-maple.c16 static int regcache_maple_read(struct regmap *map, argument
19 struct maple_tree *mt = map->cache;
38 static int regcache_maple_write(struct regmap *map, unsigned int reg, argument
41 struct maple_tree *mt = map->cache;
77 map->alloc_flags);
95 ret = mas_store_gfp(&mas, entry, map->alloc_flags);
107 static int regcache_maple_drop(struct regmap *map, unsigned int min, argument
110 struct maple_tree *mt = map->cache;
137 map->alloc_flags);
151 map
189 regcache_maple_sync_block(struct regmap *map, unsigned long *entry, struct ma_state *mas, unsigned int min, unsigned int max) argument
238 regcache_maple_sync(struct regmap *map, unsigned int min, unsigned int max) argument
293 regcache_maple_exit(struct regmap *map) argument
315 regcache_maple_insert_block(struct regmap *map, int first, int last) argument
344 regcache_maple_init(struct regmap *map) argument
[all...]
/linux-master/drivers/base/
H A Dcore.c33 #include <linux/dma-map-ops.h> /* for dma_default_coherent */
/linux-master/drivers/ata/
H A Dsata_gemini.c322 struct regmap *map; local
337 map = syscon_regmap_lookup_by_phandle(np, "syscon");
338 if (IS_ERR(map)) {
340 return PTR_ERR(map);
373 ret = regmap_update_bits(map, GEMINI_GLOBAL_MISC_CTRL, gmask, gmode);
/linux-master/security/
H A Dsecurity.c1099 * security_vm_enough_memory_mm() - Check if allocating a new mem map is allowed
5409 * security_bpf_map() - Check if access to a bpf map is allowed
5410 * @map: bpf map
5418 int security_bpf_map(struct bpf_map *map, fmode_t fmode) argument
5420 return call_int_hook(bpf_map, map, fmode);
5438 * security_bpf_map_create() - Check if BPF map creation is allowed
5439 * @map: BPF map object
5440 * @attr: BPF syscall attributes used to create BPF map
5448 security_bpf_map_create(struct bpf_map *map, union bpf_attr *attr, struct bpf_token *token) argument
5527 security_bpf_map_free(struct bpf_map *map) argument
[all...]
/linux-master/drivers/irqchip/
H A Dirq-armada-370-xp.c88 * At ->map() time, a global interrupt is unmasked at the per-CPU
90 * the current CPU, running the ->map() code. This allows to have
106 * At ->map() time, a per-CPU interrupt is unmasked at the global
368 unsigned long map = 0; local
373 map |= 1 << cpu_logical_map(cpu);
382 writel((map << 8) | d->hwirq, main_int_base +
592 .map = armada_370_xp_mpic_irq_map,
/linux-master/arch/x86/mm/
H A Dmem_encrypt_amd.c84 * page(s) that map this memory. It assumes that eventually the memory is
103 * There are limited number of early mapping slots, so map (at most)
157 bool map)
166 pmd = map ? (paddr & PMD_MASK) + pmd_flags : 0;
463 /* Update the protection map with memory encryption mask */
156 __sme_early_map_unmap_mem(void *vaddr, unsigned long size, bool map) argument
/linux-master/include/linux/gpio/
H A Ddriver.h219 * @map:
223 unsigned int *map; member in struct:gpio_irq_chip
290 * will allocate and map all IRQs during initialization.
/linux-master/drivers/scsi/
H A Dscsi_scan.c221 bool need_alloc = !sdev->budget_map.map;
239 * Request queue has to be frozen for reallocating budget map,
351 * Use .can_queue as budget map's depth because we have to
1095 * Set up budget map again since memory consumption of
1096 * the map depends on actual queue depth.
/linux-master/drivers/scsi/qla2xxx/
H A Dqla_os.c8075 struct blk_mq_queue_map *qmap = &shost->tag_set.map[HCTX_TYPE_DEFAULT];
H A Dqla_iocb.c3817 int map, pos; local
3826 * index map in firmware starts with 1; decrement index
3829 map = (sp->u.iocb_cmd.u.ctrlvp.vp_index - 1) / 8;
3831 vce->vp_idx_map[map] |= 1 << pos;
/linux-master/drivers/md/
H A Ddm-integrity.c4750 .map = dm_integrity_map,

Completed in 538 milliseconds

1234567891011>>