Searched refs:map (Results 76 - 100 of 2546) sorted by last modified time

1234567891011>>

/linux-master/fs/btrfs/
H A Dextent_map.c35 tree->map = RB_ROOT_CACHED;
161 * Previous extent map found, return as in this case the caller does not
232 * We can't modify an extent map that is in the tree and that is being
235 * the tree and 1 for this task (which is unpinning the extent map or
260 rb_erase_cached(&merge->rb_node, &tree->map);
272 rb_erase_cached(&merge->rb_node, &tree->map);
310 "no extent map found for inode %llu (root %lld) when unpinning extent range [%llu, %llu), generation %llu",
319 "found extent map for inode %llu (root %lld) with unexpected start offset %llu when unpinning extent range [%llu, %llu), generation %llu",
376 * Add new extent map to the extent tree
378 * @tree: tree to insert new map i
[all...]
H A Dextent_io.c1090 * coupled with a single extent map, and each range can have
1091 * an extent map with a different offset value relative to the
1304 * and does the loop to map the page into extents and bios.
2364 * map records are removed
2373 struct extent_map_tree *map = &btrfs_inode->extent_tree; local
2383 write_lock(&map->lock);
2384 em = lookup_extent_mapping(map, start, len);
2386 write_unlock(&map->lock);
2391 write_unlock(&map->lock);
2430 remove_extent_mapping(map, e
[all...]
/linux-master/virt/kvm/
H A Dkvm_main.c2804 * or the caller allows to map a writable pfn for a read fault
2830 * MMU, we lookup the page using GUP to map it, so the guest VCPU can
2859 /* map read fault as writable if possible */
2970 * @writable: whether it allows to map a writable host page for !@write_fault
2972 * The function will map a writable host page for these two cases:
3049 /* Do not map writable pfn in the readonly memslot. */
3149 int kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map) argument
3155 if (!map)
3174 map->page = page;
3175 map
3183 kvm_vcpu_unmap(struct kvm_vcpu *vcpu, struct kvm_host_map *map, bool dirty) argument
[all...]
/linux-master/arch/x86/kvm/
H A Dx86.c10012 struct kvm_apic_map *map; local
10020 map = rcu_dereference(vcpu->kvm->arch.apic_map);
10022 if (likely(map) && dest_id <= map->max_apic_id && map->phys_map[dest_id])
10023 target = map->phys_map[dest_id]->vcpu;
12704 * unless the memory map has changed due to process exit
H A Dlapic.c174 static inline bool kvm_apic_map_get_logical_dest(struct kvm_apic_map *map, argument
176 switch (map->logical_mode) {
178 /* Arbitrarily use the flat map so that @cluster isn't NULL. */
179 *cluster = map->xapic_flat_map;
184 u32 max_apic_id = map->max_apic_id;
189 offset = array_index_nospec(offset, map->max_apic_id + 1);
190 *cluster = &map->phys_map[offset];
199 *cluster = map->xapic_flat_map;
203 *cluster = map->xapic_cluster_map[(dest_id >> 4) & 0xf];
216 struct kvm_apic_map *map local
835 __pv_send_ipi(unsigned long *ipi_bitmap, struct kvm_apic_map *map, struct kvm_lapic_irq *irq, u32 min) argument
859 struct kvm_apic_map *map; local
1119 kvm_apic_is_broadcast_dest(struct kvm *kvm, struct kvm_lapic **src, struct kvm_lapic_irq *irq, struct kvm_apic_map *map) argument
1145 kvm_apic_map_get_dest_lapic(struct kvm *kvm, struct kvm_lapic **src, struct kvm_lapic_irq *irq, struct kvm_apic_map *map, struct kvm_lapic ***dst, unsigned long *bitmap) argument
1214 struct kvm_apic_map *map; local
1265 struct kvm_apic_map *map; local
1402 struct kvm_apic_map *map; local
[all...]
/linux-master/arch/x86/kvm/svm/
H A Dsvm.c2257 struct kvm_host_map map; local
2263 ret = kvm_vcpu_map(vcpu, gpa_to_gfn(svm->vmcb->save.rax), &map);
2270 vmcb12 = map.hva;
2282 kvm_vcpu_unmap(vcpu, &map, true);
4652 struct kvm_host_map map, map_save; local
4671 if (kvm_vcpu_map(vcpu, gpa_to_gfn(smram64->svm_guest_vmcb_gpa), &map))
4694 vmcb12 = map.hva;
4707 kvm_vcpu_unmap(vcpu, &map, true);
/linux-master/tools/perf/util/
H A Dannotate.c21 #include "map.h"
283 struct map *map = ms->map; local
285 .ms = { .map = map, },
311 target.addr = map__objdump_2mem(map, ops->target.addr);
314 map__rip_2objdump(target.ms.map, map__map_ip(target.ms.map, target.addr)) == ops->target.addr)
377 struct map *ma local
1702 struct map *map = args->ms.map; local
1935 struct map *map = args->ms.map; local
2164 struct map *map = args->ms.map; local
2635 struct map *map = ms->map; local
3028 annotation__calc_lines(struct annotation *notes, struct map *map, struct rb_root *root) argument
[all...]
/linux-master/tools/perf/ui/browsers/
H A Dannotate.c10 #include "../../util/map.h"
437 static int sym_title(struct symbol *sym, struct map *map, char *title, argument
441 map__dso(map)->long_name,
479 target_ms.map = ms->map;
483 sym_title(ms->sym, ms->map, title, sizeof(title), annotate_opts.percent_type);
665 sym_title(sym, ms->map, symbol_dso, sizeof(symbol_dso), annotate_opts.percent_type);
969 dso = map__dso(ms->map);
/linux-master/net/netfilter/
H A Dnft_set_pipapo.c106 * - map the bits we want to classify for the current field, for a given
147 * Note that as rules map to contiguous ranges of rules, given how netmask
182 * we need to map rules for 10.0.0.5 in lookup table #0 (rule #0) to 1024
189 * map to rules in next field: 0 1 1
206 * map to elements: 0x66 0x42
270 * map to rules in next field: 0 1 1
301 * map to elements: 0x66 0x42
347 * @map: Bitmap to be scanned for set bits
356 * For each bit set in map, select the bucket from mapping table with index
362 int pipapo_refill(unsigned long *map, unsigne argument
1155 pipapo_map(struct nft_pipapo_match *m, union nft_pipapo_map_bucket map[NFT_PIPAPO_MAX_FIELDS], struct nft_pipapo_elem *e) argument
[all...]
/linux-master/net/core/
H A Ddev.c1050 * id. It scans list of devices to build up a free map, then chooses
1139 * id. It scans list of devices to build up a free map, then chooses
2335 netdev_warn(dev, "Number of in use tx queues changed. Priority %i to tc mapping %i is no longer valid. Setting map to 0\n",
2372 struct xps_map *map = NULL; local
2375 map = xmap_dereference(dev_maps->attr_map[tci]);
2376 if (!map)
2379 for (pos = map->len; pos--;) {
2380 if (map->queues[pos] != index)
2383 if (map->len > 1) {
2384 map
2479 expand_xps_map(struct xps_map *map, int attr_index, u16 index, bool is_rxqs_map) argument
2525 struct xps_map *map; local
2547 struct xps_map *map, *new_map; local
3845 const struct netprio_map *map; local
4094 struct xps_map *map; local
4520 struct rps_map *map; local
[all...]
/linux-master/fs/fuse/
H A Dfuse_i.h1448 int fuse_backing_open(struct fuse_conn *fc, struct fuse_backing_map *map);
/linux-master/drivers/scsi/
H A Dscsi_lib.c1964 blk_mq_map_queues(&set->map[HCTX_TYPE_DEFAULT]);
3101 * scsi_kmap_atomic_sg - find and atomically map an sg-elemnt
3105 * @len: bytes to map, on return number of bytes mapped
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/
H A Den_main.c1432 sq->uar_map = mdev->mlx5e_res.hw_objs.bfreg.map;
1517 sq->uar_map = mdev->mlx5e_res.hw_objs.bfreg.map;
1602 sq->uar_map = mdev->mlx5e_res.hw_objs.bfreg.map;
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/
H A Dreporter_tx.c506 struct mlx5e_ptp_metadata_map *map = &ptpsq->metadata_map; local
518 txqsq->ch_ix, txqsq->sqn, ts_cq->mcq.cqn, map->undelivered_counter, map->capacity);
/linux-master/drivers/net/dsa/
H A Dmt7530.c2151 .map = mt7530_irq_map,
2192 .map = mt7988_irq_map,
/linux-master/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_kms.c368 struct ttm_bo_kmap_obj map; local
430 ret = ttm_bo_kmap(bo, kmap_offset, kmap_num, &map);
434 virtual = ttm_kmap_obj_virtual(&map, &is_iomem);
449 ttm_bo_kunmap(&map);
594 if (vps->cursor.bo->map.virtual)
626 if (!vbo || !vbo->map.virtual)
732 ret = ttm_bo_kmap(&vps->bo->tbo, 0, PFN_UP(size), &vps->bo->map);
H A Dvmwgfx_drv.c384 * The function will then map the first page and initialize a pending
394 struct ttm_bo_kmap_obj map; local
418 ret = ttm_bo_kmap(&vbo->tbo, 0, 1, &map);
420 result = ttm_kmap_obj_virtual(&map, &dummy);
424 ttm_bo_kunmap(&map);
430 DRM_ERROR("Dummy query buffer map failed.\n");
682 "DMA map mode: %s\n", names[dev_priv->map_mode]);
H A Dvmwgfx_bo.h68 * @map: Kmap object for semi-persistent mappings
84 struct ttm_bo_kmap_obj map; member in struct:vmw_bo
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
H A Dnv50.c52 void *map; member in struct:nv50_instobj
105 iowrite32_native(data, nv50_instobj(memory)->map + offset);
111 return ioread32_native(nv50_instobj(memory)->map + offset);
133 /* Attempt to allocate BAR2 address-space and map the object
152 emap = eobj->map;
153 eobj->map = NULL;
175 iobj->map = ioremap_wc(device->func->resource_addr(device, 3) +
177 if (!iobj->map) {
203 * unmapping it here, in case we need to map it again later.
205 if (likely(iobj->lru.next) && iobj->map) {
222 void __iomem *map = NULL; local
322 void *map; local
[all...]
/linux-master/drivers/gpu/drm/nouveau/
H A Dnouveau_bios.c1828 u8 map[16] = { }; local
1854 if (!map[i2c])
1855 map[i2c] = ++idx;
1856 dcbt->entry[i].connector = map[i2c] - 1;
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_cs.c1752 * @map: Placeholder to return found BO mapping
1760 struct amdgpu_bo_va_mapping **map)
1775 *map = mapping;
1758 amdgpu_cs_find_mapping(struct amdgpu_cs_parser *parser, uint64_t addr, struct amdgpu_bo **bo, struct amdgpu_bo_va_mapping **map) argument
/linux-master/fs/bcachefs/
H A Drecovery_passes.c85 static const u8 map[] = { local
92 for (unsigned i = 0; i < ARRAY_SIZE(map); i++)
94 ret |= BIT_ULL(map[i]);
/linux-master/drivers/vhost/
H A Dvhost.c39 "Maximum number of memory regions in memory map. (default: 64)");
1029 struct vhost_iotlb_map *map; local
1034 list_for_each_entry(map, &umem->list, link) {
1035 unsigned long a = map->addr;
1037 if (vhost_overflow(map->addr, map->size))
1041 if (!access_ok((void __user *)a, map->size))
1044 map->start,
1045 map->size))
1055 const struct vhost_iotlb_map *map local
1607 vhost_vq_meta_update(struct vhost_virtqueue *vq, const struct vhost_iotlb_map *map, int type) argument
1621 const struct vhost_iotlb_map *map; local
2335 const struct vhost_iotlb_map *map; local
[all...]
/linux-master/arch/x86/kernel/cpu/
H A Dtopology.c52 static struct { DECLARE_BITMAP(map, MAX_LOCAL_APIC); } apic_maps[TOPO_MAX_DOMAIN] __ro_after_init;
70 #define domain_weight(_dom) bitmap_weight(apic_maps[_dom].map, MAX_LOCAL_APIC)
161 unsigned long *map)
169 for (id = find_next_bit(map, end, lvlid); id < end; id = find_next_bit(map, end, ++id))
218 set_bit(topo_apicid(apic_id, dom), apic_maps[dom].map);
285 if (!test_bit(lvlid, apic_maps[at_level].map))
288 return bitmap_weight(apic_maps[at_level].map, lvlid);
318 if (!test_bit(lvlid, apic_maps[at_level].map))
324 return topo_unit_count(lvlid, at_level, apic_maps[which_units].map);
160 topo_unit_count(u32 lvlid, enum x86_topology_domains at_level, unsigned long *map) argument
[all...]
/linux-master/drivers/pci/
H A Dquirks.c2465 pci_warn(dev, "Can't map e100 registers\n");
3704 pci_warn(pdev, "Can't map ConnectX-4 initialization segment\n");
5993 void __iomem *map; local
6008 map = pci_iomap(pdev, 0, 0x23000);
6009 if (!map) {
6010 pci_err(pdev, "Can't map MMIO space\n");
6018 if (ioread32(map + 0x2240c) & 0x2) {
6025 iounmap(map);

Completed in 426 milliseconds

1234567891011>>