/linux-master/fs/btrfs/ |
H A D | extent_map.c | 35 tree->map = RB_ROOT_CACHED; 161 * Previous extent map found, return as in this case the caller does not 232 * We can't modify an extent map that is in the tree and that is being 235 * the tree and 1 for this task (which is unpinning the extent map or 260 rb_erase_cached(&merge->rb_node, &tree->map); 272 rb_erase_cached(&merge->rb_node, &tree->map); 310 "no extent map found for inode %llu (root %lld) when unpinning extent range [%llu, %llu), generation %llu", 319 "found extent map for inode %llu (root %lld) with unexpected start offset %llu when unpinning extent range [%llu, %llu), generation %llu", 376 * Add new extent map to the extent tree 378 * @tree: tree to insert new map i [all...] |
H A D | extent_io.c | 1090 * coupled with a single extent map, and each range can have 1091 * an extent map with a different offset value relative to the 1304 * and does the loop to map the page into extents and bios. 2364 * map records are removed 2373 struct extent_map_tree *map = &btrfs_inode->extent_tree; local 2383 write_lock(&map->lock); 2384 em = lookup_extent_mapping(map, start, len); 2386 write_unlock(&map->lock); 2391 write_unlock(&map->lock); 2430 remove_extent_mapping(map, e [all...] |
/linux-master/virt/kvm/ |
H A D | kvm_main.c | 2804 * or the caller allows to map a writable pfn for a read fault 2830 * MMU, we lookup the page using GUP to map it, so the guest VCPU can 2859 /* map read fault as writable if possible */ 2970 * @writable: whether it allows to map a writable host page for !@write_fault 2972 * The function will map a writable host page for these two cases: 3049 /* Do not map writable pfn in the readonly memslot. */ 3149 int kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map) argument 3155 if (!map) 3174 map->page = page; 3175 map 3183 kvm_vcpu_unmap(struct kvm_vcpu *vcpu, struct kvm_host_map *map, bool dirty) argument [all...] |
/linux-master/arch/x86/kvm/ |
H A D | x86.c | 10012 struct kvm_apic_map *map; local 10020 map = rcu_dereference(vcpu->kvm->arch.apic_map); 10022 if (likely(map) && dest_id <= map->max_apic_id && map->phys_map[dest_id]) 10023 target = map->phys_map[dest_id]->vcpu; 12704 * unless the memory map has changed due to process exit
|
H A D | lapic.c | 174 static inline bool kvm_apic_map_get_logical_dest(struct kvm_apic_map *map, argument 176 switch (map->logical_mode) { 178 /* Arbitrarily use the flat map so that @cluster isn't NULL. */ 179 *cluster = map->xapic_flat_map; 184 u32 max_apic_id = map->max_apic_id; 189 offset = array_index_nospec(offset, map->max_apic_id + 1); 190 *cluster = &map->phys_map[offset]; 199 *cluster = map->xapic_flat_map; 203 *cluster = map->xapic_cluster_map[(dest_id >> 4) & 0xf]; 216 struct kvm_apic_map *map local 835 __pv_send_ipi(unsigned long *ipi_bitmap, struct kvm_apic_map *map, struct kvm_lapic_irq *irq, u32 min) argument 859 struct kvm_apic_map *map; local 1119 kvm_apic_is_broadcast_dest(struct kvm *kvm, struct kvm_lapic **src, struct kvm_lapic_irq *irq, struct kvm_apic_map *map) argument 1145 kvm_apic_map_get_dest_lapic(struct kvm *kvm, struct kvm_lapic **src, struct kvm_lapic_irq *irq, struct kvm_apic_map *map, struct kvm_lapic ***dst, unsigned long *bitmap) argument 1214 struct kvm_apic_map *map; local 1265 struct kvm_apic_map *map; local 1402 struct kvm_apic_map *map; local [all...] |
/linux-master/arch/x86/kvm/svm/ |
H A D | svm.c | 2257 struct kvm_host_map map; local 2263 ret = kvm_vcpu_map(vcpu, gpa_to_gfn(svm->vmcb->save.rax), &map); 2270 vmcb12 = map.hva; 2282 kvm_vcpu_unmap(vcpu, &map, true); 4652 struct kvm_host_map map, map_save; local 4671 if (kvm_vcpu_map(vcpu, gpa_to_gfn(smram64->svm_guest_vmcb_gpa), &map)) 4694 vmcb12 = map.hva; 4707 kvm_vcpu_unmap(vcpu, &map, true);
|
/linux-master/tools/perf/util/ |
H A D | annotate.c | 21 #include "map.h" 283 struct map *map = ms->map; local 285 .ms = { .map = map, }, 311 target.addr = map__objdump_2mem(map, ops->target.addr); 314 map__rip_2objdump(target.ms.map, map__map_ip(target.ms.map, target.addr)) == ops->target.addr) 377 struct map *ma local 1702 struct map *map = args->ms.map; local 1935 struct map *map = args->ms.map; local 2164 struct map *map = args->ms.map; local 2635 struct map *map = ms->map; local 3028 annotation__calc_lines(struct annotation *notes, struct map *map, struct rb_root *root) argument [all...] |
/linux-master/tools/perf/ui/browsers/ |
H A D | annotate.c | 10 #include "../../util/map.h" 437 static int sym_title(struct symbol *sym, struct map *map, char *title, argument 441 map__dso(map)->long_name, 479 target_ms.map = ms->map; 483 sym_title(ms->sym, ms->map, title, sizeof(title), annotate_opts.percent_type); 665 sym_title(sym, ms->map, symbol_dso, sizeof(symbol_dso), annotate_opts.percent_type); 969 dso = map__dso(ms->map);
|
/linux-master/net/netfilter/ |
H A D | nft_set_pipapo.c | 106 * - map the bits we want to classify for the current field, for a given 147 * Note that as rules map to contiguous ranges of rules, given how netmask 182 * we need to map rules for 10.0.0.5 in lookup table #0 (rule #0) to 1024 189 * map to rules in next field: 0 1 1 206 * map to elements: 0x66 0x42 270 * map to rules in next field: 0 1 1 301 * map to elements: 0x66 0x42 347 * @map: Bitmap to be scanned for set bits 356 * For each bit set in map, select the bucket from mapping table with index 362 int pipapo_refill(unsigned long *map, unsigne argument 1155 pipapo_map(struct nft_pipapo_match *m, union nft_pipapo_map_bucket map[NFT_PIPAPO_MAX_FIELDS], struct nft_pipapo_elem *e) argument [all...] |
/linux-master/net/core/ |
H A D | dev.c | 1050 * id. It scans list of devices to build up a free map, then chooses 1139 * id. It scans list of devices to build up a free map, then chooses 2335 netdev_warn(dev, "Number of in use tx queues changed. Priority %i to tc mapping %i is no longer valid. Setting map to 0\n", 2372 struct xps_map *map = NULL; local 2375 map = xmap_dereference(dev_maps->attr_map[tci]); 2376 if (!map) 2379 for (pos = map->len; pos--;) { 2380 if (map->queues[pos] != index) 2383 if (map->len > 1) { 2384 map 2479 expand_xps_map(struct xps_map *map, int attr_index, u16 index, bool is_rxqs_map) argument 2525 struct xps_map *map; local 2547 struct xps_map *map, *new_map; local 3845 const struct netprio_map *map; local 4094 struct xps_map *map; local 4520 struct rps_map *map; local [all...] |
/linux-master/fs/fuse/ |
H A D | fuse_i.h | 1448 int fuse_backing_open(struct fuse_conn *fc, struct fuse_backing_map *map);
|
/linux-master/drivers/scsi/ |
H A D | scsi_lib.c | 1964 blk_mq_map_queues(&set->map[HCTX_TYPE_DEFAULT]); 3101 * scsi_kmap_atomic_sg - find and atomically map an sg-elemnt 3105 * @len: bytes to map, on return number of bytes mapped
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_main.c | 1432 sq->uar_map = mdev->mlx5e_res.hw_objs.bfreg.map; 1517 sq->uar_map = mdev->mlx5e_res.hw_objs.bfreg.map; 1602 sq->uar_map = mdev->mlx5e_res.hw_objs.bfreg.map;
|
/linux-master/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | reporter_tx.c | 506 struct mlx5e_ptp_metadata_map *map = &ptpsq->metadata_map; local 518 txqsq->ch_ix, txqsq->sqn, ts_cq->mcq.cqn, map->undelivered_counter, map->capacity);
|
/linux-master/drivers/net/dsa/ |
H A D | mt7530.c | 2151 .map = mt7530_irq_map, 2192 .map = mt7988_irq_map,
|
/linux-master/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_kms.c | 368 struct ttm_bo_kmap_obj map; local 430 ret = ttm_bo_kmap(bo, kmap_offset, kmap_num, &map); 434 virtual = ttm_kmap_obj_virtual(&map, &is_iomem); 449 ttm_bo_kunmap(&map); 594 if (vps->cursor.bo->map.virtual) 626 if (!vbo || !vbo->map.virtual) 732 ret = ttm_bo_kmap(&vps->bo->tbo, 0, PFN_UP(size), &vps->bo->map);
|
H A D | vmwgfx_drv.c | 384 * The function will then map the first page and initialize a pending 394 struct ttm_bo_kmap_obj map; local 418 ret = ttm_bo_kmap(&vbo->tbo, 0, 1, &map); 420 result = ttm_kmap_obj_virtual(&map, &dummy); 424 ttm_bo_kunmap(&map); 430 DRM_ERROR("Dummy query buffer map failed.\n"); 682 "DMA map mode: %s\n", names[dev_priv->map_mode]);
|
H A D | vmwgfx_bo.h | 68 * @map: Kmap object for semi-persistent mappings 84 struct ttm_bo_kmap_obj map; member in struct:vmw_bo
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | nv50.c | 52 void *map; member in struct:nv50_instobj 105 iowrite32_native(data, nv50_instobj(memory)->map + offset); 111 return ioread32_native(nv50_instobj(memory)->map + offset); 133 /* Attempt to allocate BAR2 address-space and map the object 152 emap = eobj->map; 153 eobj->map = NULL; 175 iobj->map = ioremap_wc(device->func->resource_addr(device, 3) + 177 if (!iobj->map) { 203 * unmapping it here, in case we need to map it again later. 205 if (likely(iobj->lru.next) && iobj->map) { 222 void __iomem *map = NULL; local 322 void *map; local [all...] |
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_bios.c | 1828 u8 map[16] = { }; local 1854 if (!map[i2c]) 1855 map[i2c] = ++idx; 1856 dcbt->entry[i].connector = map[i2c] - 1;
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_cs.c | 1752 * @map: Placeholder to return found BO mapping 1760 struct amdgpu_bo_va_mapping **map) 1775 *map = mapping; 1758 amdgpu_cs_find_mapping(struct amdgpu_cs_parser *parser, uint64_t addr, struct amdgpu_bo **bo, struct amdgpu_bo_va_mapping **map) argument
|
/linux-master/fs/bcachefs/ |
H A D | recovery_passes.c | 85 static const u8 map[] = { local 92 for (unsigned i = 0; i < ARRAY_SIZE(map); i++) 94 ret |= BIT_ULL(map[i]);
|
/linux-master/drivers/vhost/ |
H A D | vhost.c | 39 "Maximum number of memory regions in memory map. (default: 64)"); 1029 struct vhost_iotlb_map *map; local 1034 list_for_each_entry(map, &umem->list, link) { 1035 unsigned long a = map->addr; 1037 if (vhost_overflow(map->addr, map->size)) 1041 if (!access_ok((void __user *)a, map->size)) 1044 map->start, 1045 map->size)) 1055 const struct vhost_iotlb_map *map local 1607 vhost_vq_meta_update(struct vhost_virtqueue *vq, const struct vhost_iotlb_map *map, int type) argument 1621 const struct vhost_iotlb_map *map; local 2335 const struct vhost_iotlb_map *map; local [all...] |
/linux-master/arch/x86/kernel/cpu/ |
H A D | topology.c | 52 static struct { DECLARE_BITMAP(map, MAX_LOCAL_APIC); } apic_maps[TOPO_MAX_DOMAIN] __ro_after_init; 70 #define domain_weight(_dom) bitmap_weight(apic_maps[_dom].map, MAX_LOCAL_APIC) 161 unsigned long *map) 169 for (id = find_next_bit(map, end, lvlid); id < end; id = find_next_bit(map, end, ++id)) 218 set_bit(topo_apicid(apic_id, dom), apic_maps[dom].map); 285 if (!test_bit(lvlid, apic_maps[at_level].map)) 288 return bitmap_weight(apic_maps[at_level].map, lvlid); 318 if (!test_bit(lvlid, apic_maps[at_level].map)) 324 return topo_unit_count(lvlid, at_level, apic_maps[which_units].map); 160 topo_unit_count(u32 lvlid, enum x86_topology_domains at_level, unsigned long *map) argument [all...] |
/linux-master/drivers/pci/ |
H A D | quirks.c | 2465 pci_warn(dev, "Can't map e100 registers\n"); 3704 pci_warn(pdev, "Can't map ConnectX-4 initialization segment\n"); 5993 void __iomem *map; local 6008 map = pci_iomap(pdev, 0, 0x23000); 6009 if (!map) { 6010 pci_err(pdev, "Can't map MMIO space\n"); 6018 if (ioread32(map + 0x2240c) & 0x2) { 6025 iounmap(map);
|