/linux-master/net/mac80211/ |
H A D | mlme.c | 5931 new_active_links = sdata->u.mgd.ttlm_info.map & 5933 new_dormant_links = ~sdata->u.mgd.ttlm_info.map & 5974 sdata_info(sdata, "Invalid advertised T2L map direction\n"); 6002 * not advertise a TID-to-link mapping that does not map all TIDs to the 6011 ttlm_info->map = ieee80211_get_ttlm(map_size, pos); 6012 if (!ttlm_info->map) { 6014 "Invalid advertised T2L map for TID 0\n"); 6021 u16 map = ieee80211_get_ttlm(map_size, pos); local 6023 if (map != ttlm_info->map) { 6439 u16 new_active_links, new_dormant_links, new_suspended_links, map = 0; local 6492 __le16 map; local 6690 u16 map; local [all...] |
/linux-master/mm/ |
H A D | hugetlb.c | 548 hugetlb_resv_map_add(struct resv_map *map, struct list_head *rg, long from, argument 555 nrg = get_file_region_entry_from_cache(map, from, to); 556 record_hugetlb_cgroup_uncharge_info(cg, h, map, nrg); 558 coalesce_file_region(map, nrg); 691 * map. Regions will be taken from the cache to fill in this range. 699 * Return the number of new huge pages added to the map. This number is greater 754 * Examine the existing reserve map and determine how many 758 * map to add the specified range [f, t). region_chg does 760 * map. A number of new file_region structures is added to the cache as a 769 * reservation map fo 1163 set_vma_resv_map(struct vm_area_struct *vma, struct resv_map *map) argument [all...] |
/linux-master/include/linux/ |
H A D | mm.h | 179 * Default maximum number of active map areas, this limits the number of vmas 2541 * Whether we should manually check if we can map individual PTEs writable, 3956 void register_page_bootmem_memmap(unsigned long section_nr, struct page *map,
|
/linux-master/fs/ntfs3/ |
H A D | ntfs_fs.h | 482 bool are_bits_clear(const void *map, size_t bit, size_t nbits); 483 bool are_bits_set(const void *map, size_t bit, size_t nbits); 484 size_t get_set_bits_ex(const void *map, size_t bit, size_t nbits); 855 void ntfs_bitmap_set_le(void *map, unsigned int start, int len); 856 void ntfs_bitmap_clear_le(void *map, unsigned int start, int len);
|
/linux-master/drivers/net/ethernet/marvell/octeontx2/af/ |
H A D | rvu_npc.c | 81 u32 map; local 85 map = pkind->pfchan_map[i]; 86 if (((map >> 16) & 0x3F) == pf) 2060 * packet. And map this action to a counter to count dropped 2373 static u16 npc_mcam_find_zero_area(unsigned long *map, u16 size, u16 start, argument 2382 index = find_next_zero_bit(map, size, start); 2387 next = find_next_bit(map, end, index); 2404 static u16 npc_mcam_get_free_count(unsigned long *map, u16 start, u16 end) argument 2413 index = find_next_zero_bit(map, end, start); 2417 next = find_next_bit(map, en [all...] |
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt.c | 1356 struct bnxt_tpa_idx_map *map = rxr->rx_tpa_idx_map; local 1359 if (test_bit(idx, map->agg_idx_bmap)) 1360 idx = find_first_zero_bit(map->agg_idx_bmap, 1362 __set_bit(idx, map->agg_idx_bmap); 1363 map->agg_id_tbl[agg_id] = idx; 1369 struct bnxt_tpa_idx_map *map = rxr->rx_tpa_idx_map; local 1371 __clear_bit(idx, map->agg_idx_bmap); 1376 struct bnxt_tpa_idx_map *map = rxr->rx_tpa_idx_map; local 1378 return map->agg_id_tbl[agg_id]; 3275 struct bnxt_tpa_idx_map *map; local [all...] |
/linux-master/drivers/net/ethernet/broadcom/ |
H A D | b44.c | 629 * for the RX buffer, DMA map all of it, skb_reserve the 30 bytes, then 635 struct ring_info *src_map, *map; local 646 map = &bp->rx_buffers[dest_idx]; 685 map->skb = skb; 686 map->mapping = mapping; 766 dma_addr_t map = rp->mapping; local 770 dma_sync_single_for_cpu(bp->sdev->dma_dev, map, 804 dma_unmap_single(bp->sdev->dma_dev, map,
|
/linux-master/drivers/net/dsa/mv88e6xxx/ |
H A D | port.h | 391 int mv88e6xxx_port_set_vlan_map(struct mv88e6xxx_chip *chip, int port, u16 map); 462 int mv88e6xxx_port_set_map_da(struct mv88e6xxx_chip *chip, int port, bool map);
|
H A D | chip.c | 258 .map = mv88e6xxx_g1_irq_domain_map, 3001 * CPU and map in the PVT. First dst->last_switch elements are taken by 3068 "port %d failed to restore map-DA: %pe\n", 3534 /* Port based VLAN map: give each port the same default address 6739 u16 map = 0; local 6745 /* Build the map of all ports to distribute flows destined for 6750 map |= BIT(dsa_towards_port(ds, dp->ds->index, dp->index)); 6752 return mv88e6xxx_g2_trunk_mapping_write(chip, id, map);
|
/linux-master/drivers/md/ |
H A D | dm.c | 83 struct dm_table *map; member in struct:clone_info 416 struct dm_table *map; local 421 map = dm_get_live_table(md, srcu_idx); 422 if (!map || !dm_table_get_size(map)) 426 if (map->num_targets != 1) 429 ti = dm_table_get_target(map, 0); 682 * function to access the md->map field, and make sure they call 690 return srcu_dereference(md->map, &md->io_barrier); 712 return rcu_dereference(md->map); 1218 struct dm_table *map; variable in typeref:struct:dm_table 1760 init_clone_info(struct clone_info *ci, struct dm_io *io, struct dm_table *map, struct bio *bio, bool is_abnormal) argument 1780 dm_split_and_process_bio(struct mapped_device *md, struct dm_table *map, struct bio *bio) argument 1853 struct dm_table *map; local 2294 struct dm_table *map = rcu_dereference_protected(md->map, 1); local 2487 struct dm_table *map; local 2639 struct dm_table *live_map = NULL, *map = ERR_PTR(-EINVAL); local 2711 __dm_suspend(struct mapped_device *md, struct dm_table *map, unsigned int suspend_flags, unsigned int task_state, int dmf_suspended_flag) argument 2821 struct dm_table *map = NULL; local 2860 __dm_resume(struct mapped_device *md, struct dm_table *map) argument 2887 struct dm_table *map = NULL; local 2928 struct dm_table *map = NULL; local 2959 struct dm_table *map; local [all...] |
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_guc_ct.c | 183 struct iosys_map *map) 194 h2g->desc = *map; 197 h2g->cmds = IOSYS_MAP_INIT_OFFSET(map, CTB_DESC_SIZE * 2); 201 struct iosys_map *map) 212 g2h->desc = IOSYS_MAP_INIT_OFFSET(map, CTB_DESC_SIZE); 215 g2h->cmds = IOSYS_MAP_INIT_OFFSET(map, CTB_DESC_SIZE * 2 + 462 struct iosys_map map = IOSYS_MAP_INIT_OFFSET(&h2g->cmds, local 473 xe_map_memset(xe, &map, 0, 0, 507 xe_map_memcpy_to(xe, &map, 0, cmd, H2G_CT_HEADERS * sizeof(u32)); 508 xe_map_memcpy_to(xe, &map, H2G_CT_HEADER 182 guc_ct_ctb_h2g_init(struct xe_device *xe, struct guc_ctb *h2g, struct iosys_map *map) argument 200 guc_ct_ctb_g2h_init(struct xe_device *xe, struct guc_ctb *g2h, struct iosys_map *map) argument 1320 struct iosys_map map = local [all...] |
H A D | xe_huc.c | 139 static u32 huc_emit_pxp_auth_msg(struct xe_device *xe, struct iosys_map *map, argument 142 xe_map_memset(xe, map, wr_offset, 0, sizeof(struct pxp43_new_huc_auth_in)); 144 huc_auth_msg_wr(xe, map, wr_offset, header.api_version, PXP_APIVER(4, 3)); 145 huc_auth_msg_wr(xe, map, wr_offset, header.command_id, PXP43_CMDID_NEW_HUC_AUTH); 146 huc_auth_msg_wr(xe, map, wr_offset, header.status, 0); 147 huc_auth_msg_wr(xe, map, wr_offset, header.buffer_len, 149 huc_auth_msg_wr(xe, map, wr_offset, huc_base_address, huc_offset); 150 huc_auth_msg_wr(xe, map, wr_offset, huc_size, huc_size);
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_gem_atomic_helper.c | 83 * // access shadow buffer via shadow_plane_state->map 87 * struct &drm_shadow_plane_state.map. The mappings are valid while the state 109 * // access shadow buffer via shadow_plane_state->map 349 * space and stores them in struct &drm_shadow_plane_state.map. The first data 365 return drm_gem_fb_vmap(fb, shadow_plane_state->map, shadow_plane_state->data); 387 drm_gem_fb_vunmap(fb, shadow_plane_state->map);
|
/linux-master/drivers/gpio/ |
H A D | gpio-tegra186.c | 980 irq->map = devm_kcalloc(&pdev->dev, gpio->gpio.ngpio, 981 sizeof(*irq->map), GFP_KERNEL); 982 if (!irq->map) 989 irq->map[offset + j] = irq->parents[port->bank];
|
/linux-master/arch/arc/kernel/ |
H A D | intc-arcv2.c | 156 .map = arcv2_irq_map,
|
/linux-master/fs/btrfs/tests/ |
H A D | extent-map-tests.c | 20 while (!RB_EMPTY_ROOT(&em_tree->map.rb_root)) { 21 node = rb_first_cached(&em_tree->map); 43 * Suppose that no extent map has been loaded into memory yet, there is a file 125 test_err("case1 [%llu %llu]: no extent map returned", 218 test_err("case2 [0 1K]: no extent map returned"); 285 test_err("case3 [%llu %llu): no extent map returned", 312 * Suppose that no extent map has been loaded into memory yet. 409 test_err("case4 [%llu %llu): no extent map returned", 431 * Suppose that no extent map has been loaded into memory yet. 489 test_err("cannot add extent map [ 888 struct btrfs_chunk_map *map; local [all...] |
/linux-master/fs/btrfs/ |
H A D | scrub.c | 1292 struct btrfs_chunk_map *map, u64 *offset, 1298 const int data_stripes = nr_data_stripes(map); 1300 last_offset = (physical - map->stripes[num].physical) * data_stripes; 1315 rot = stripe_nr % map->num_stripes; 1318 stripe_index = rot % map->num_stripes; 1926 struct btrfs_chunk_map *map, 1938 const int data_stripes = nr_data_stripes(map); 1963 stripe_index = (i + rot) % map->num_stripes; 1964 physical = map->stripes[stripe_index].physical + 1970 map 1291 get_raid56_logic_offset(u64 physical, int num, struct btrfs_chunk_map *map, u64 *offset, u64 *stripe_start) argument 1923 scrub_raid56_parity_stripe(struct scrub_ctx *sctx, struct btrfs_device *scrub_dev, struct btrfs_block_group *bg, struct btrfs_chunk_map *map, u64 full_stripe_start) argument 2093 scrub_simple_mirror(struct scrub_ctx *sctx, struct btrfs_block_group *bg, struct btrfs_chunk_map *map, u64 logical_start, u64 logical_length, struct btrfs_device *device, u64 physical, int mirror_num) argument 2156 simple_stripe_full_stripe_len(const struct btrfs_chunk_map *map) argument 2165 simple_stripe_get_logical(struct btrfs_chunk_map *map, struct btrfs_block_group *bg, int stripe_index) argument 2182 simple_stripe_mirror_num(struct btrfs_chunk_map *map, int stripe_index) argument 2192 scrub_simple_stripe(struct scrub_ctx *sctx, struct btrfs_block_group *bg, struct btrfs_chunk_map *map, struct btrfs_device *device, int stripe_index) argument 2225 scrub_stripe(struct scrub_ctx *sctx, struct btrfs_block_group *bg, struct btrfs_chunk_map *map, struct btrfs_device *scrub_dev, int stripe_index) argument 2400 struct btrfs_chunk_map *map; local [all...] |
H A D | inode.c | 2914 * The new extent is pinned in the extent map, and we don't want 3222 * we don't mess with the extent map tree in the NOCOW case, but 7268 * than one extent map, if at this point has already 7269 * created an ordered extent for a previous extent map 7271 * concurrent write against that previous extent map's 7353 static int btrfs_get_blocks_direct_write(struct extent_map **map, argument 7361 struct extent_map *em = *map; 7406 /* Our caller expects us to free the input extent map. */ 7408 *map = NULL; 7423 *map 10693 struct btrfs_chunk_map *map = NULL; local [all...] |
H A D | extent_map.c | 35 tree->map = RB_ROOT_CACHED; 161 * Previous extent map found, return as in this case the caller does not 232 * We can't modify an extent map that is in the tree and that is being 235 * the tree and 1 for this task (which is unpinning the extent map or 260 rb_erase_cached(&merge->rb_node, &tree->map); 272 rb_erase_cached(&merge->rb_node, &tree->map); 310 "no extent map found for inode %llu (root %lld) when unpinning extent range [%llu, %llu), generation %llu", 319 "found extent map for inode %llu (root %lld) with unexpected start offset %llu when unpinning extent range [%llu, %llu), generation %llu", 376 * Add new extent map to the extent tree 378 * @tree: tree to insert new map i [all...] |
/linux-master/virt/kvm/ |
H A D | kvm_main.c | 2804 * or the caller allows to map a writable pfn for a read fault 2830 * MMU, we lookup the page using GUP to map it, so the guest VCPU can 2859 /* map read fault as writable if possible */ 2970 * @writable: whether it allows to map a writable host page for !@write_fault 2972 * The function will map a writable host page for these two cases: 3049 /* Do not map writable pfn in the readonly memslot. */ 3149 int kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map) argument 3155 if (!map) 3174 map->page = page; 3175 map 3183 kvm_vcpu_unmap(struct kvm_vcpu *vcpu, struct kvm_host_map *map, bool dirty) argument [all...] |
/linux-master/arch/x86/kvm/ |
H A D | x86.c | 10012 struct kvm_apic_map *map; local 10020 map = rcu_dereference(vcpu->kvm->arch.apic_map); 10022 if (likely(map) && dest_id <= map->max_apic_id && map->phys_map[dest_id]) 10023 target = map->phys_map[dest_id]->vcpu; 12704 * unless the memory map has changed due to process exit
|
H A D | lapic.c | 174 static inline bool kvm_apic_map_get_logical_dest(struct kvm_apic_map *map, argument 176 switch (map->logical_mode) { 178 /* Arbitrarily use the flat map so that @cluster isn't NULL. */ 179 *cluster = map->xapic_flat_map; 184 u32 max_apic_id = map->max_apic_id; 189 offset = array_index_nospec(offset, map->max_apic_id + 1); 190 *cluster = &map->phys_map[offset]; 199 *cluster = map->xapic_flat_map; 203 *cluster = map->xapic_cluster_map[(dest_id >> 4) & 0xf]; 216 struct kvm_apic_map *map local 835 __pv_send_ipi(unsigned long *ipi_bitmap, struct kvm_apic_map *map, struct kvm_lapic_irq *irq, u32 min) argument 859 struct kvm_apic_map *map; local 1119 kvm_apic_is_broadcast_dest(struct kvm *kvm, struct kvm_lapic **src, struct kvm_lapic_irq *irq, struct kvm_apic_map *map) argument 1145 kvm_apic_map_get_dest_lapic(struct kvm *kvm, struct kvm_lapic **src, struct kvm_lapic_irq *irq, struct kvm_apic_map *map, struct kvm_lapic ***dst, unsigned long *bitmap) argument 1214 struct kvm_apic_map *map; local 1265 struct kvm_apic_map *map; local 1402 struct kvm_apic_map *map; local [all...] |
/linux-master/arch/x86/kvm/svm/ |
H A D | svm.c | 2257 struct kvm_host_map map; local 2263 ret = kvm_vcpu_map(vcpu, gpa_to_gfn(svm->vmcb->save.rax), &map); 2270 vmcb12 = map.hva; 2282 kvm_vcpu_unmap(vcpu, &map, true); 4652 struct kvm_host_map map, map_save; local 4671 if (kvm_vcpu_map(vcpu, gpa_to_gfn(smram64->svm_guest_vmcb_gpa), &map)) 4694 vmcb12 = map.hva; 4707 kvm_vcpu_unmap(vcpu, &map, true);
|
/linux-master/ |
H A D | Makefile | 1003 KBUILD_CPPFLAGS += $(call cc-option,-fmacro-prefix-map=$(srctree)/=) 1073 # INSTALL_PATH specifies where to place the updated kernel and system map 1572 @echo ' (requires a recent binutils and recent build (System.map))'
|
/linux-master/tools/perf/util/ |
H A D | annotate.c | 21 #include "map.h" 283 struct map *map = ms->map; local 285 .ms = { .map = map, }, 311 target.addr = map__objdump_2mem(map, ops->target.addr); 314 map__rip_2objdump(target.ms.map, map__map_ip(target.ms.map, target.addr)) == ops->target.addr) 377 struct map *ma local 1702 struct map *map = args->ms.map; local 1935 struct map *map = args->ms.map; local 2164 struct map *map = args->ms.map; local 2635 struct map *map = ms->map; local 3028 annotation__calc_lines(struct annotation *notes, struct map *map, struct rb_root *root) argument [all...] |