Searched refs:map (Results 1 - 25 of 2546) sorted by last modified time

1234567891011>>

/linux-master/net/mac80211/
H A Dmlme.c5931 new_active_links = sdata->u.mgd.ttlm_info.map &
5933 new_dormant_links = ~sdata->u.mgd.ttlm_info.map &
5974 sdata_info(sdata, "Invalid advertised T2L map direction\n");
6002 * not advertise a TID-to-link mapping that does not map all TIDs to the
6011 ttlm_info->map = ieee80211_get_ttlm(map_size, pos);
6012 if (!ttlm_info->map) {
6014 "Invalid advertised T2L map for TID 0\n");
6021 u16 map = ieee80211_get_ttlm(map_size, pos); local
6023 if (map != ttlm_info->map) {
6439 u16 new_active_links, new_dormant_links, new_suspended_links, map = 0; local
6492 __le16 map; local
6690 u16 map; local
[all...]
/linux-master/mm/
H A Dhugetlb.c548 hugetlb_resv_map_add(struct resv_map *map, struct list_head *rg, long from, argument
555 nrg = get_file_region_entry_from_cache(map, from, to);
556 record_hugetlb_cgroup_uncharge_info(cg, h, map, nrg);
558 coalesce_file_region(map, nrg);
691 * map. Regions will be taken from the cache to fill in this range.
699 * Return the number of new huge pages added to the map. This number is greater
754 * Examine the existing reserve map and determine how many
758 * map to add the specified range [f, t). region_chg does
760 * map. A number of new file_region structures is added to the cache as a
769 * reservation map fo
1163 set_vma_resv_map(struct vm_area_struct *vma, struct resv_map *map) argument
[all...]
/linux-master/include/linux/
H A Dmm.h179 * Default maximum number of active map areas, this limits the number of vmas
2541 * Whether we should manually check if we can map individual PTEs writable,
3956 void register_page_bootmem_memmap(unsigned long section_nr, struct page *map,
/linux-master/fs/ntfs3/
H A Dntfs_fs.h482 bool are_bits_clear(const void *map, size_t bit, size_t nbits);
483 bool are_bits_set(const void *map, size_t bit, size_t nbits);
484 size_t get_set_bits_ex(const void *map, size_t bit, size_t nbits);
855 void ntfs_bitmap_set_le(void *map, unsigned int start, int len);
856 void ntfs_bitmap_clear_le(void *map, unsigned int start, int len);
/linux-master/drivers/net/ethernet/marvell/octeontx2/af/
H A Drvu_npc.c81 u32 map; local
85 map = pkind->pfchan_map[i];
86 if (((map >> 16) & 0x3F) == pf)
2060 * packet. And map this action to a counter to count dropped
2373 static u16 npc_mcam_find_zero_area(unsigned long *map, u16 size, u16 start, argument
2382 index = find_next_zero_bit(map, size, start);
2387 next = find_next_bit(map, end, index);
2404 static u16 npc_mcam_get_free_count(unsigned long *map, u16 start, u16 end) argument
2413 index = find_next_zero_bit(map, end, start);
2417 next = find_next_bit(map, en
[all...]
/linux-master/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt.c1356 struct bnxt_tpa_idx_map *map = rxr->rx_tpa_idx_map; local
1359 if (test_bit(idx, map->agg_idx_bmap))
1360 idx = find_first_zero_bit(map->agg_idx_bmap,
1362 __set_bit(idx, map->agg_idx_bmap);
1363 map->agg_id_tbl[agg_id] = idx;
1369 struct bnxt_tpa_idx_map *map = rxr->rx_tpa_idx_map; local
1371 __clear_bit(idx, map->agg_idx_bmap);
1376 struct bnxt_tpa_idx_map *map = rxr->rx_tpa_idx_map; local
1378 return map->agg_id_tbl[agg_id];
3275 struct bnxt_tpa_idx_map *map; local
[all...]
/linux-master/drivers/net/ethernet/broadcom/
H A Db44.c629 * for the RX buffer, DMA map all of it, skb_reserve the 30 bytes, then
635 struct ring_info *src_map, *map; local
646 map = &bp->rx_buffers[dest_idx];
685 map->skb = skb;
686 map->mapping = mapping;
766 dma_addr_t map = rp->mapping; local
770 dma_sync_single_for_cpu(bp->sdev->dma_dev, map,
804 dma_unmap_single(bp->sdev->dma_dev, map,
/linux-master/drivers/net/dsa/mv88e6xxx/
H A Dport.h391 int mv88e6xxx_port_set_vlan_map(struct mv88e6xxx_chip *chip, int port, u16 map);
462 int mv88e6xxx_port_set_map_da(struct mv88e6xxx_chip *chip, int port, bool map);
H A Dchip.c258 .map = mv88e6xxx_g1_irq_domain_map,
3001 * CPU and map in the PVT. First dst->last_switch elements are taken by
3068 "port %d failed to restore map-DA: %pe\n",
3534 /* Port based VLAN map: give each port the same default address
6739 u16 map = 0; local
6745 /* Build the map of all ports to distribute flows destined for
6750 map |= BIT(dsa_towards_port(ds, dp->ds->index, dp->index));
6752 return mv88e6xxx_g2_trunk_mapping_write(chip, id, map);
/linux-master/drivers/md/
H A Ddm.c83 struct dm_table *map; member in struct:clone_info
416 struct dm_table *map; local
421 map = dm_get_live_table(md, srcu_idx);
422 if (!map || !dm_table_get_size(map))
426 if (map->num_targets != 1)
429 ti = dm_table_get_target(map, 0);
682 * function to access the md->map field, and make sure they call
690 return srcu_dereference(md->map, &md->io_barrier);
712 return rcu_dereference(md->map);
1218 struct dm_table *map; variable in typeref:struct:dm_table
1760 init_clone_info(struct clone_info *ci, struct dm_io *io, struct dm_table *map, struct bio *bio, bool is_abnormal) argument
1780 dm_split_and_process_bio(struct mapped_device *md, struct dm_table *map, struct bio *bio) argument
1853 struct dm_table *map; local
2294 struct dm_table *map = rcu_dereference_protected(md->map, 1); local
2487 struct dm_table *map; local
2639 struct dm_table *live_map = NULL, *map = ERR_PTR(-EINVAL); local
2711 __dm_suspend(struct mapped_device *md, struct dm_table *map, unsigned int suspend_flags, unsigned int task_state, int dmf_suspended_flag) argument
2821 struct dm_table *map = NULL; local
2860 __dm_resume(struct mapped_device *md, struct dm_table *map) argument
2887 struct dm_table *map = NULL; local
2928 struct dm_table *map = NULL; local
2959 struct dm_table *map; local
[all...]
/linux-master/drivers/gpu/drm/xe/
H A Dxe_guc_ct.c183 struct iosys_map *map)
194 h2g->desc = *map;
197 h2g->cmds = IOSYS_MAP_INIT_OFFSET(map, CTB_DESC_SIZE * 2);
201 struct iosys_map *map)
212 g2h->desc = IOSYS_MAP_INIT_OFFSET(map, CTB_DESC_SIZE);
215 g2h->cmds = IOSYS_MAP_INIT_OFFSET(map, CTB_DESC_SIZE * 2 +
462 struct iosys_map map = IOSYS_MAP_INIT_OFFSET(&h2g->cmds, local
473 xe_map_memset(xe, &map, 0, 0,
507 xe_map_memcpy_to(xe, &map, 0, cmd, H2G_CT_HEADERS * sizeof(u32));
508 xe_map_memcpy_to(xe, &map, H2G_CT_HEADER
182 guc_ct_ctb_h2g_init(struct xe_device *xe, struct guc_ctb *h2g, struct iosys_map *map) argument
200 guc_ct_ctb_g2h_init(struct xe_device *xe, struct guc_ctb *g2h, struct iosys_map *map) argument
1320 struct iosys_map map = local
[all...]
H A Dxe_huc.c139 static u32 huc_emit_pxp_auth_msg(struct xe_device *xe, struct iosys_map *map, argument
142 xe_map_memset(xe, map, wr_offset, 0, sizeof(struct pxp43_new_huc_auth_in));
144 huc_auth_msg_wr(xe, map, wr_offset, header.api_version, PXP_APIVER(4, 3));
145 huc_auth_msg_wr(xe, map, wr_offset, header.command_id, PXP43_CMDID_NEW_HUC_AUTH);
146 huc_auth_msg_wr(xe, map, wr_offset, header.status, 0);
147 huc_auth_msg_wr(xe, map, wr_offset, header.buffer_len,
149 huc_auth_msg_wr(xe, map, wr_offset, huc_base_address, huc_offset);
150 huc_auth_msg_wr(xe, map, wr_offset, huc_size, huc_size);
/linux-master/drivers/gpu/drm/
H A Ddrm_gem_atomic_helper.c83 * // access shadow buffer via shadow_plane_state->map
87 * struct &drm_shadow_plane_state.map. The mappings are valid while the state
109 * // access shadow buffer via shadow_plane_state->map
349 * space and stores them in struct &drm_shadow_plane_state.map. The first data
365 return drm_gem_fb_vmap(fb, shadow_plane_state->map, shadow_plane_state->data);
387 drm_gem_fb_vunmap(fb, shadow_plane_state->map);
/linux-master/drivers/gpio/
H A Dgpio-tegra186.c980 irq->map = devm_kcalloc(&pdev->dev, gpio->gpio.ngpio,
981 sizeof(*irq->map), GFP_KERNEL);
982 if (!irq->map)
989 irq->map[offset + j] = irq->parents[port->bank];
/linux-master/arch/arc/kernel/
H A Dintc-arcv2.c156 .map = arcv2_irq_map,
/linux-master/fs/btrfs/tests/
H A Dextent-map-tests.c20 while (!RB_EMPTY_ROOT(&em_tree->map.rb_root)) {
21 node = rb_first_cached(&em_tree->map);
43 * Suppose that no extent map has been loaded into memory yet, there is a file
125 test_err("case1 [%llu %llu]: no extent map returned",
218 test_err("case2 [0 1K]: no extent map returned");
285 test_err("case3 [%llu %llu): no extent map returned",
312 * Suppose that no extent map has been loaded into memory yet.
409 test_err("case4 [%llu %llu): no extent map returned",
431 * Suppose that no extent map has been loaded into memory yet.
489 test_err("cannot add extent map [
888 struct btrfs_chunk_map *map; local
[all...]
/linux-master/fs/btrfs/
H A Dscrub.c1292 struct btrfs_chunk_map *map, u64 *offset,
1298 const int data_stripes = nr_data_stripes(map);
1300 last_offset = (physical - map->stripes[num].physical) * data_stripes;
1315 rot = stripe_nr % map->num_stripes;
1318 stripe_index = rot % map->num_stripes;
1926 struct btrfs_chunk_map *map,
1938 const int data_stripes = nr_data_stripes(map);
1963 stripe_index = (i + rot) % map->num_stripes;
1964 physical = map->stripes[stripe_index].physical +
1970 map
1291 get_raid56_logic_offset(u64 physical, int num, struct btrfs_chunk_map *map, u64 *offset, u64 *stripe_start) argument
1923 scrub_raid56_parity_stripe(struct scrub_ctx *sctx, struct btrfs_device *scrub_dev, struct btrfs_block_group *bg, struct btrfs_chunk_map *map, u64 full_stripe_start) argument
2093 scrub_simple_mirror(struct scrub_ctx *sctx, struct btrfs_block_group *bg, struct btrfs_chunk_map *map, u64 logical_start, u64 logical_length, struct btrfs_device *device, u64 physical, int mirror_num) argument
2156 simple_stripe_full_stripe_len(const struct btrfs_chunk_map *map) argument
2165 simple_stripe_get_logical(struct btrfs_chunk_map *map, struct btrfs_block_group *bg, int stripe_index) argument
2182 simple_stripe_mirror_num(struct btrfs_chunk_map *map, int stripe_index) argument
2192 scrub_simple_stripe(struct scrub_ctx *sctx, struct btrfs_block_group *bg, struct btrfs_chunk_map *map, struct btrfs_device *device, int stripe_index) argument
2225 scrub_stripe(struct scrub_ctx *sctx, struct btrfs_block_group *bg, struct btrfs_chunk_map *map, struct btrfs_device *scrub_dev, int stripe_index) argument
2400 struct btrfs_chunk_map *map; local
[all...]
H A Dinode.c2914 * The new extent is pinned in the extent map, and we don't want
3222 * we don't mess with the extent map tree in the NOCOW case, but
7268 * than one extent map, if at this point has already
7269 * created an ordered extent for a previous extent map
7271 * concurrent write against that previous extent map's
7353 static int btrfs_get_blocks_direct_write(struct extent_map **map, argument
7361 struct extent_map *em = *map;
7406 /* Our caller expects us to free the input extent map. */
7408 *map = NULL;
7423 *map
10693 struct btrfs_chunk_map *map = NULL; local
[all...]
H A Dextent_map.c35 tree->map = RB_ROOT_CACHED;
161 * Previous extent map found, return as in this case the caller does not
232 * We can't modify an extent map that is in the tree and that is being
235 * the tree and 1 for this task (which is unpinning the extent map or
260 rb_erase_cached(&merge->rb_node, &tree->map);
272 rb_erase_cached(&merge->rb_node, &tree->map);
310 "no extent map found for inode %llu (root %lld) when unpinning extent range [%llu, %llu), generation %llu",
319 "found extent map for inode %llu (root %lld) with unexpected start offset %llu when unpinning extent range [%llu, %llu), generation %llu",
376 * Add new extent map to the extent tree
378 * @tree: tree to insert new map i
[all...]
/linux-master/virt/kvm/
H A Dkvm_main.c2804 * or the caller allows to map a writable pfn for a read fault
2830 * MMU, we lookup the page using GUP to map it, so the guest VCPU can
2859 /* map read fault as writable if possible */
2970 * @writable: whether it allows to map a writable host page for !@write_fault
2972 * The function will map a writable host page for these two cases:
3049 /* Do not map writable pfn in the readonly memslot. */
3149 int kvm_vcpu_map(struct kvm_vcpu *vcpu, gfn_t gfn, struct kvm_host_map *map) argument
3155 if (!map)
3174 map->page = page;
3175 map
3183 kvm_vcpu_unmap(struct kvm_vcpu *vcpu, struct kvm_host_map *map, bool dirty) argument
[all...]
/linux-master/arch/x86/kvm/
H A Dx86.c10012 struct kvm_apic_map *map; local
10020 map = rcu_dereference(vcpu->kvm->arch.apic_map);
10022 if (likely(map) && dest_id <= map->max_apic_id && map->phys_map[dest_id])
10023 target = map->phys_map[dest_id]->vcpu;
12704 * unless the memory map has changed due to process exit
H A Dlapic.c174 static inline bool kvm_apic_map_get_logical_dest(struct kvm_apic_map *map, argument
176 switch (map->logical_mode) {
178 /* Arbitrarily use the flat map so that @cluster isn't NULL. */
179 *cluster = map->xapic_flat_map;
184 u32 max_apic_id = map->max_apic_id;
189 offset = array_index_nospec(offset, map->max_apic_id + 1);
190 *cluster = &map->phys_map[offset];
199 *cluster = map->xapic_flat_map;
203 *cluster = map->xapic_cluster_map[(dest_id >> 4) & 0xf];
216 struct kvm_apic_map *map local
835 __pv_send_ipi(unsigned long *ipi_bitmap, struct kvm_apic_map *map, struct kvm_lapic_irq *irq, u32 min) argument
859 struct kvm_apic_map *map; local
1119 kvm_apic_is_broadcast_dest(struct kvm *kvm, struct kvm_lapic **src, struct kvm_lapic_irq *irq, struct kvm_apic_map *map) argument
1145 kvm_apic_map_get_dest_lapic(struct kvm *kvm, struct kvm_lapic **src, struct kvm_lapic_irq *irq, struct kvm_apic_map *map, struct kvm_lapic ***dst, unsigned long *bitmap) argument
1214 struct kvm_apic_map *map; local
1265 struct kvm_apic_map *map; local
1402 struct kvm_apic_map *map; local
[all...]
/linux-master/arch/x86/kvm/svm/
H A Dsvm.c2257 struct kvm_host_map map; local
2263 ret = kvm_vcpu_map(vcpu, gpa_to_gfn(svm->vmcb->save.rax), &map);
2270 vmcb12 = map.hva;
2282 kvm_vcpu_unmap(vcpu, &map, true);
4652 struct kvm_host_map map, map_save; local
4671 if (kvm_vcpu_map(vcpu, gpa_to_gfn(smram64->svm_guest_vmcb_gpa), &map))
4694 vmcb12 = map.hva;
4707 kvm_vcpu_unmap(vcpu, &map, true);
/linux-master/
H A DMakefile1003 KBUILD_CPPFLAGS += $(call cc-option,-fmacro-prefix-map=$(srctree)/=)
1073 # INSTALL_PATH specifies where to place the updated kernel and system map
1572 @echo ' (requires a recent binutils and recent build (System.map))'
/linux-master/tools/perf/util/
H A Dannotate.c21 #include "map.h"
283 struct map *map = ms->map; local
285 .ms = { .map = map, },
311 target.addr = map__objdump_2mem(map, ops->target.addr);
314 map__rip_2objdump(target.ms.map, map__map_ip(target.ms.map, target.addr)) == ops->target.addr)
377 struct map *ma local
1702 struct map *map = args->ms.map; local
1935 struct map *map = args->ms.map; local
2164 struct map *map = args->ms.map; local
2635 struct map *map = ms->map; local
3028 annotation__calc_lines(struct annotation *notes, struct map *map, struct rb_root *root) argument
[all...]

Completed in 633 milliseconds

1234567891011>>