/linux-master/drivers/net/ethernet/mellanox/mlx5/core/diag/ |
H A D | en_rep_tracepoint.h | 24 TP_fast_assign(const struct mlx5e_neigh *mn = &nhe->m_neigh; 34 if (mn->family == AF_INET) { 35 *p32 = mn->dst_ip.v4; 37 } else if (mn->family == AF_INET6) { 38 *pin6 = mn->dst_ip.v6;
|
H A D | en_tc_tracepoint.h | 85 TP_fast_assign(const struct mlx5e_neigh *mn = &nhe->m_neigh; 94 if (mn->family == AF_INET) { 95 *p32 = mn->dst_ip.v4; 97 } else if (mn->family == AF_INET6) { 98 *pin6 = mn->dst_ip.v6;
|
/linux-master/drivers/clk/qcom/ |
H A D | clk-rcg.c | 106 static u32 md_to_m(struct mn *mn, u32 md) argument 108 md >>= mn->m_val_shift; 109 md &= BIT(mn->width) - 1; 132 static u32 mn_to_md(struct mn *mn, u32 m, u32 n, u32 md) argument 136 mask_w = BIT(mn->width) - 1; 137 mask = (mask_w << mn->m_val_shift) | mask_w; 141 m <<= mn->m_val_shift; 149 static u32 ns_m_to_n(struct mn *m argument 156 reg_to_mnctr_mode(struct mn *mn, u32 val) argument 163 mn_to_ns(struct mn *mn, u32 m, u32 n, u32 ns) argument 182 mn_to_reg(struct mn *mn, u32 m, u32 n, u32 val) argument 202 struct mn *mn; local 346 struct mn *mn = &rcg->mn; local 372 struct mn *mn; local 477 struct mn *mn = &rcg->mn; local [all...] |
H A D | clk-rcg.h | 21 * struct mn - M/N:D counter 22 * @mnctr_en_bit: bit to enable mn counter 23 * @mnctr_reset_bit: bit to assert mn counter reset 24 * @mnctr_mode_shift: lowest bit of mn counter mode field 30 struct mn { struct 68 * @mn: mn counter 79 struct mn mn; member in struct:clk_rcg 105 * @mn 118 struct mn mn[2]; member in struct:clk_dyn_rcg [all...] |
H A D | gcc-mdm9615.c | 188 .mn = { 239 .mn = { 290 .mn = { 341 .mn = { 392 .mn = { 455 .mn = { 506 .mn = { 557 .mn = { 608 .mn = { 659 .mn [all...] |
H A D | gcc-msm8660.c | 106 .mn = { 157 .mn = { 208 .mn = { 259 .mn = { 310 .mn = { 361 .mn = { 412 .mn = { 463 .mn = { 514 .mn = { 565 .mn [all...] |
H A D | gcc-ipq806x.c | 414 .mn = { 465 .mn = { 516 .mn = { 567 .mn = { 618 .mn = { 669 .mn = { 733 .mn = { 784 .mn = { 835 .mn = { 886 .mn [all...] |
H A D | lcc-ipq806x.c | 113 .mn = { 230 .mn = { 314 .mn = { 371 .mn = {
|
H A D | mmcc-msm8960.c | 176 .mn = { 227 .mn = { 278 .mn = { 336 .mn = { 404 .mn = { 472 .mn = { 727 .mn = { 833 .mn[0] = { 841 .mn[1] = { 895 .mn[ [all...] |
H A D | gcc-msm8960.c | 352 .mn = { 403 .mn = { 454 .mn = { 505 .mn = { 556 .mn = { 607 .mn = { 658 .mn = { 709 .mn = { 760 .mn = { 811 .mn [all...] |
H A D | lcc-msm8960.c | 96 .mn = { \ 258 .mn = { 328 .mn = {
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
H A D | ram.c | 35 struct nvkm_mm_node *mn; member in struct:nvkm_vram 52 .mem = vram->mn, 61 return (u64)nvkm_mm_size(nvkm_vram(memory)->mn) << NVKM_RAM_MM_SHIFT; 68 if (!nvkm_mm_contiguous(vram->mn)) 70 return (u64)nvkm_mm_addr(vram->mn) << NVKM_RAM_MM_SHIFT; 89 struct nvkm_mm_node *next = vram->mn; 101 kfree(vram->mn); 138 vram->mn = kzalloc(sizeof(*vram->mn), GFP_KERNEL); 139 if (!vram->mn) [all...] |
/linux-master/drivers/misc/sgi-gru/ |
H A D | grutlbpurge.c | 209 static int gru_invalidate_range_start(struct mmu_notifier *mn, argument 212 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, 224 static void gru_invalidate_range_end(struct mmu_notifier *mn, argument 227 struct gru_mm_struct *gms = container_of(mn, struct gru_mm_struct, 252 static void gru_free_notifier(struct mmu_notifier *mn) argument 254 kfree(container_of(mn, struct gru_mm_struct, ms_notifier)); 267 struct mmu_notifier *mn; local 269 mn = mmu_notifier_get_locked(&gru_mmuops, current->mm); 270 if (IS_ERR(mn)) 271 return ERR_CAST(mn); [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/core/ |
H A D | memory.c | 38 nvkm_mm_free(&fb->tags.mm, &tags->mn); 61 if (tags->mn && tags->mn->length != nr) { 77 if (!nvkm_mm_head(&fb->tags.mm, 0, 1, nr, nr, 1, &tags->mn)) { 79 clr(device, tags->mn->offset, tags->mn->length); 90 tags->mn = NULL;
|
/linux-master/drivers/infiniband/hw/hfi1/ |
H A D | mmu_rb.c | 59 INIT_HLIST_NODE(&h->mn.hlist); 61 h->mn.ops = &mn_opts; 68 ret = mmu_notifier_register(&h->mn, current->mm); 86 mmgrab(handler->mn.mm); 89 mmu_notifier_unregister(&handler->mn, handler->mn.mm); 115 mmdrop(handler->mn.mm); 129 if (current->mm != handler->mn.mm) 233 if (current->mm != handler->mn.mm) 261 static int mmu_notifier_range_start(struct mmu_notifier *mn, argument [all...] |
H A D | mmu_rb.h | 34 * they fit together in one cache line. mn is relatively rarely 39 struct mmu_notifier mn; member in struct:mmu_rb_handler
|
/linux-master/drivers/iommu/arm/arm-smmu-v3/ |
H A D | arm-smmu-v3-sva.c | 16 struct mmu_notifier mn; member in struct:arm_smmu_mmu_notifier 24 #define mn_to_smmu(mn) container_of(mn, struct arm_smmu_mmu_notifier, mn) 214 static void arm_smmu_mm_arch_invalidate_secondary_tlbs(struct mmu_notifier *mn, argument 219 struct arm_smmu_mmu_notifier *smmu_mn = mn_to_smmu(mn); 252 static void arm_smmu_mm_release(struct mmu_notifier *mn, struct mm_struct *mm) argument 254 struct arm_smmu_mmu_notifier *smmu_mn = mn_to_smmu(mn); 277 static void arm_smmu_mmu_notifier_free(struct mmu_notifier *mn) argument 279 kfree(mn_to_smmu(mn)); [all...] |
/linux-master/drivers/gpu/drm/radeon/ |
H A D | radeon_mn.c | 42 * @mn: our notifier 49 static bool radeon_mn_invalidate(struct mmu_interval_notifier *mn, argument 53 struct radeon_bo *bo = container_of(mn, struct radeon_bo, notifier);
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_cs.h | 68 struct amdgpu_mn *mn; member in struct:amdgpu_cs_parser
|
/linux-master/mm/ |
H A D | page_ext.c | 416 struct memory_notify *mn = arg; local 421 ret = online_page_ext(mn->start_pfn, 422 mn->nr_pages, mn->status_change_nid); 425 offline_page_ext(mn->start_pfn, 426 mn->nr_pages); 429 offline_page_ext(mn->start_pfn, 430 mn->nr_pages);
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | gk20a.c | 53 struct nvkm_mm_node *mn; member in struct:gk20a_instobj 130 return (u64)gk20a_instobj(memory)->mn->offset << 12; 136 return (u64)gk20a_instobj(memory)->mn->length << 12; 288 .mem = node->mn, 304 dma_free_attrs(dev, (u64)node->base.mn->length << PAGE_SHIFT, 317 struct nvkm_mm_node *r = node->base.mn; 335 for (i = 0; i < node->base.mn->length; i++) { 416 node->base.mn = &node->r; 494 node->base.mn = r; 543 size, align, (u64)node->mn [all...] |
/linux-master/arch/x86/kernel/apic/ |
H A D | x2apic_uv_x.c | 1160 struct mn { struct 1168 static void get_mn(struct mn *mnp) 1190 struct mn mn; local 1192 get_mn(&mn); 1193 hi->gpa_mask = mn.m_val ? 1194 (1UL << (mn.m_val + mn.n_val)) - 1 : 1197 hi->m_val = mn.m_val; 1198 hi->n_val = mn [all...] |
/linux-master/net/sched/ |
H A D | sch_teql.c | 230 struct neighbour *mn; local 232 mn = __neigh_lookup_errno(n->tbl, n->primary_key, dev); 234 if (IS_ERR(mn)) 235 return PTR_ERR(mn); 236 n = mn;
|
/linux-master/drivers/scsi/qla2xxx/ |
H A D | qla_bsg.c | 1025 struct verify_chip_entry_84xx *mn = NULL; local 1071 mn = dma_pool_zalloc(ha->s_dma_pool, GFP_KERNEL, &mn_dma); 1072 if (!mn) { 1082 mn->entry_type = VERIFY_CHIP_IOCB_TYPE; 1083 mn->entry_count = 1; 1089 mn->options = cpu_to_le16(options); 1090 mn->fw_ver = cpu_to_le32(fw_ver); 1091 mn->fw_size = cpu_to_le32(data_len); 1092 mn->fw_seq_size = cpu_to_le32(data_len); 1093 put_unaligned_le64(fw_dma, &mn 1135 struct access_chip_84xx *mn = NULL; local [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_object.c | 242 struct i915_mmap_offset *mmo, *mn; local 266 rbtree_postorder_for_each_entry_safe(mmo, mn, &obj->mmo.offsets, offset) 318 struct i915_mmap_offset *mmo, *mn; local 322 rbtree_postorder_for_each_entry_safe(mmo, mn,
|