/linux-master/mm/ |
H A D | mempolicy.c | 503 static void queue_folios_pmd(pmd_t *pmd, struct mm_walk *walk) argument 508 if (unlikely(is_pmd_migration_entry(*pmd))) { 512 folio = pmd_folio(*pmd); 535 static int queue_folios_pte_range(pmd_t *pmd, unsigned long addr, argument 546 ptl = pmd_trans_huge_lock(pmd, vma); 548 queue_folios_pmd(pmd, walk); 553 mapped_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
|
H A D | percpu.c | 3185 pmd_t *pmd; local 3204 pmd = memblock_alloc(PMD_TABLE_SIZE, PMD_TABLE_SIZE); 3205 if (!pmd) 3207 pud_populate(&init_mm, pud, pmd); 3210 pmd = pmd_offset(pud, addr); 3211 if (!pmd_present(*pmd)) { 3217 pmd_populate_kernel(&init_mm, pmd, new);
|
H A D | ksm.c | 615 static int break_ksm_pmd_entry(pmd_t *pmd, unsigned long addr, unsigned long next, argument 624 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); 1377 pmd_t *pmd; local 1390 pmd = mm_find_pmd(mm, addr); 1391 if (!pmd) 1398 pmde = pmdp_get_lockless(pmd); 1406 ptep = pte_offset_map_lock(mm, pmd, addr, &ptl);
|
H A D | memcontrol.c | 6408 * Caller should make sure that pmd_trans_huge(pmd) is true. 6411 unsigned long addr, pmd_t pmd, union mc_target *target) 6417 if (unlikely(is_swap_pmd(pmd))) { 6419 !is_pmd_migration_entry(pmd)); 6422 page = pmd_page(pmd); 6442 unsigned long addr, pmd_t pmd, union mc_target *target) 6448 static int mem_cgroup_count_precharge_pte_range(pmd_t *pmd, argument 6456 ptl = pmd_trans_huge_lock(pmd, vma); 6463 if (get_mctgt_type_thp(vma, addr, *pmd, NULL) == MC_TARGET_PAGE) 6469 pte = pte_offset_map_lock(vma->vm_mm, pmd, add 6410 get_mctgt_type_thp(struct vm_area_struct *vma, unsigned long addr, pmd_t pmd, union mc_target *target) argument 6441 get_mctgt_type_thp(struct vm_area_struct *vma, unsigned long addr, pmd_t pmd, union mc_target *target) argument 6645 mem_cgroup_move_charge_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, struct mm_walk *walk) argument [all...] |
H A D | filemap.c | 3234 ptep = pte_offset_map(vmf->pmd, vmf->address); 3438 if (pmd_trans_huge(*vmf->pmd)) { 3444 if (pmd_none(*vmf->pmd) && folio_test_pmd_mappable(folio)) { 3454 if (pmd_none(*vmf->pmd) && vmf->prealloc_pte) 3455 pmd_install(mm, vmf->pmd, &vmf->prealloc_pte); 3625 vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, addr, &vmf->ptl);
|
H A D | swap_state.c | 841 pte = pte_offset_map(vmf->pmd, addr);
|
H A D | hugetlb.c | 416 * participate in pmd sharing. This is only a possible 5390 * returned ptep could go away if part of a shared pmd and 5824 * for pmd sharing. And, i_mmap_rwsem is required to set up 5825 * pmd sharing. This is important as page tables for this 7048 * vma specific semaphore used for pmd sharing and fault/truncation 7281 * Determine if start,end range within vma could be mapped by shared pmd. 7283 * shared pmd mappings. 7308 * Search for a shareable pmd page for hugetlb. In any case calls pmd_alloc() 7310 * !shared pmd case because we can allocate the pmd late 7467 pmd_t *pmd; local [all...] |
/linux-master/arch/sh/kernel/cpu/sh3/ |
H A D | entry.S | 98 ! Find the pmd/pte entry and loadtlb
|
/linux-master/fs/ |
H A D | userfaultfd.c | 299 pmd_t *pmd, _pmd; local 315 pmd = pmd_offset(pud, address); 317 _pmd = pmdp_get_lockless(pmd); 331 pte = pte_offset_map(pmd, address);
|
/linux-master/include/linux/ |
H A D | rmap.h | 678 pmd_t *pmd; member in struct:page_vma_mapped_walk
|
/linux-master/include/trace/events/ |
H A D | xen.h | 166 __entry->pmdval = pmdval.pmd),
|
/linux-master/drivers/net/ethernet/intel/ice/ |
H A D | ice_ptp_hw.c | 2239 u64 total_offset, pmd, val; local 2298 err = ice_phy_calc_pmd_adj_e82x(hw, port, link_spd, fec_mode, &pmd); 2306 total_offset += pmd; 2308 total_offset -= pmd;
|
H A D | ice_common.c | 3869 options[i].pmd = FIELD_GET(ICE_AQC_PORT_OPT_PMD_COUNT_M, 3870 options[i].pmd); 3874 options[i].pmd, options[i].max_lane_speed);
|
H A D | ice_adminq_cmd.h | 1579 u8 pmd; member in struct:ice_aqc_get_port_options_elem
|
/linux-master/sound/soc/codecs/ |
H A D | cs42l43.c | 1468 static int cs42l43_dapm_wait_completion(struct completion *pmu, struct completion *pmd, argument 1478 reinit_completion(pmd); 1484 time_left = wait_for_completion_timeout(pmd, msecs_to_jiffies(timeout_ms));
|
/linux-master/arch/x86/kvm/mmu/ |
H A D | mmu.c | 3048 pmd_t pmd; local 3091 pmd = READ_ONCE(*pmd_offset(&pud, hva)); 3092 if (pmd_none(pmd) || !pmd_present(pmd)) 3095 if (pmd_leaf(pmd)) 3165 * the pmd can't be split from under us.
|
/linux-master/arch/mips/mm/ |
H A D | tlbex.c | 679 unsigned int pmd, int lid) 681 UASM_i_LW(p, tmp, 0, pmd); 758 * TMP will be clobbered, PTR will hold the pmd entry. 838 uasm_i_ld(p, ptr, 0, ptr); /* get pmd pointer */ 839 uasm_i_dsrl_safe(p, tmp, tmp, PMD_SHIFT-3); /* get pmd offset in bytes */ 841 uasm_i_daddu(p, ptr, ptr, tmp); /* add in pmd offset */ 1143 uasm_i_ld(p, LOC_PTEP, 0, ptr); /* get pmd pointer */ 1154 uasm_i_daddu(p, ptr, ptr, scratch); /* add in pmd offset */ 1162 /* get pmd offset in bytes */ 1170 uasm_i_daddu(p, ptr, ptr, scratch); /* add in pmd offse 678 build_is_huge_pte(u32 **p, struct uasm_reloc **r, unsigned int tmp, unsigned int pmd, int lid) argument [all...] |
/linux-master/kernel/events/ |
H A D | core.c | 7557 pmd_t *pmdp, pmd; local 7586 pmd = pmdp_get_lockless(pmdp); 7587 if (!pmd_present(pmd)) 7590 if (pmd_leaf(pmd)) 7591 return pmd_leaf_size(pmd); 7593 ptep = pte_offset_map(&pmd, addr);
|