Searched refs:pmd (Results 276 - 293 of 293) sorted by relevance

<<1112

/linux-master/mm/
H A Dmempolicy.c503 static void queue_folios_pmd(pmd_t *pmd, struct mm_walk *walk) argument
508 if (unlikely(is_pmd_migration_entry(*pmd))) {
512 folio = pmd_folio(*pmd);
535 static int queue_folios_pte_range(pmd_t *pmd, unsigned long addr, argument
546 ptl = pmd_trans_huge_lock(pmd, vma);
548 queue_folios_pmd(pmd, walk);
553 mapped_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
H A Dpercpu.c3185 pmd_t *pmd; local
3204 pmd = memblock_alloc(PMD_TABLE_SIZE, PMD_TABLE_SIZE);
3205 if (!pmd)
3207 pud_populate(&init_mm, pud, pmd);
3210 pmd = pmd_offset(pud, addr);
3211 if (!pmd_present(*pmd)) {
3217 pmd_populate_kernel(&init_mm, pmd, new);
H A Dksm.c615 static int break_ksm_pmd_entry(pmd_t *pmd, unsigned long addr, unsigned long next, argument
624 pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl);
1377 pmd_t *pmd; local
1390 pmd = mm_find_pmd(mm, addr);
1391 if (!pmd)
1398 pmde = pmdp_get_lockless(pmd);
1406 ptep = pte_offset_map_lock(mm, pmd, addr, &ptl);
H A Dmemcontrol.c6408 * Caller should make sure that pmd_trans_huge(pmd) is true.
6411 unsigned long addr, pmd_t pmd, union mc_target *target)
6417 if (unlikely(is_swap_pmd(pmd))) {
6419 !is_pmd_migration_entry(pmd));
6422 page = pmd_page(pmd);
6442 unsigned long addr, pmd_t pmd, union mc_target *target)
6448 static int mem_cgroup_count_precharge_pte_range(pmd_t *pmd, argument
6456 ptl = pmd_trans_huge_lock(pmd, vma);
6463 if (get_mctgt_type_thp(vma, addr, *pmd, NULL) == MC_TARGET_PAGE)
6469 pte = pte_offset_map_lock(vma->vm_mm, pmd, add
6410 get_mctgt_type_thp(struct vm_area_struct *vma, unsigned long addr, pmd_t pmd, union mc_target *target) argument
6441 get_mctgt_type_thp(struct vm_area_struct *vma, unsigned long addr, pmd_t pmd, union mc_target *target) argument
6645 mem_cgroup_move_charge_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, struct mm_walk *walk) argument
[all...]
H A Dfilemap.c3234 ptep = pte_offset_map(vmf->pmd, vmf->address);
3438 if (pmd_trans_huge(*vmf->pmd)) {
3444 if (pmd_none(*vmf->pmd) && folio_test_pmd_mappable(folio)) {
3454 if (pmd_none(*vmf->pmd) && vmf->prealloc_pte)
3455 pmd_install(mm, vmf->pmd, &vmf->prealloc_pte);
3625 vmf->pte = pte_offset_map_lock(vma->vm_mm, vmf->pmd, addr, &vmf->ptl);
H A Dswap_state.c841 pte = pte_offset_map(vmf->pmd, addr);
H A Dhugetlb.c416 * participate in pmd sharing. This is only a possible
5390 * returned ptep could go away if part of a shared pmd and
5824 * for pmd sharing. And, i_mmap_rwsem is required to set up
5825 * pmd sharing. This is important as page tables for this
7048 * vma specific semaphore used for pmd sharing and fault/truncation
7281 * Determine if start,end range within vma could be mapped by shared pmd.
7283 * shared pmd mappings.
7308 * Search for a shareable pmd page for hugetlb. In any case calls pmd_alloc()
7310 * !shared pmd case because we can allocate the pmd late
7467 pmd_t *pmd; local
[all...]
/linux-master/arch/sh/kernel/cpu/sh3/
H A Dentry.S98 ! Find the pmd/pte entry and loadtlb
/linux-master/fs/
H A Duserfaultfd.c299 pmd_t *pmd, _pmd; local
315 pmd = pmd_offset(pud, address);
317 _pmd = pmdp_get_lockless(pmd);
331 pte = pte_offset_map(pmd, address);
/linux-master/include/linux/
H A Drmap.h678 pmd_t *pmd; member in struct:page_vma_mapped_walk
/linux-master/include/trace/events/
H A Dxen.h166 __entry->pmdval = pmdval.pmd),
/linux-master/drivers/net/ethernet/intel/ice/
H A Dice_ptp_hw.c2239 u64 total_offset, pmd, val; local
2298 err = ice_phy_calc_pmd_adj_e82x(hw, port, link_spd, fec_mode, &pmd);
2306 total_offset += pmd;
2308 total_offset -= pmd;
H A Dice_common.c3869 options[i].pmd = FIELD_GET(ICE_AQC_PORT_OPT_PMD_COUNT_M,
3870 options[i].pmd);
3874 options[i].pmd, options[i].max_lane_speed);
H A Dice_adminq_cmd.h1579 u8 pmd; member in struct:ice_aqc_get_port_options_elem
/linux-master/sound/soc/codecs/
H A Dcs42l43.c1468 static int cs42l43_dapm_wait_completion(struct completion *pmu, struct completion *pmd, argument
1478 reinit_completion(pmd);
1484 time_left = wait_for_completion_timeout(pmd, msecs_to_jiffies(timeout_ms));
/linux-master/arch/x86/kvm/mmu/
H A Dmmu.c3048 pmd_t pmd; local
3091 pmd = READ_ONCE(*pmd_offset(&pud, hva));
3092 if (pmd_none(pmd) || !pmd_present(pmd))
3095 if (pmd_leaf(pmd))
3165 * the pmd can't be split from under us.
/linux-master/arch/mips/mm/
H A Dtlbex.c679 unsigned int pmd, int lid)
681 UASM_i_LW(p, tmp, 0, pmd);
758 * TMP will be clobbered, PTR will hold the pmd entry.
838 uasm_i_ld(p, ptr, 0, ptr); /* get pmd pointer */
839 uasm_i_dsrl_safe(p, tmp, tmp, PMD_SHIFT-3); /* get pmd offset in bytes */
841 uasm_i_daddu(p, ptr, ptr, tmp); /* add in pmd offset */
1143 uasm_i_ld(p, LOC_PTEP, 0, ptr); /* get pmd pointer */
1154 uasm_i_daddu(p, ptr, ptr, scratch); /* add in pmd offset */
1162 /* get pmd offset in bytes */
1170 uasm_i_daddu(p, ptr, ptr, scratch); /* add in pmd offse
678 build_is_huge_pte(u32 **p, struct uasm_reloc **r, unsigned int tmp, unsigned int pmd, int lid) argument
[all...]
/linux-master/kernel/events/
H A Dcore.c7557 pmd_t *pmdp, pmd; local
7586 pmd = pmdp_get_lockless(pmdp);
7587 if (!pmd_present(pmd))
7590 if (pmd_leaf(pmd))
7591 return pmd_leaf_size(pmd);
7593 ptep = pte_offset_map(&pmd, addr);

Completed in 383 milliseconds

<<1112