Lines Matching defs:pmd

52  * The pmd table for the upper-most set of pages.
75 pmdval_t pmd;
83 .pmd = PMD_SECT_UNCACHED,
88 .pmd = PMD_SECT_BUFFERED,
93 .pmd = PMD_SECT_WT,
98 .pmd = PMD_SECT_WB,
103 .pmd = PMD_SECT_WBWA,
113 * via the "pmd" value. This is used to ensure that on ARMv6 and later,
118 void __init init_default_cache_policy(unsigned long pmd)
122 initial_pmd_value = pmd;
124 pmd &= PMD_SECT_CACHE_MASK;
127 if (cache_policies[i].pmd == pmd) {
377 pmd_t *pmd;
386 pmd = fixmap_pmd(FIXADDR_TOP);
387 pmd_populate_kernel(&init_mm, pmd, bm_pte);
681 mem_types[MT_MEMORY_RWX].prot_sect |= ecc_mask | cp->pmd;
683 mem_types[MT_MEMORY_RW].prot_sect |= ecc_mask | cp->pmd;
685 mem_types[MT_MEMORY_RO].prot_sect |= ecc_mask | cp->pmd;
689 mem_types[MT_ROM].prot_sect |= cp->pmd;
691 switch (cp->pmd) {
748 static pte_t * __init arm_pte_alloc(pmd_t *pmd, unsigned long addr,
752 if (pmd_none(*pmd)) {
754 __pmd_populate(pmd, __pa(pte), prot);
756 BUG_ON(pmd_bad(*pmd));
757 return pte_offset_kernel(pmd, addr);
760 static pte_t * __init early_pte_alloc(pmd_t *pmd, unsigned long addr,
763 return arm_pte_alloc(pmd, addr, prot, early_alloc);
766 static void __init alloc_init_pte(pmd_t *pmd, unsigned long addr,
772 pte_t *pte = arm_pte_alloc(pmd, addr, type->prot_l1, alloc);
780 static void __init __map_init_section(pmd_t *pmd, unsigned long addr,
784 pmd_t *p = pmd;
797 pmd++;
800 *pmd = __pmd(phys | type->prot_sect | (ng ? PMD_SECT_nG : 0));
802 } while (pmd++, addr += SECTION_SIZE, addr != end);
812 pmd_t *pmd = pmd_offset(pud, addr);
828 __map_init_section(pmd, addr, next, phys, type, ng);
830 alloc_init_pte(pmd, addr, next,
836 } while (pmd++, addr = next, addr != end);
918 pmd_t *pmd = pmd_offset(pud, addr);
922 *pmd++ = __pmd(phys | type->prot_sect | PMD_SECT_SUPER |
1093 pmd_t *pmd;
1107 pmd = pmd_off_k(addr);
1108 if (pmd_none(*pmd))
1119 pmd = pmd_off_k(addr) + 1;
1120 if (pmd_none(*pmd))
1240 * Find the first non-pmd-aligned page, and point
1242 * limit down to be pmd-aligned, which happens at the
1246 * bank can be non-pmd-aligned. The only exception is
1270 * Round the memblock limit down to a pmd size. This
1272 * last full pmd, which should be mapped.
1308 * equal to MODULES_VADDR and then we exit the pmd clearing. If we
1391 * Clear page table except top pmd used by early fixmaps