/linux-master/arch/powerpc/include/asm/nohash/32/ |
H A D | pte-8xx.h | 124 unsigned long clr, unsigned long set, int huge); 137 int huge = psize > mmu_virtual_psize ? 1 : 0; local 139 pte_update(vma->vm_mm, address, ptep, clr, set, huge); 176 static inline int number_of_cells_per_pte(pmd_t *pmd, pte_basic_t val, int huge) argument 178 if (!huge) 189 unsigned long clr, unsigned long set, int huge) 197 num = number_of_cells_per_pte(pmd, new, huge); 188 pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p, unsigned long clr, unsigned long set, int huge) argument
|
/linux-master/tools/testing/selftests/mm/ |
H A D | charge_reserved_hugetlb.sh | 54 if [[ -e /mnt/huge ]]; then 55 rm -rf /mnt/huge/* 56 umount /mnt/huge || echo error 57 rmdir /mnt/huge 262 if [[ -e /mnt/huge ]]; then 263 rm -rf /mnt/huge/* 264 umount /mnt/huge 265 rmdir /mnt/huge 292 mkdir -p /mnt/huge 293 mount -t hugetlbfs -o pagesize=${MB}M,size=256M none /mnt/huge [all...] |
H A D | run_vmtests.sh | 67 test transparent huge pages 69 test hugetlbfs huge pages 119 for huge in -t -T "-H -m $hugetlb_mb"; do 128 # just test partial gup when hit a huge in whatever form 131 $huge $test_cmd $write $share $num 139 # get huge pagesize and freepages from /proc/meminfo 150 # huge pages totaling 256MB (262144KB) in size. The userfaultfd 151 # hugetlb test requires a minimum of 2 * nr_cpus huge pages. Take 153 # number of huge pages available. 179 printf "Not enough huge page [all...] |
H A D | hugetlb_reparenting_test.sh | 39 MNT='/mnt/huge/' 192 rm -rf /mnt/huge/*
|
/linux-master/arch/powerpc/include/asm/book3s/64/ |
H A D | hash.h | 161 pte_t *ptep, unsigned long pte, int huge); 168 int huge) 185 /* huge pages use the old page table lock */ 186 if (!huge) 191 hpte_need_flush(mm, addr, ptep, old, huge); 164 hash__pte_update(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned long clr, unsigned long set, int huge) argument
|
H A D | radix.h | 176 int huge) 181 if (!huge) 172 radix__pte_update(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned long clr, unsigned long set, int huge) argument
|
H A D | pgtable.h | 333 unsigned long set, int huge) 336 return radix__pte_update(mm, addr, ptep, clr, set, huge); 337 return hash__pte_update(mm, addr, ptep, clr, set, huge); 331 pte_update(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned long clr, unsigned long set, int huge) argument
|
/linux-master/arch/powerpc/include/asm/nohash/ |
H A D | pgtable.h | 7 unsigned long clr, unsigned long set, int huge); 51 unsigned long clr, unsigned long set, int huge) 64 /* huge pages use the old page table lock */ 65 if (!huge) 113 int huge = psize > mmu_virtual_psize ? 1 : 0; local 115 pte_update(vma->vm_mm, address, ptep, 0, set, huge); 50 pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p, unsigned long clr, unsigned long set, int huge) argument
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | hash_tlb.c | 41 pte_t *ptep, unsigned long pte, int huge) 61 if (huge) { 40 hpte_need_flush(struct mm_struct *mm, unsigned long addr, pte_t *ptep, unsigned long pte, int huge) argument
|
/linux-master/arch/alpha/lib/ |
H A D | ev6-clear_user.S | 86 subq $1, 16, $4 # .. .. .. E : If < 16, we can not use the huge loop 87 and $16, 0x3f, $2 # .. .. E .. : Forward work for huge loop 88 subq $2, 0x40, $3 # .. E .. .. : bias counter (huge loop)
|
/linux-master/arch/loongarch/mm/ |
H A D | init.c | 143 int huge = pmd_val(*pmd) & _PAGE_HUGE; local 145 if (huge) 148 return huge;
|
/linux-master/drivers/misc/lkdtm/ |
H A D | bugs.c | 304 static volatile unsigned int huge = INT_MAX - 2; variable 311 value = huge; 326 value = huge;
|
/linux-master/mm/ |
H A D | memory-failure.c | 899 [MF_MSG_HUGE] = "huge page", 900 [MF_MSG_FREE_HUGE] = "free huge page", 1640 pr_info("%#lx: could not lock mapping for mapped huge page\n", pfn); 2534 bool huge = false; local 2588 huge = true; 2604 huge = true; 2622 if (!huge) 2676 bool huge = folio_test_hugetlb(folio); local 2683 if (!huge && folio_test_large(folio)) { 2692 if (!huge) [all...] |
H A D | shmem.c | 121 int huge; member in struct:shmem_options 508 * Definitions for "huge tmpfs": tmpfs mounted with the huge= option 511 * disables huge pages for the mount; 513 * enables huge pages for the mount; 515 * only allocate huge pages if the page will be fully within i_size, 518 * only allocate huge pages if requested with fadvise()/madvise(); 531 * disables huge on shm_mnt and all mounts, for emergency use; 533 * enables huge on shm_mnt and all mounts, w/o needing option, for testing; 558 switch (SHMEM_SB(inode->i_sb)->huge) { 596 shmem_format_huge(int huge) argument 1634 shmem_alloc_and_add_folio(gfp_t gfp, struct inode *inode, pgoff_t index, struct mm_struct *fault_mm, bool huge) argument 4733 int huge; local [all...] |
H A D | zsmalloc.c | 21 * For the huge page, this is always 0, so we use this field 27 * PG_owner_priv_1: identifies the huge component page 245 unsigned int huge:HUGE_BITS; member in struct:zspage::__anon1351 265 /* huge object: pages_per_zspage == 1 && maxobj_per_zspage == 1 */ 268 zspage->huge = 1; 273 return zspage->huge; 1257 * zs_huge_class_size() - Returns the size (in bytes) of the first huge 1261 * The function returns the size of the first huge class - any object of equal 1267 * Return: the size (in bytes) of the first huge zsmalloc &size_class. 2136 * so huge_class_size holds the size of the first huge [all...] |
H A D | mm_init.c | 1475 * Assume the largest contiguous order of interest is a huge page. 2462 bool huge; local 2523 huge = is_vm_area_hugepages(table); 2540 virt ? (huge ? "vmalloc hugepage" : "vmalloc") : "linear");
|
H A D | filemap.c | 855 bool huge = folio_test_hugetlb(folio); local 863 if (!huge) { 915 if (!huge) {
|
/linux-master/arch/parisc/mm/ |
H A D | init.c | 400 bool huge = false; local 410 huge = true; 415 huge = true; 421 if (huge)
|
/linux-master/include/linux/ |
H A D | shmem_fs.h | 64 unsigned char huge; /* Whether to try for hugepages */ member in struct:shmem_sb_info
|
/linux-master/fs/ubifs/ |
H A D | ubifs-media.h | 469 * @huge: huge type device descriptor 473 * uses standard Linux "new" and "huge" device node encodings. 477 __le64 huge; member in union:ubifs_dev_desc
|
H A D | super.c | 220 else if (ui->data_len == sizeof(dev->huge)) 221 rdev = huge_decode_dev(le64_to_cpu(dev->huge));
|
/linux-master/lib/ |
H A D | test_maple_tree.c | 292 unsigned long huge = 4000UL * 1000 * 1000; local 295 i = huge; 298 for (j = huge; j >= i; j /= 2) { 317 unsigned long huge; local 322 huge = 2147483647UL; 324 huge = 4000UL * 1000 * 1000; 327 while (i < huge) { 329 for (j = i; j >= huge; j *= 2) { 341 unsigned long huge = 8000UL * 1000 * 1000; local 343 check_insert(mt, huge, (voi [all...] |
/linux-master/arch/m68k/fpsp040/ |
H A D | kernel_ex.S | 27 huge: .long 0x7ffe0000,0xffffffff,0xffffffff label
|
/linux-master/arch/powerpc/include/asm/book3s/32/ |
H A D | pgtable.h | 259 unsigned long clr, unsigned long set, int huge) 258 pte_update(struct mm_struct *mm, unsigned long addr, pte_t *p, unsigned long clr, unsigned long set, int huge) argument
|
/linux-master/arch/sparc/include/asm/ |
H A D | pgtable_64.h | 79 #error PMD_SHIFT must equal HPAGE_SHIFT for transparent huge pages. 218 * the first physical page in the machine is at some huge physical address, 801 /* pmd_bad() is only called on non-trans-huge PMDs. Our encoding is
|