/barrelfish-2018-10-04/include/target/arm/barrelfish_kpi/ |
H A D | paging_arm_v5.h | 26 #define LARGE_PAGE_SIZE (1u << LARGE_PAGE_BITS) macro 27 #define LARGE_PAGE_MASK (LARGE_PAGE_SIZE - 1)
|
H A D | paging_arm_v7.h | 26 #define LARGE_PAGE_SIZE (1u << LARGE_PAGE_BITS) macro 27 #define LARGE_PAGE_MASK (LARGE_PAGE_SIZE - 1)
|
/barrelfish-2018-10-04/lib/posixcompat/ |
H A D | sbrk.c | 17 #define SBRK_REGION_BYTES (8UL * 512UL * LARGE_PAGE_SIZE) 19 #define SBRK_MIN_MAPPING (16 * LARGE_PAGE_SIZE) 20 #define SBRK_REGION_ALIGNMENT LARGE_PAGE_SIZE 23 #define SBRK_REGION_BYTES (8UL * 512UL * LARGE_PAGE_SIZE) 25 #define SBRK_MIN_MAPPING (16 * LARGE_PAGE_SIZE)
|
/barrelfish-2018-10-04/lib/barrelfish/ |
H A D | morecore.c | 49 if (bytes < LARGE_PAGE_SIZE) { 50 bytes = LARGE_PAGE_SIZE; 53 bytes = ROUND_UP(bytes, LARGE_PAGE_SIZE); 119 morecore_flags |= (alignment == LARGE_PAGE_SIZE ? VREGION_FLAGS_LARGE : 0);
|
H A D | init.c | 172 case LARGE_PAGE_SIZE: 259 || morecore_pagesize == LARGE_PAGE_SIZE );
|
/barrelfish-2018-10-04/include/arch/aarch64/barrelfish_kpi/ |
H A D | paging_arch.h | 30 #define LARGE_PAGE_SIZE VMSAv8_64_L2_BLOCK_SIZE macro
|
/barrelfish-2018-10-04/include/arch/x86_32/barrelfish_kpi/ |
H A D | paging_arch.h | 34 #define LARGE_PAGE_SIZE X86_32_LARGE_PAGE_SIZE macro
|
/barrelfish-2018-10-04/include/arch/x86_64/barrelfish_kpi/ |
H A D | paging_arch.h | 29 #define LARGE_PAGE_SIZE X86_64_LARGE_PAGE_SIZE macro
|
/barrelfish-2018-10-04/usr/bench/largepage/ |
H A D | largepage_64_bench.c | 32 #define LARGE_PAGE_SIZE X86_64_LARGE_PAGE_SIZE macro 256 bytes = LARGE_PAGE_SIZE; 263 assert(bytes >= LARGE_PAGE_SIZE); 271 test_region((uint32_t*) SAFE_PMAP_ADDR, LARGE_PAGE_SIZE, false, true); 274 bytes = LARGE_PAGE_SIZE; 281 assert(bytes >= LARGE_PAGE_SIZE); 292 test_region((uint32_t*)SAFE_PMAP_ADDR_L, LARGE_PAGE_SIZE, true, true);
|
/barrelfish-2018-10-04/usr/tests/twolevel_cspace/ |
H A D | test_create_l1l2.c | 68 setup(LARGE_PAGE_SIZE); 100 setup(LARGE_PAGE_SIZE);
|
/barrelfish-2018-10-04/usr/tests/large_page/ |
H A D | map_test_32.c | 20 #define TOTAL_SIZE (PAGE_COUNT * LARGE_PAGE_SIZE)
|
/barrelfish-2018-10-04/lib/cpuid/ |
H A D | cpuid_intel.c | 166 CPUID_FILL_TLB(0x02, tci, 0xff, 2, 1, CPUID_CACHE_TYPE_INSTR, LARGE_PAGE_SIZE); 168 CPUID_FILL_TLB(0x04, tci, 4, 8, 1, CPUID_CACHE_TYPE_DATA, LARGE_PAGE_SIZE); 169 CPUID_FILL_TLB(0x05, tci, 4, 32, 1, CPUID_CACHE_TYPE_DATA, LARGE_PAGE_SIZE); 179 CPUID_FILL_TLB(0x0b, tci, 4, 4, 1, CPUID_CACHE_TYPE_INSTR, LARGE_PAGE_SIZE); 238 CPUID_FILL_TLB(0x50, tci, 0xff, 64, 1, CPUID_CACHE_TYPE_INSTR, LARGE_PAGE_SIZE); 240 CPUID_FILL_TLB(0x51, tci, 0xff, 128, 1, CPUID_CACHE_TYPE_INSTR, LARGE_PAGE_SIZE); 244 CPUID_FILL_TLB(0x55, tci, 0xff, 7, 1, CPUID_CACHE_TYPE_INSTR, LARGE_PAGE_SIZE); 246 CPUID_FILL_TLB(0x56, tci, 4, 16, 1, CPUID_CACHE_TYPE_DATA, LARGE_PAGE_SIZE); 252 CPUID_FILL_TLB(0x5a, tci, 4, 32, 1, CPUID_CACHE_TYPE_DATA, LARGE_PAGE_SIZE); 280 CPUID_FILL_TLB(0x76, tci, 0xff, 8, 1, CPUID_CACHE_TYPE_INSTR, LARGE_PAGE_SIZE); [all...] |
H A D | cpuid_amd.c | 319 ti->pagesize = LARGE_PAGE_SIZE; 341 ti->pagesize = (idx < 2) ? LARGE_PAGE_SIZE : BASE_PAGE_SIZE;
|
/barrelfish-2018-10-04/usr/tests/capops/ |
H A D | retype.c | 70 setup(LARGE_PAGE_SIZE); 160 setup(LARGE_PAGE_SIZE); 240 setup(LARGE_PAGE_SIZE); 302 setup(LARGE_PAGE_SIZE); 312 err = cap_retype(cap, bunch_o_ram, LARGE_PAGE_SIZE, ObjType_Frame, 321 err = cap_retype(cap, bunch_o_ram, LARGE_PAGE_SIZE - 31*BASE_PAGE_SIZE, ObjType_Frame,
|
/barrelfish-2018-10-04/usr/vnode_map_32/ |
H A D | vnode_map_32.c | 117 err = pmap->f.determine_addr_raw(pmap, a.size, LARGE_PAGE_SIZE, (genvaddr_t*)&address); 160 err = pmap->f.determine_addr_raw(pmap, a.size, LARGE_PAGE_SIZE, (genvaddr_t*)&address);
|
/barrelfish-2018-10-04/lib/numa/ |
H A D | alloc.c | 46 case LARGE_PAGE_SIZE: 48 return LARGE_PAGE_SIZE; 60 case LARGE_PAGE_SIZE: 62 return LARGE_PAGE_SIZE;
|
/barrelfish-2018-10-04/lib/barrelfish/vspace/ |
H A D | mmu_aware.c | 137 // mapoffset is aligned to at least LARGE_PAGE_SIZE. 138 alloc_size = ROUND_UP(req_size, LARGE_PAGE_SIZE); 215 VREGION_FLAGS_READ_WRITE | VREGION_FLAGS_LARGE, LARGE_PAGE_SIZE,
|
/barrelfish-2018-10-04/usr/vnode_map_test/ |
H A D | vnode_map_test.c | 271 err = pmap->f.determine_addr_raw(pmap, a.size, LARGE_PAGE_SIZE, &address); 306 err = pmap->f.determine_addr_raw(pmap, a.size, LARGE_PAGE_SIZE, &address);
|
/barrelfish-2018-10-04/tools/elver/ |
H A D | elver.c | 42 #define LARGE_PAGE_SIZE 0x200000 macro 298 for(uint32_t i = 0; i < ONE_GIB; i += LARGE_PAGE_SIZE) {
|
/barrelfish-2018-10-04/kernel/arch/x86_64/ |
H A D | page_mappings_arch.c | 115 if (get_size(src) < LARGE_PAGE_SIZE) { 359 return LARGE_PAGE_SIZE; 407 pagesize = LARGE_PAGE_SIZE;
|
/barrelfish-2018-10-04/lib/barrelfish/arch/arm/ |
H A D | pmap.c | 415 fi.bytes >= LARGE_PAGE_SIZE && 523 fi.bytes >= LARGE_PAGE_SIZE && 526 page_size = LARGE_PAGE_SIZE; 750 fi.bytes >= LARGE_PAGE_SIZE && 754 page_size = LARGE_PAGE_SIZE; 1011 size_t free_count = DIVIDE_ROUND_UP(size, LARGE_PAGE_SIZE);
|
/barrelfish-2018-10-04/kernel/arch/x86_32/ |
H A D | page_mappings_arch.c | 366 pagesize = LARGE_PAGE_SIZE;
|
/barrelfish-2018-10-04/lib/barrelfish/arch/x86_32/ |
H A D | pmap.c | 647 pagesize = LARGE_PAGE_SIZE; 846 size_t free_count = DIVIDE_ROUND_UP(size, LARGE_PAGE_SIZE);
|
/barrelfish-2018-10-04/kernel/arch/armv7/ |
H A D | paging.c | 383 if ((src_lpaddr & (LARGE_PAGE_SIZE - 1))) {
|
/barrelfish-2018-10-04/lib/barrelfish/arch/x86_64/ |
H A D | pmap.c | 599 size = ROUND_UP(size, LARGE_PAGE_SIZE);
|