/fuchsia/zircon/kernel/arch/arm64/ |
H A D | mmu.cpp | 392 // currently we only support allocating a single page 395 vm_page_t* page; 396 zx_status_t status = pmm_alloc_page(0, &page, paddrp); 401 page->state = VM_PAGE_STATE_MMU; 404 LOCAL_KTRACE0("page table alloc"); 413 // currently we only support freeing a single page 416 vm_page_t* page; 418 LOCAL_KTRACE0("page table free"); 420 page = paddr_to_vm_page(paddr); 421 if (!page) { [all...] |
H A D | rules.mk | 90 GLOBAL_LDFLAGS += -z max-page-size=4096
|
/fuchsia/zircon/system/dev/nand/nandpart/test/ |
H A D | aml-bad-block-test.cpp | 23 // It is conveniant for mapping for nand page size to be same as DRAM page size. 95 unittest_printf("Trying to write to a page that is out of range!\n"); 101 for (NandPage page = start_page; page < end_page; page++) { 102 context->table_entries.erase(page); 179 unittest_printf("Trying to write to a page that isn't erased!\n"); 398 for (uint32_t page = 0; page < [all...] |
/fuchsia/zircon/third_party/ulib/jemalloc/bin/ |
H A D | jemalloc-config.in | 17 --mandir : Print manual page installation directory.
|
/fuchsia/zircon/system/uapp/nand-util/ |
H A D | main.cpp | 39 --read (-r) --absolute xxx : Read the page number xxx (0-based). 42 --absolute (-a) xxx : Use an absolute page number. 43 --page (-p) xxx : Use the xxx page number (from within a block). 86 bool DumpPage(uint32_t page) const; 162 bool NandBroker::DumpPage(uint32_t page) const { 163 if (!ReadPages(page, 1)) { 168 uint32_t address = page * info_.page_size; 225 {"page", required_argument, nullptr, 'p'}, 293 printf("Provide either a block + page o [all...] |
/fuchsia/zircon/system/ulib/hid-parser/include/hid-parser/ |
H A D | parser.h | 128 uint16_t page; member in struct:hid::Usage
|
/fuchsia/zircon/kernel/arch/arm64/include/arch/arm64/ |
H A D | mmu.h | 53 #elif MMU_IDENT_SIZE_SHIFT <= 29 /* Use 2MB block mappings (4K page size) */ 55 #elif MMU_IDENT_SIZE_SHIFT <= 30 /* Use 512MB block mappings (64K page size) */ 57 #elif MMU_IDENT_SIZE_SHIFT <= 39 /* Use 1GB block mappings (4K page size) */ 59 #elif MMU_IDENT_SIZE_SHIFT <= 42 /* Use 512MB block mappings (64K page size) */ 92 #error User address space size must be larger than page size 106 #error Kernel address space size must be larger than page size 120 #error Ident address space size must be larger than page size 134 #error Guest physical address space size must be larger than page size 246 /* Default configuration for main kernel page table: 280 /* Enable cached page tabl [all...] |
/fuchsia/zircon/kernel/arch/x86/include/arch/ |
H A D | hypervisor.h | 17 #include <hypervisor/page.h>
|
/fuchsia/zircon/kernel/arch/x86/page_tables/ |
H A D | page_tables.cpp | 23 // Return the page size for this level 39 // Whether an address is aligned to the page size of this level 118 // Utility for coalescing cache line flushes when modifying page tables. This 119 // allows us to mutate adjacent page table entries without having to flush for 166 // Utility for managing consistency of the page tables from a cache and TLB 168 // refer to it, and that changes to the page tables have appropriate visiblity 170 // class, even if the page table change failed. 178 void queue_free(vm_page_t* page) TA_NO_THREAD_SAFETY_ANALYSIS { 181 list_add_tail(&to_free_, &page->queue_node); 213 // support deferring invoking pmm_free() until after we've left the page 514 vm_page_t* page = paddr_to_vm_page(ptable_phys); local [all...] |
/fuchsia/zircon/system/public/zircon/hw/ |
H A D | usb-mass-storage.h | 117 uint8_t page; member in struct:__anon774
|
/fuchsia/zircon/kernel/vm/include/vm/ |
H A D | vm_object.h | 22 #include <vm/page.h> 88 // page that is already not pinned (do not expose this function to 168 // create a copy-on-write clone vmo at the page-aligned offset and length 181 // get a pointer to the page structure and/or physical address at the specified offset. 184 vm_page_t** page, paddr_t* pa) TA_REQ(lock_) {
|
/fuchsia/zircon/third_party/ulib/jemalloc/src/ |
H A D | stats.c | 39 size_t page; local 43 CTL_GET("arenas.page", &page, size_t); 158 nregs, slab_size / page, util, nfills, 168 nregs, slab_size / page, util, nslabs, 256 size_t page, pactive, pdirty, mapped, retained; local 265 CTL_GET("arenas.page", &page, size_t); 387 "active: %12zu\n", pactive * page); 667 "Unused dirty page deca [all...] |
/fuchsia/zircon/kernel/arch/arm64/include/arch/ |
H A D | hypervisor.h | 15 #include <hypervisor/page.h> 86 // Provides a smart pointer to an EL2State allocated in its own page. 88 // We allocate an EL2State into its own page as the structure is passed between 90 // EL2State will not cross a page boundary and be incorrectly accessed in EL2.
|
/fuchsia/zircon/third_party/ulib/jemalloc/include/jemalloc/internal/ |
H A D | size_classes.sh | 280 * psz: 'yes' if a multiple of the page size, 'no' otherwise. 282 * pgs: Slab page count if a small bin size class, 0 otherwise. 305 # Iterate through page sizes and compute how many bins there are.
|
/fuchsia/zircon/kernel/vm/ |
H A D | vm_object_paged.cpp | 49 // round up the size to the next page size boundary and make sure we dont wrap 98 // make sure size is page aligned 124 // make sure size is page aligned 213 vm_page_t* page = paddr_to_vm_page(pa); 214 ASSERT(page); 216 if (page->state == VM_PAGE_STATE_WIRED) { 218 } else if (page->state == VM_PAGE_STATE_FREE) { 221 page->state = VM_PAGE_STATE_WIRED; 223 panic("page used to back static vmo in unusable state: paddr %#" PRIxPTR " state %u\n", pa, 224 page [all...] |
H A D | vm_unittest.cpp | 23 // Allocates a single page, translates it to a vm_page_t and frees it. 27 vm_page_t* page; local 29 zx_status_t status = pmm_alloc_page(0, &page, &pa); 30 ASSERT_EQ(ZX_OK, status, "pmm_alloc single page"); 31 ASSERT_NE(nullptr, page, "pmm_alloc single page"); 32 ASSERT_NE(0u, pa, "pmm_alloc single page"); 35 ASSERT_EQ(page2, page, "paddr_to_vm_page on single page"); 37 pmm_free_page(page); [all...] |
H A D | vm_mapping.cpp | 397 // Add a page to the mapping run. If this fails, the VmMappingCoalescer is 487 // if committing, then tell it to soft fault in a page 501 // iterate through the range, grabbing a page from the underlying object and 512 // no page to map 514 // fail when we can't commit every requested page 616 // user page fault on non user mapped region 641 // via UnmapVmoRangeLocked(). Since we're responsible for that page, signal to ourself to skip 647 // fault in or grab an existing page 649 vm_page_t* page; local 650 zx_status_t status = object_->GetPageLocked(vmo_offset, pf_flags, nullptr, &page, [all...] |
/fuchsia/zircon/kernel/arch/arm64/hypervisor/ |
H A D | vmexit.cpp | 141 // TODO(abdulla): Make this understand concatenated page tables. 143 pte_t page = index_shift > MMU_GUEST_PAGE_SIZE_SHIFT ? local 148 if (desc == page) {
|
/fuchsia/zircon/kernel/arch/x86/ |
H A D | rules.mk | 121 GLOBAL_LDFLAGS += -z max-page-size=4096
|
/fuchsia/zircon/system/ulib/hid-parser/ |
H A D | parser.cpp | 228 Usage {table_.attributes.usage.page, usage}, 319 table_.attributes.usage.page =
|
/fuchsia/zircon/kernel/arch/x86/hypervisor/ |
H A D | vmexit.cpp | 813 /* Returns the page address for a given page table entry. 815 * If the page address is for a large page, we additionally calculate the offset 816 * to the correct guest physical page that backs the large page. 867 uint8_t* page = static_cast<uint8_t*>(paddr_to_physmap(pa)); local 869 mandatory_memcpy(data, page + page_offset, from_page); 879 page = static_cast<uint8_t*>(paddr_to_physmap(pa)); 880 mandatory_memcpy(data + from_page, page, siz [all...] |
/fuchsia/zircon/third_party/ulib/jemalloc/test/unit/ |
H A D | mallctl.c | 562 TEST_ARENAS_CONSTANT(size_t, page, PAGE);
|
/fuchsia/zircon/system/dev/block/usb-mass-storage/ |
H A D | usb-mass-storage.c | 247 command.page = 0x3F; // all pages, current values
|