/fuchsia/zircon/kernel/dev/iommu/intel/ |
H A D | iommu_page.cpp | 14 IommuPage::IommuPage(vm_page_t* page, uintptr_t virt) : page_(page), virt_(virt) { argument 25 vm_page_t* page; local 26 zx_status_t status = pmm_alloc_page(0, &page); 30 page->state = VM_PAGE_STATE_IOMMU; 39 page->paddr(), 43 pmm_free_page(page); 49 *out = IommuPage(page, reinterpret_cast<uintptr_t>(vaddr));
|
H A D | context_table_state.cpp | 20 IommuPage page) 21 : parent_(parent), root_entry_(root_entry), page_(fbl::move(page)), 47 IommuPage page; local 48 zx_status_t status = IommuPage::AllocatePage(&page); 56 fbl::move(page))); 18 ContextTableState(uint8_t bus, bool extended, bool upper, IommuImpl* parent, volatile ds::RootEntrySubentry* root_entry, IommuPage page) argument
|
H A D | iommu_page.h | 49 IommuPage(vm_page_t* page, uintptr_t virt);
|
H A D | context_table_state.h | 54 volatile ds::RootEntrySubentry* root_entry, IommuPage page);
|
/fuchsia/zircon/kernel/vm/ |
H A D | pmm_node.cpp | 21 void set_state_alloc(vm_page* page) { argument 22 LTRACEF("page %p: prev state %s\n", page, page_state_to_string(page->state)); 24 DEBUG_ASSERT(page->state == VM_PAGE_STATE_FREE); 26 page->state = VM_PAGE_STATE_ALLOC; 84 vm_page *temp, *page; variable 85 list_for_every_entry_safe (list, page, temp, vm_page, queue_node) { 86 list_delete(&page->queue_node); 87 list_add_tail(&free_list_, &page 97 vm_page* page = list_remove_head_type(&free_list_, vm_page, queue_node); local 139 vm_page* page = list_remove_head_type(&free_list_, vm_page, queue_node); local 184 vm_page_t* page = a.FindSpecific(address); local 269 FreePageLocked(vm_page* page) argument 293 FreePage(vm_page* page) argument 303 vm_page* page = list_remove_head_type(list, vm_page, queue_node); local 360 vm_page* page; local 368 FreeFill(vm_page_t* page) argument 374 CheckFreeFill(vm_page_t* page) argument [all...] |
H A D | page.cpp | 14 #include <vm/page.h> 43 printf("page %p: address %#" PRIxPTR " state %s flags %#x\n", this, paddr(), 63 vm_page* page = reinterpret_cast<vm_page*>(argv[2].u); local 65 page->dump(); 71 vm_page* page = reinterpret_cast<vm_page*>(argv[2].u); local 73 paddr_t pa = page->paddr(); 76 printf("bad page or page not mapped in kernel space\n");
|
H A D | pmm.cpp | 62 zx_status_t pmm_alloc_page(uint alloc_flags, vm_page_t** page) { argument 63 return pmm_node.AllocPage(alloc_flags, page, nullptr); 66 zx_status_t pmm_alloc_page(uint alloc_flags, vm_page_t** page, paddr_t* pa) { argument 67 return pmm_node.AllocPage(alloc_flags, page, pa); 80 // if we're called with a single page, just fall through to the regular allocation routine 82 vm_page_t* page; local 83 zx_status_t status = pmm_node.AllocPage(alloc_flags, &page, pa); 87 list_add_tail(list, &page->queue_node); 98 void pmm_free_page(vm_page* page) { argument 99 pmm_node.FreePage(page); [all...] |
H A D | pmm_node.h | 28 paddr_t PageToPaddr(const vm_page_t* page) TA_NO_THREAD_SAFETY_ANALYSIS; 32 zx_status_t AllocPage(uint alloc_flags, vm_page_t** page, paddr_t* pa); 36 void FreePage(vm_page* page); 59 void FreePageLocked(vm_page* page) TA_REQ(lock_); 71 // page queues 79 void FreeFill(vm_page_t* page); 80 void CheckFreeFill(vm_page_t* page);
|
H A D | pmm_arena.h | 35 // Counts the number of pages in every state. For each page in the arena, 45 // return a pointer to a specific page 49 bool page_belongs_to_arena(const vm_page* page) const { 50 return (page->paddr() >= base() && page->paddr() < (base() + size()));
|
H A D | vm_page_list.cpp | 78 LTRACEF_LEVEL(2, "%p page %p, offset %#" PRIx64 " node_offset %#" PRIx64 " index %zu\n", this, p, offset, 81 // lookup the tree node that holds this page 110 // lookup the tree node that holds this page 126 // lookup the tree node that holds this page 132 // free this page 133 auto page = pln->RemovePage(index); 134 if (page) { 135 // if it was the last page in the node, remove the node from the tree 141 pmm_free_page(page); 155 // per page ge [all...] |
H A D | rules.mk | 22 $(LOCAL_DIR)/page.cpp \
|
/fuchsia/zircon/system/utest/libfzl/ |
H A D | mapped-vmo.cpp | 14 constexpr size_t page = PAGE_SIZE; member in namespace:__anon1432 21 zx_status_t status = fzl::MappedVmo::Create(page, vmo_name, &mapped_vmo); 26 EXPECT_EQ(mapped_vmo->GetSize(), page); 30 for (size_t i = 0; i < page; ++i) { 49 zx_status_t status = fzl::MappedVmo::Create(page, vmo_name, &mapped_vmo); 52 uint8_t bytes[page]; 53 memset(bytes, 0xff, page); 55 status = zx_vmo_read(mapped_vmo->GetVmo(), bytes, 0, page); 57 for (size_t i = 0; i < page; ++i) { 69 zx_status_t status = fzl::MappedVmo::Create(page, vmo_nam [all...] |
/fuchsia/zircon/system/ulib/ddk/ |
H A D | phys-iter.c | 19 // iter->page is index of page containing buf->vmo_offset, 20 // and iter->last_page is index of page containing buf->vmo_offset + buf->length 21 iter->page = 0; 41 zx_paddr_t phys = phys_addrs[iter->page]; 49 // end on a page boundary 57 // if vmo_offset is unaligned we need to adjust out_paddr, accumulate partial page length 58 // in return_length and skip to next page. 59 // we will make sure the range ends on a page boundary so we don't need to worry about 64 iter->page [all...] |
/fuchsia/zircon/system/uapp/nand-util/ |
H A D | aml.cpp | 40 for (uint32_t page = 0; page < info.pages_per_block; page++) { 41 if (memcmp(oob + page * info.oob_size, "nbbt", 4) != 0) {
|
/fuchsia/zircon/system/ulib/ddk/include/ddk/ |
H A D | phys-iter.h | 17 uint64_t vmo_offset; // offset into first page to start iterating on 26 uint64_t page; // index of page in buf->phys that contains offset member in struct:__anon917 27 uint64_t last_page; // last valid page index in buf->phys
|
/fuchsia/zircon/kernel/platform/pc/include/platform/ |
H A D | console.h | 16 void set_visual_page(int page); 17 void set_active_page(int page);
|
/fuchsia/zircon/kernel/vm/include/vm/ |
H A D | pmm.h | 11 #include <vm/page.h> 40 // Allocate a single page of physical memory. 50 // append the allocate page structures to the tail of the passed in list. 57 // Free a single page. 58 void pmm_free_page(vm_page_t* page) __NONNULL((1)); 66 // Counts the number of pages in every state. For every page in every arena,
|
/fuchsia/zircon/kernel/arch/x86/hypervisor/ |
H A D | vmx_cpu_state.cpp | 84 // unconditionally allocating a page is adequate. 90 VmxPage& page = (*pages)[cpu_num]; local 101 // Check that a page-walk length of 4 is supported. 138 // Setup VMXON page. 139 VmxRegion* region = page.VirtualAddress<VmxRegion>(); 143 zx_status_t status = vmxon(page.PhysicalAddress()); 174 for (auto& page : pages) { 175 zx_status_t status = page.Alloc(vmx_info, 0);
|
/fuchsia/zircon/system/dev/nand/nandpart/ |
H A D | aml-bad-block.cpp | 331 uint32_t page = 0; local 333 for (; page + bbt_page_count <= nand_info_.pages_per_block; page += bbt_page_count) { 336 zxlogf(TRACE, "Reading pages [%u, %u)\n", page, page + bbt_page_count); 337 const uint32_t nand_page = block_entry_->block * nand_info_.pages_per_block + page; 342 zxlogf(TRACE, "nandpart: Unable to read page %u\n", page); 349 zxlogf(TRACE, "nandpart: Page %u does not contain valid BBT entry\n", page + i); 358 zxlogf(TRACE, "BBT entry in pages (%u, %u] is valid\n", page, pag [all...] |
/fuchsia/zircon/system/utest/hid-parser/ |
H A D | hid-parser-test.cpp | 109 // First 3 fields are the buttons, with usages 1, 2, 3, in the button page. 113 EXPECT_EQ(fields[ix].attr.usage.page, hid::usage::Page::kButton); 128 EXPECT_EQ(fields[4].attr.usage.page, hid::usage::Page::kGenericDesktop); 138 EXPECT_EQ(fields[5].attr.usage.page, hid::usage::Page::kGenericDesktop); 152 EXPECT_EQ(collection->usage.page, hid::usage::Page::kGenericDesktop); 159 EXPECT_EQ(collection->usage.page, hid::usage::Page::kGenericDesktop); 204 // First 3 fields are the buttons, with usages 1, 2, 3, in the button page. 208 EXPECT_EQ(fields[ix].attr.usage.page, hid::usage::Page::kButton); 223 EXPECT_EQ(fields[4].attr.usage.page, hid::usage::Page::kGenericDesktop); 231 EXPECT_EQ(fields[5].attr.usage.page, hi [all...] |
/fuchsia/zircon/kernel/platform/pc/ |
H A D | console.cpp | 38 /* video page buffer */ 60 void set_visual_page(int page) { argument 61 unsigned short page_offset = page * VPAGE_SIZE; 62 visual_page = page; 70 void set_active_page(int page) { argument 73 curr_x = curs_x[page]; 74 curr_y = curs_y[page]; 75 active_page = page;
|
/fuchsia/zircon/kernel/arch/arm64/hypervisor/ |
H A D | el2_cpu_state_priv.h | 12 #include <hypervisor/page.h>
|
/fuchsia/zircon/kernel/object/include/object/ |
H A D | buffer_chain.h | 16 #include <vm/page.h> 30 // | page | | page | 100 // Construct a Buffer in each page and add them to a temporary list. 102 vm_page_t* page; local 103 list_for_every_entry (&pages, page, vm_page_t, queue_node) { 104 DEBUG_ASSERT(page->state == VM_PAGE_STATE_ALLOC); 105 page->state = VM_PAGE_STATE_IPC; 106 void* va = paddr_to_physmap(page->paddr());
|
/fuchsia/zircon/third_party/ulib/backtrace/ |
H A D | mmap.c | 97 void *page; local 141 /* Allocate a new page. */ 145 page = mmap (NULL, asksize, PROT_READ | PROT_WRITE, 147 if (page == MAP_FAILED) 156 backtrace_free (state, (char *) page + size, asksize - size, 159 ret = page;
|
/fuchsia/zircon/system/core/pwrbtn-monitor/ |
H A D | pwrbtn-monitor.cpp | 33 return u1.page == u2.page && u1.usage == u2.usage; 44 .page = static_cast<uint16_t>(hid::usage::Page::kGenericDesktop), 49 .page = static_cast<uint16_t>(hid::usage::Page::kGenericDesktop),
|