/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/linux/ |
H A D | vmalloc.h | 38 void *vmap(struct page **pages, unsigned int count, unsigned long flags,
|
/barrelfish-2018-10-04/usr/eclipseclp/Shm/src/ |
H A D | alloc.c | 124 * Management of free pages: 126 * We maintain a bitmap pages->map[] of all memory pages. 129 * Additionally, we have free lists pages->free[] for page clusters. 130 * pages->free[i] holds a list of i-page cluster descriptor. 131 * pages->free[0] holds a list of clusters of size PAGE_LISTS or larger. 147 * not be in the pages themselves, that causes unnecessary page accesses. 173 if (! pages->map[Block]) \ 174 pages->map[Block] = _new_bitmap_block(hd); \ 175 Ptr = &(pages 215 struct page_admin *pages = hd->pages; local 342 struct page_admin *pages = hd->pages; local 409 struct page_admin *pages = hd->pages; local 550 struct page_admin *pages = hd->pages; local 1016 struct page_admin *pages = hd->pages; local [all...] |
H A D | private_mem.c | 82 struct page_admin pages; member in struct:prmem 157 hd->pages = &private_memory->pages; 182 prmem_struct_addr = (char*)hd->pages - ((char*)&dummy.pages - (char*)&dummy);
|
H A D | shared_mem.c | 204 hd->pages = &hd->shared_header->pages; 282 hd->pages = &hd->shared_header->pages;
|
H A D | memman.h | 117 #define MIN_OS_PAGE_REQUEST 8 /* min pages to get from OS */ 134 word allocated; /* # pages gotten from OS */ 135 word freed; /* # pages in free list */ 137 struct page_log *log_page; /* log of more'd pages */ 141 bits32 *map[BITMAP_BLOCKS]; /* bitmap of pages (1 = free) */ 230 struct page_admin *pages; member in struct:heap_descriptor 258 struct page_admin pages; /* page manager structure */ member in struct:shm_desc
|
/barrelfish-2018-10-04/kernel/include/ |
H A D | paging_generic.h | 33 errval_t paging_tlb_flush_range(struct cte *frame, size_t offset, size_t pages);
|
H A D | capabilities.h | 133 size_t pages, size_t mflags, 136 size_t pages, size_t mflags); 138 uintptr_t pages, uintptr_t kpi_paging_flags);
|
/barrelfish-2018-10-04/include/vm/ |
H A D | vm_glue.c | 197 * The limit for transient usage of wired pages should be 198 * larger than for "permanent" wired pages (mlock()). 321 vm_thread_new(struct thread *td, int pages) argument 330 if (pages <= 1) 331 pages = kstack_pages; 332 else if (pages > KSTACK_MAX_PAGES) 333 pages = KSTACK_MAX_PAGES; 335 if (pages == kstack_pages) { 353 ksobj = vm_object_allocate(OBJT_DEFAULT, pages); 363 if (vmem_xalloc(kernel_arena, (pages 409 vm_thread_stack_dispose(vm_object_t ksobj, vm_offset_t ks, int pages) argument 441 int pages; local 544 int i, pages; local 571 int pages; local [all...] |
H A D | vm_reserv.c | 65 * pages ("superpages"). Speculative allocation enables the fully-automatic 73 * The number of small pages that are contained in a level 0 reservation 90 * within the reservation's array of small pages. 155 * physical pages for the range [pindex, pindex + VM_LEVEL_0_NPAGES) of offsets 157 * small physical pages that are in use at any given time. When and if the 169 vm_page_t pages; /* first page of a superpage */ member in struct:vm_reserv 170 int popcnt; /* # of pages in use */ 172 popmap_t popmap[NPOPMAP]; /* bit vector of used pages */ 184 * reservation's "pages" field. Invalid reservation structures have a NULL 185 * "pages" fiel [all...] |
H A D | vm_extern.h | 116 int vm_thread_new(struct thread *td, int pages);
|
/barrelfish-2018-10-04/kernel/arch/x86_64/ |
H A D | page_mappings_arch.c | 169 // TODO: cleanup already mapped pages 261 // TODO: cleanup already mapped pages 376 * \arg pages the number of pages to modify 382 size_t pages, 391 for (int i = 0; i < pages; i++) { 400 for (int i = 0; i < pages; i++) { 410 for (int i = 0; i < pages; i++) { 431 * \arg pages the number of pages t 381 generic_modify_flags(struct cte *leaf_pt, size_t offset, size_t pages, paging_x86_64_flags_t flags) argument 436 page_mappings_modify_flags(struct capability *mapping, size_t offset, size_t pages, size_t mflags, genvaddr_t va_hint) argument 498 ptable_modify_flags(struct capability *leaf_pt, size_t offset, size_t pages, size_t mflags) argument [all...] |
/barrelfish-2018-10-04/kernel/arch/armv8/ |
H A D | init.c | 68 uint64_t pages = ARMV8_CORE_DATA_PAGES; local 72 if (desc->Type == EfiConventionalMemory && desc->NumberOfPages > pages) { 74 pages = desc->NumberOfPages;
|
H A D | paging.c | 729 uintptr_t pages, uintptr_t kpi_paging_flags) 741 for (int i = 0; i < pages; i++) { 747 return paging_tlb_flush_range(cte_for_cap(mapping), 0, pages); 728 paging_modify_flags(struct capability *mapping, uintptr_t offset, uintptr_t pages, uintptr_t kpi_paging_flags) argument
|
/barrelfish-2018-10-04/kernel/include/arch/k1om/kernel_boot_param/ |
H A D | screen_info.h | 59 uint16_t pages; /* 0x32 */ member in struct:screen_info
|
/barrelfish-2018-10-04/lib/libc/stdlib/jemalloc/ |
H A D | Makefile.inc | 7 mutex.c nstime.c pages.c prng.c prof.c quarantine.c rtree.c stats.c \
|
/barrelfish-2018-10-04/kernel/arch/x86_32/ |
H A D | page_mappings_arch.c | 315 size_t pages, size_t mflags, genvaddr_t va_hint) 335 if (offset + pages > X86_32_PTABLE_SIZE) { // mapping size ok 354 for (int i = 0; i < pages; i++) { 361 for (int i = 0; i < pages; i++) { 374 // invlpg should work for large/huge pages 375 for (int i = 0; i < pages; i++) { 314 page_mappings_modify_flags(struct capability *mapping, size_t offset, size_t pages, size_t mflags, genvaddr_t va_hint) argument
|
/barrelfish-2018-10-04/usr/drivers/usb/usb_manager/controller/ehci/ |
H A D | usb_ehci_xfer.c | 359 struct usb_dma_page *pages; ///< pointer to a set of dma pages member in struct:usb_ehci_qtd_setup_param 360 uint16_t num_pages; ///< the number of dma pages that are needed 480 td->qtd_bp[pages_count].address = (setup->pages->phys_addr 483 assert(!((setup->pages->phys_addr + buf_offset) & (0xFFF))); 490 td->qtd_bp[pages_count].address = (setup->pages->phys_addr 606 setup.pages = xfer->frame_buffers[0]; 629 setup.pages = xfer->frame_buffers[frames]; 684 setup.pages = NULL;
|
/barrelfish-2018-10-04/kernel/ |
H A D | paging_generic.c | 311 debug(SUBSYS_PAGING, "page_mappings_unmap(%hu pages)\n", info->pte_count); 335 // flush TLB for unmapped pages if we got a valid virtual address 350 errval_t paging_tlb_flush_range(struct cte *mapping_cte, size_t offset, size_t pages) argument 377 vaddr, vaddr+(pages * BASE_PAGE_SIZE)); 378 // flush TLB entries for all modified pages 416 // TODO: check what tlb flushing instructions expect for large/huge pages 417 for (int i = 0; i < pages; i++) {
|
/barrelfish-2018-10-04/include/ |
H A D | numa.h | 7 * Linux man pages "numa" 180 * XXX according to the man pages this function is only used internally 342 * should only be used for large areas consisting of multiple pages. 357 * should only be used for large areas consisting of multiple pages. 726 * \brief moves a list of pages in the address space of the current domain 729 * \param count number of pages to move 730 * \param pages list of pages 731 * \param nodes list of nodes to which the pages can be moved 733 * \param flags flags for moving the pages [all...] |
/barrelfish-2018-10-04/lib/barrelfish/ |
H A D | heap.c | 126 * \brief Allocate and map in one or more pages of memory 128 static void *pages_alloc(size_t pages) argument 152 // Allocate requested number of pages and insert freelist header
|
/barrelfish-2018-10-04/doc/015-disk-driver-arch/ |
H A D | conclusion.tex | 31 cache that stores pages or larger chunks of data. A read operation would then
|
/barrelfish-2018-10-04/include/barrelfish/ |
H A D | invocations.h | 240 * \param off Offset (in #pages) of the first page to get new set of flags 242 * \param pages Number of pages that should get new set of flags 250 size_t pages, 255 pages, flags, va_hint).error; 248 invoke_mapping_modify_flags(struct capref mapping, size_t offset, size_t pages, size_t flags, genvaddr_t va_hint) argument
|
/barrelfish-2018-10-04/lib/barrelfish/arch/aarch64/ |
H A D | pmap.c | 19 * resource bootstrapping. The bootstrap ram allocator allocates pages. 28 * allocating single pages and the difficulty in bootstrapping 458 size_t pages = DIVIDE_ROUND_UP(bytes, BASE_PAGE_SIZE); local 461 size_t l3entries = DIVIDE_ROUND_UP(pages, 512); 472 return pages + l3entries + l2entries + l1entries + l0entries; 729 size_t pages, vregion_flags_t flags) 737 if (inside_region(ptable, ptentry, pages)) { 741 // pages, new flags. Invocation should check compatibility of 746 err = invoke_mapping_modify_flags(page->mapping, off, pages, pmap_flags, 0); 728 do_single_modify_flags(struct pmap_aarch64 *pmap, genvaddr_t vaddr, size_t pages, vregion_flags_t flags) argument
|
/barrelfish-2018-10-04/lib/barrelfish/arch/x86_32/ |
H A D | pmap.c | 39 // flags for large pages 232 //4M/2M (PAE) pages 421 // 4M pages/2M pages(PAE) 427 // 4K pages 476 if (page && page->is_vnode) { // not 2M/4M pages 626 * \param pages number of pages to modify 630 size_t pages, vregion_flags_t flags) 650 if (inside_region(pt, ptentry, pages)) { 629 do_single_modify_flags(struct pmap_x86 *pmap, genvaddr_t vaddr, size_t pages, vregion_flags_t flags) argument 711 size_t pages = size / page_size; local [all...] |
/barrelfish-2018-10-04/kernel/arch/armv7/ |
H A D | paging.c | 610 uintptr_t pages, uintptr_t kpi_paging_flags) 623 for (int i = 0; i < pages; i++) { 632 return paging_tlb_flush_range(cte_for_cap(mapping), offset, pages); 609 paging_modify_flags(struct capability *mapping, uintptr_t offset, uintptr_t pages, uintptr_t kpi_paging_flags) argument
|