Searched refs:pages (Results 1 - 25 of 70) sorted by relevance

123

/barrelfish-2018-10-04/lib/devif/backends/net/mlx4/include/linux/
H A Dvmalloc.h38 void *vmap(struct page **pages, unsigned int count, unsigned long flags,
/barrelfish-2018-10-04/usr/eclipseclp/Shm/src/
H A Dalloc.c124 * Management of free pages:
126 * We maintain a bitmap pages->map[] of all memory pages.
129 * Additionally, we have free lists pages->free[] for page clusters.
130 * pages->free[i] holds a list of i-page cluster descriptor.
131 * pages->free[0] holds a list of clusters of size PAGE_LISTS or larger.
147 * not be in the pages themselves, that causes unnecessary page accesses.
173 if (! pages->map[Block]) \
174 pages->map[Block] = _new_bitmap_block(hd); \
175 Ptr = &(pages
215 struct page_admin *pages = hd->pages; local
342 struct page_admin *pages = hd->pages; local
409 struct page_admin *pages = hd->pages; local
550 struct page_admin *pages = hd->pages; local
1016 struct page_admin *pages = hd->pages; local
[all...]
H A Dprivate_mem.c82 struct page_admin pages; member in struct:prmem
157 hd->pages = &private_memory->pages;
182 prmem_struct_addr = (char*)hd->pages - ((char*)&dummy.pages - (char*)&dummy);
H A Dshared_mem.c204 hd->pages = &hd->shared_header->pages;
282 hd->pages = &hd->shared_header->pages;
H A Dmemman.h117 #define MIN_OS_PAGE_REQUEST 8 /* min pages to get from OS */
134 word allocated; /* # pages gotten from OS */
135 word freed; /* # pages in free list */
137 struct page_log *log_page; /* log of more'd pages */
141 bits32 *map[BITMAP_BLOCKS]; /* bitmap of pages (1 = free) */
230 struct page_admin *pages; member in struct:heap_descriptor
258 struct page_admin pages; /* page manager structure */ member in struct:shm_desc
/barrelfish-2018-10-04/kernel/include/
H A Dpaging_generic.h33 errval_t paging_tlb_flush_range(struct cte *frame, size_t offset, size_t pages);
H A Dcapabilities.h133 size_t pages, size_t mflags,
136 size_t pages, size_t mflags);
138 uintptr_t pages, uintptr_t kpi_paging_flags);
/barrelfish-2018-10-04/include/vm/
H A Dvm_glue.c197 * The limit for transient usage of wired pages should be
198 * larger than for "permanent" wired pages (mlock()).
321 vm_thread_new(struct thread *td, int pages) argument
330 if (pages <= 1)
331 pages = kstack_pages;
332 else if (pages > KSTACK_MAX_PAGES)
333 pages = KSTACK_MAX_PAGES;
335 if (pages == kstack_pages) {
353 ksobj = vm_object_allocate(OBJT_DEFAULT, pages);
363 if (vmem_xalloc(kernel_arena, (pages
409 vm_thread_stack_dispose(vm_object_t ksobj, vm_offset_t ks, int pages) argument
441 int pages; local
544 int i, pages; local
571 int pages; local
[all...]
H A Dvm_reserv.c65 * pages ("superpages"). Speculative allocation enables the fully-automatic
73 * The number of small pages that are contained in a level 0 reservation
90 * within the reservation's array of small pages.
155 * physical pages for the range [pindex, pindex + VM_LEVEL_0_NPAGES) of offsets
157 * small physical pages that are in use at any given time. When and if the
169 vm_page_t pages; /* first page of a superpage */ member in struct:vm_reserv
170 int popcnt; /* # of pages in use */
172 popmap_t popmap[NPOPMAP]; /* bit vector of used pages */
184 * reservation's "pages" field. Invalid reservation structures have a NULL
185 * "pages" fiel
[all...]
H A Dvm_extern.h116 int vm_thread_new(struct thread *td, int pages);
/barrelfish-2018-10-04/kernel/arch/x86_64/
H A Dpage_mappings_arch.c169 // TODO: cleanup already mapped pages
261 // TODO: cleanup already mapped pages
376 * \arg pages the number of pages to modify
382 size_t pages,
391 for (int i = 0; i < pages; i++) {
400 for (int i = 0; i < pages; i++) {
410 for (int i = 0; i < pages; i++) {
431 * \arg pages the number of pages t
381 generic_modify_flags(struct cte *leaf_pt, size_t offset, size_t pages, paging_x86_64_flags_t flags) argument
436 page_mappings_modify_flags(struct capability *mapping, size_t offset, size_t pages, size_t mflags, genvaddr_t va_hint) argument
498 ptable_modify_flags(struct capability *leaf_pt, size_t offset, size_t pages, size_t mflags) argument
[all...]
/barrelfish-2018-10-04/kernel/arch/armv8/
H A Dinit.c68 uint64_t pages = ARMV8_CORE_DATA_PAGES; local
72 if (desc->Type == EfiConventionalMemory && desc->NumberOfPages > pages) {
74 pages = desc->NumberOfPages;
H A Dpaging.c729 uintptr_t pages, uintptr_t kpi_paging_flags)
741 for (int i = 0; i < pages; i++) {
747 return paging_tlb_flush_range(cte_for_cap(mapping), 0, pages);
728 paging_modify_flags(struct capability *mapping, uintptr_t offset, uintptr_t pages, uintptr_t kpi_paging_flags) argument
/barrelfish-2018-10-04/kernel/include/arch/k1om/kernel_boot_param/
H A Dscreen_info.h59 uint16_t pages; /* 0x32 */ member in struct:screen_info
/barrelfish-2018-10-04/lib/libc/stdlib/jemalloc/
H A DMakefile.inc7 mutex.c nstime.c pages.c prng.c prof.c quarantine.c rtree.c stats.c \
/barrelfish-2018-10-04/kernel/arch/x86_32/
H A Dpage_mappings_arch.c315 size_t pages, size_t mflags, genvaddr_t va_hint)
335 if (offset + pages > X86_32_PTABLE_SIZE) { // mapping size ok
354 for (int i = 0; i < pages; i++) {
361 for (int i = 0; i < pages; i++) {
374 // invlpg should work for large/huge pages
375 for (int i = 0; i < pages; i++) {
314 page_mappings_modify_flags(struct capability *mapping, size_t offset, size_t pages, size_t mflags, genvaddr_t va_hint) argument
/barrelfish-2018-10-04/usr/drivers/usb/usb_manager/controller/ehci/
H A Dusb_ehci_xfer.c359 struct usb_dma_page *pages; ///< pointer to a set of dma pages member in struct:usb_ehci_qtd_setup_param
360 uint16_t num_pages; ///< the number of dma pages that are needed
480 td->qtd_bp[pages_count].address = (setup->pages->phys_addr
483 assert(!((setup->pages->phys_addr + buf_offset) & (0xFFF)));
490 td->qtd_bp[pages_count].address = (setup->pages->phys_addr
606 setup.pages = xfer->frame_buffers[0];
629 setup.pages = xfer->frame_buffers[frames];
684 setup.pages = NULL;
/barrelfish-2018-10-04/kernel/
H A Dpaging_generic.c311 debug(SUBSYS_PAGING, "page_mappings_unmap(%hu pages)\n", info->pte_count);
335 // flush TLB for unmapped pages if we got a valid virtual address
350 errval_t paging_tlb_flush_range(struct cte *mapping_cte, size_t offset, size_t pages) argument
377 vaddr, vaddr+(pages * BASE_PAGE_SIZE));
378 // flush TLB entries for all modified pages
416 // TODO: check what tlb flushing instructions expect for large/huge pages
417 for (int i = 0; i < pages; i++) {
/barrelfish-2018-10-04/include/
H A Dnuma.h7 * Linux man pages "numa"
180 * XXX according to the man pages this function is only used internally
342 * should only be used for large areas consisting of multiple pages.
357 * should only be used for large areas consisting of multiple pages.
726 * \brief moves a list of pages in the address space of the current domain
729 * \param count number of pages to move
730 * \param pages list of pages
731 * \param nodes list of nodes to which the pages can be moved
733 * \param flags flags for moving the pages
[all...]
/barrelfish-2018-10-04/lib/barrelfish/
H A Dheap.c126 * \brief Allocate and map in one or more pages of memory
128 static void *pages_alloc(size_t pages) argument
152 // Allocate requested number of pages and insert freelist header
/barrelfish-2018-10-04/doc/015-disk-driver-arch/
H A Dconclusion.tex31 cache that stores pages or larger chunks of data. A read operation would then
/barrelfish-2018-10-04/include/barrelfish/
H A Dinvocations.h240 * \param off Offset (in #pages) of the first page to get new set of flags
242 * \param pages Number of pages that should get new set of flags
250 size_t pages,
255 pages, flags, va_hint).error;
248 invoke_mapping_modify_flags(struct capref mapping, size_t offset, size_t pages, size_t flags, genvaddr_t va_hint) argument
/barrelfish-2018-10-04/lib/barrelfish/arch/aarch64/
H A Dpmap.c19 * resource bootstrapping. The bootstrap ram allocator allocates pages.
28 * allocating single pages and the difficulty in bootstrapping
458 size_t pages = DIVIDE_ROUND_UP(bytes, BASE_PAGE_SIZE); local
461 size_t l3entries = DIVIDE_ROUND_UP(pages, 512);
472 return pages + l3entries + l2entries + l1entries + l0entries;
729 size_t pages, vregion_flags_t flags)
737 if (inside_region(ptable, ptentry, pages)) {
741 // pages, new flags. Invocation should check compatibility of
746 err = invoke_mapping_modify_flags(page->mapping, off, pages, pmap_flags, 0);
728 do_single_modify_flags(struct pmap_aarch64 *pmap, genvaddr_t vaddr, size_t pages, vregion_flags_t flags) argument
/barrelfish-2018-10-04/lib/barrelfish/arch/x86_32/
H A Dpmap.c39 // flags for large pages
232 //4M/2M (PAE) pages
421 // 4M pages/2M pages(PAE)
427 // 4K pages
476 if (page && page->is_vnode) { // not 2M/4M pages
626 * \param pages number of pages to modify
630 size_t pages, vregion_flags_t flags)
650 if (inside_region(pt, ptentry, pages)) {
629 do_single_modify_flags(struct pmap_x86 *pmap, genvaddr_t vaddr, size_t pages, vregion_flags_t flags) argument
711 size_t pages = size / page_size; local
[all...]
/barrelfish-2018-10-04/kernel/arch/armv7/
H A Dpaging.c610 uintptr_t pages, uintptr_t kpi_paging_flags)
623 for (int i = 0; i < pages; i++) {
632 return paging_tlb_flush_range(cte_for_cap(mapping), offset, pages);
609 paging_modify_flags(struct capability *mapping, uintptr_t offset, uintptr_t pages, uintptr_t kpi_paging_flags) argument

Completed in 264 milliseconds

123