/freebsd-13-stable/sys/dev/hptrr/ |
H A D | him.h | 336 HPT_ADDRESS *pages; member in struct:ctl_pages
|
/freebsd-13-stable/sys/kern/ |
H A D | kern_thread.c | 745 thread_alloc(int pages) argument 757 if (!vm_thread_new(td, pages)) { 770 thread_alloc_stack(struct thread *td, int pages) argument 775 if (!vm_thread_new(td, pages))
|
H A D | vfs_aio.c | 1286 struct vm_page** pages; local 1301 pages = malloc(sizeof(vm_page_t) * (atop(round_page( 1306 pages = pbuf->b_pages; 1322 (vm_offset_t)buf, bp->bio_length, prot, pages, 1328 free(pages, M_TEMP); 1335 pmap_qenter((vm_offset_t)pbuf->b_data, pages, npages); 1340 bp->bio_ma = pages;
|
/freebsd-13-stable/sys/dev/usb/wlan/ |
H A D | if_upgtvar.h | 369 uint8_t pages; member in struct:upgt_memory
|
H A D | if_upgt.c | 848 for (i = 0; i < sc->sc_memory.pages; i++) { 1540 for (i = 0; i < sc->sc_memory.pages; i++) { 1727 sc->sc_memory.pages = i; 1729 DPRINTF(sc, UPGT_DEBUG_FW, "memory pages=%d\n", sc->sc_memory.pages);
|
/freebsd-13-stable/contrib/mandoc/ |
H A D | mandocdb.c | 182 static struct mpage *mpage_head; /* list of distinct manual pages */ 183 static struct ohash mpages; /* table of distinct manual pages */ 1380 * Bad pages are not worth better heuristics. 1529 * Bad pages are not worth better heuristics. 2087 mpage->dba = dba_page_new(dba->pages, 2125 dba_array_FOREACH(dba->pages, page) { 2134 dba_array_del(dba->pages); 2157 dba_array_start(dba->pages); 2158 if (dba_array_next(dba->pages) == NULL) {
|
/freebsd-13-stable/sys/contrib/vchiq/interface/vchiq_arm/ |
H A D | vchiq_arm.c | 1504 struct page **pages; local 1518 pages = kmalloc(sizeof(struct page *) * num_pages, GFP_KERNEL); 1519 if (pages == NULL) { 1521 "Unable to allocation memory for %d pages\n", 1533 pages, /* pages (array of page pointers) */ 1549 page = pages[page_idx]; 1567 page_cache_release(pages[page_idx]); 1569 kfree(pages);
|
/freebsd-13-stable/sys/dev/drm2/ttm/ |
H A D | ttm_bo_vm.c | 223 m = ttm->pages[OFF_TO_IDX(offset)];
|
/freebsd-13-stable/sys/mips/mips/ |
H A D | pmap.c | 202 * At bootup, we reserve 2 virtual pages per CPU for mapping highmem pages. To 211 vm_offset_t pages; local 225 pages = kva_alloc(PAGE_SIZE * 3); 226 if (pages == 0) 228 pc->pc_cmap1_ptep = pmap_pte(kernel_pmap, pages); 229 pc->pc_cmap2_ptep = pmap_pte(kernel_pmap, pages + PAGE_SIZE); 231 pmap_pte(kernel_pmap, pages + (PAGE_SIZE * 2)); 232 pc->pc_cmap1_addr = pages; 233 pc->pc_cmap2_addr = pages [all...] |
/freebsd-13-stable/stand/libsa/zfs/ |
H A D | zfs.c | 1805 int rv, pages, perpage, currpage; local 1831 /* Calculate and store the number of pages of BEs */ 1833 pages = (zfs_env_count / perpage) + ((zfs_env_count % perpage) > 0 ? 1 : 0); 1834 snprintf(becount, 4, "%d", pages); 1840 if (currpage > pages) {
|
/freebsd-13-stable/sys/amd64/include/ |
H A D | pmap.h | 199 * We use the same numbering of the page table pages for 5-level and 203 pages */ 205 pages */ 207 pages */ 372 struct vm_radix pm_root; /* spare page table pages */ 486 void pmap_invalidate_cache_pages(vm_page_t *pages, int count);
|
/freebsd-13-stable/contrib/tnftp/src/ |
H A D | Makefile.in | 645 echo "error: found man pages containing the \`missing help2man' replacement text:" >&2; \ 647 echo " to fix them, install help2man, remove and regenerate the man pages;" >&2; \
|
/freebsd-13-stable/sbin/nvmecontrol/modules/wdc/ |
H A D | wdc.c | 786 int pages; local 794 pages = *walker++;
|
/freebsd-13-stable/contrib/bmake/mk/ |
H A D | own.mk | 153 # Define MANZ to have the man pages compressed (gzip)
|
/freebsd-13-stable/contrib/bc/ |
H A D | configure | 57 printf ' [--disable-history --disable-man-pages --disable-nls] \\\n' 116 printf ' -M, --disable-man-pages\n' 600 disable-man-pages) install_manpages=0 ;; 617 disable-man-pages* | disable-nls* | disable-strip*)
|
H A D | configure.sh | 57 printf ' [--disable-history --disable-man-pages --disable-nls] \\\n' 116 printf ' -M, --disable-man-pages\n' 600 disable-man-pages) install_manpages=0 ;; 617 disable-man-pages* | disable-nls* | disable-strip*)
|
/freebsd-13-stable/sys/contrib/openzfs/module/os/linux/zfs/ |
H A D | zpl_file.c | 534 * both the data in the ARC and existing pages in the page cache 537 * will ever result in new pages being added to the page cache. 539 * New pages are added to the page cache only via .readpage() which 541 * virtual memory region. These pages may be modified without 555 * mapping page cache pages directly on to the ARC buffers. The 557 * selection of which pages back a particular index. The trick 624 * Populate a set of pages with data for the Linux page cache. This 631 struct list_head *pages, unsigned nr_pages) 633 return (read_cache_pages(mapping, pages, 687 * pages (se 630 zpl_readpages(struct file *filp, struct address_space *mapping, struct list_head *pages, unsigned nr_pages) argument [all...] |
/freebsd-13-stable/sys/dev/agp/ |
H A D | agp.c | 537 * Allocate the pages early, before acquiring the lock, 548 * the pages will be allocated and zeroed. 567 * Bind the individual pages and flush the chipset's 589 * and unwire the pages. 641 * Unbind the individual pages and flush the chipset's 642 * TLB. Unwire the pages so they can be swapped. 983 agp_bind_pages(device_t dev, vm_page_t *pages, vm_size_t size, argument 999 m = pages[OFF_TO_IDX(i)];
|
/freebsd-13-stable/crypto/openssh/contrib/suse/ |
H A D | openssh.spec | 98 - Marked all manual pages as %doc(umentation)
|
/freebsd-13-stable/sys/powerpc/aim/ |
H A D | mmu_radix.c | 263 /* 1GB pages aren't currently supported. */ 401 "Number of kernel page table pages allocated on bootup"); 465 void mmu_radix_page_array_startup(long pages); 642 #define NUPML1E (RADIX_PGD_SIZE/sizeof(uint64_t)) /* number of userland PML1 pages */ 643 #define NUPDPE (NUPML1E * NL2EPG)/* number of userland PDP pages */ 1355 * drastic measures to free some pages so we can allocate 1834 vm_paddr_t ptpages, pages; local 1856 pages = ptpages; 1858 *pte = htobe64(pages | RPTE_VALID | RPTE_SHIFT); 1859 pages 6319 mmu_radix_page_array_startup(long pages) argument [all...] |
/freebsd-13-stable/sys/dev/oce/ |
H A D | oce_mbox.c | 1663 /* oce_page_list will also prepare pages */ 1664 num_pages = oce_page_list(rq->ring, &fwcmd->params.req.pages[0]); 1726 num_pages = oce_page_list(wq->ring, &fwcmd->params.req.pages[0]); 1776 num_pages = oce_page_list(eq->ring, &fwcmd->params.req.pages[0]); 1831 num_pages = oce_page_list(cq->ring, &fwcmd->params.req.pages[0]); 2296 /* oce_page_list will also prepare pages */ 2297 num_pages = oce_page_list(rq->ring, &fwcmd->params.req.pages[0]);
|
/freebsd-13-stable/sys/arm/arm/ |
H A D | pmap-v6.c | 766 * it saves physical pages and KVA per a process, and it's simple change. 820 * preallocated pages for kernel L2 page tables so that vm_page 821 * structures representing these pages will be created. The vm_page 828 * Insert allocated L2 page table pages to PT2TAB and make 853 * Note that pages holding PT2s are stored in PT2TAB as pt2_entry_t 910 * Note that we have allocated NKPT2PG pages for L2 page tables in advance 991 /* Map all the pages. */ 1065 /* Map all the pages. */ 1204 * mapping of pages. 1212 * Local CMAP1/CMAP2 are used for zeroing and copying pages 1255 vm_offset_t pages; local [all...] |
/freebsd-13-stable/sys/dev/mthca/ |
H A D | mthca_cmd.c | 657 __be64 *pages; local 668 pages = mailbox->buf; 674 * We have to pass pages that are aligned to their 689 pages[nent * 2] = cpu_to_be64(virt); 693 pages[nent * 2 + 1] = 852 * Round up number of system pages needed in case 1564 mthca_dbg(dev, "Unmapping %d pages at %llx from ICM.\n", 1590 * Round up number of system pages needed in case
|
/freebsd-13-stable/sys/dev/drm2/ |
H A D | drm_drv.c | 213 drm_free_agp(entry->memory, entry->pages);
|
/freebsd-13-stable/sys/dev/qlnx/qlnxr/ |
H A D | qlnxr_def.h | 148 * An S/RQ PBL contains a list a pointers to pages. Each page contains S/RQE 177 * calculation that comes from having a u16 for the number of pages i.e. 0xffff 178 * is the maximum number of pages (in single layer). 732 u64 *pages; member in struct:qlnxr_mr
|