Searched refs:pages (Results 101 - 125 of 162) sorted by relevance

1234567

/freebsd-13-stable/sys/dev/hptrr/
H A Dhim.h336 HPT_ADDRESS *pages; member in struct:ctl_pages
/freebsd-13-stable/sys/kern/
H A Dkern_thread.c745 thread_alloc(int pages) argument
757 if (!vm_thread_new(td, pages)) {
770 thread_alloc_stack(struct thread *td, int pages) argument
775 if (!vm_thread_new(td, pages))
H A Dvfs_aio.c1286 struct vm_page** pages; local
1301 pages = malloc(sizeof(vm_page_t) * (atop(round_page(
1306 pages = pbuf->b_pages;
1322 (vm_offset_t)buf, bp->bio_length, prot, pages,
1328 free(pages, M_TEMP);
1335 pmap_qenter((vm_offset_t)pbuf->b_data, pages, npages);
1340 bp->bio_ma = pages;
/freebsd-13-stable/sys/dev/usb/wlan/
H A Dif_upgtvar.h369 uint8_t pages; member in struct:upgt_memory
H A Dif_upgt.c848 for (i = 0; i < sc->sc_memory.pages; i++) {
1540 for (i = 0; i < sc->sc_memory.pages; i++) {
1727 sc->sc_memory.pages = i;
1729 DPRINTF(sc, UPGT_DEBUG_FW, "memory pages=%d\n", sc->sc_memory.pages);
/freebsd-13-stable/contrib/mandoc/
H A Dmandocdb.c182 static struct mpage *mpage_head; /* list of distinct manual pages */
183 static struct ohash mpages; /* table of distinct manual pages */
1380 * Bad pages are not worth better heuristics.
1529 * Bad pages are not worth better heuristics.
2087 mpage->dba = dba_page_new(dba->pages,
2125 dba_array_FOREACH(dba->pages, page) {
2134 dba_array_del(dba->pages);
2157 dba_array_start(dba->pages);
2158 if (dba_array_next(dba->pages) == NULL) {
/freebsd-13-stable/sys/contrib/vchiq/interface/vchiq_arm/
H A Dvchiq_arm.c1504 struct page **pages; local
1518 pages = kmalloc(sizeof(struct page *) * num_pages, GFP_KERNEL);
1519 if (pages == NULL) {
1521 "Unable to allocation memory for %d pages\n",
1533 pages, /* pages (array of page pointers) */
1549 page = pages[page_idx];
1567 page_cache_release(pages[page_idx]);
1569 kfree(pages);
/freebsd-13-stable/sys/dev/drm2/ttm/
H A Dttm_bo_vm.c223 m = ttm->pages[OFF_TO_IDX(offset)];
/freebsd-13-stable/sys/mips/mips/
H A Dpmap.c202 * At bootup, we reserve 2 virtual pages per CPU for mapping highmem pages. To
211 vm_offset_t pages; local
225 pages = kva_alloc(PAGE_SIZE * 3);
226 if (pages == 0)
228 pc->pc_cmap1_ptep = pmap_pte(kernel_pmap, pages);
229 pc->pc_cmap2_ptep = pmap_pte(kernel_pmap, pages + PAGE_SIZE);
231 pmap_pte(kernel_pmap, pages + (PAGE_SIZE * 2));
232 pc->pc_cmap1_addr = pages;
233 pc->pc_cmap2_addr = pages
[all...]
/freebsd-13-stable/stand/libsa/zfs/
H A Dzfs.c1805 int rv, pages, perpage, currpage; local
1831 /* Calculate and store the number of pages of BEs */
1833 pages = (zfs_env_count / perpage) + ((zfs_env_count % perpage) > 0 ? 1 : 0);
1834 snprintf(becount, 4, "%d", pages);
1840 if (currpage > pages) {
/freebsd-13-stable/sys/amd64/include/
H A Dpmap.h199 * We use the same numbering of the page table pages for 5-level and
203 pages */
205 pages */
207 pages */
372 struct vm_radix pm_root; /* spare page table pages */
486 void pmap_invalidate_cache_pages(vm_page_t *pages, int count);
/freebsd-13-stable/contrib/tnftp/src/
H A DMakefile.in645 echo "error: found man pages containing the \`missing help2man' replacement text:" >&2; \
647 echo " to fix them, install help2man, remove and regenerate the man pages;" >&2; \
/freebsd-13-stable/sbin/nvmecontrol/modules/wdc/
H A Dwdc.c786 int pages; local
794 pages = *walker++;
/freebsd-13-stable/contrib/bmake/mk/
H A Down.mk153 # Define MANZ to have the man pages compressed (gzip)
/freebsd-13-stable/contrib/bc/
H A Dconfigure57 printf ' [--disable-history --disable-man-pages --disable-nls] \\\n'
116 printf ' -M, --disable-man-pages\n'
600 disable-man-pages) install_manpages=0 ;;
617 disable-man-pages* | disable-nls* | disable-strip*)
H A Dconfigure.sh57 printf ' [--disable-history --disable-man-pages --disable-nls] \\\n'
116 printf ' -M, --disable-man-pages\n'
600 disable-man-pages) install_manpages=0 ;;
617 disable-man-pages* | disable-nls* | disable-strip*)
/freebsd-13-stable/sys/contrib/openzfs/module/os/linux/zfs/
H A Dzpl_file.c534 * both the data in the ARC and existing pages in the page cache
537 * will ever result in new pages being added to the page cache.
539 * New pages are added to the page cache only via .readpage() which
541 * virtual memory region. These pages may be modified without
555 * mapping page cache pages directly on to the ARC buffers. The
557 * selection of which pages back a particular index. The trick
624 * Populate a set of pages with data for the Linux page cache. This
631 struct list_head *pages, unsigned nr_pages)
633 return (read_cache_pages(mapping, pages,
687 * pages (se
630 zpl_readpages(struct file *filp, struct address_space *mapping, struct list_head *pages, unsigned nr_pages) argument
[all...]
/freebsd-13-stable/sys/dev/agp/
H A Dagp.c537 * Allocate the pages early, before acquiring the lock,
548 * the pages will be allocated and zeroed.
567 * Bind the individual pages and flush the chipset's
589 * and unwire the pages.
641 * Unbind the individual pages and flush the chipset's
642 * TLB. Unwire the pages so they can be swapped.
983 agp_bind_pages(device_t dev, vm_page_t *pages, vm_size_t size, argument
999 m = pages[OFF_TO_IDX(i)];
/freebsd-13-stable/crypto/openssh/contrib/suse/
H A Dopenssh.spec98 - Marked all manual pages as %doc(umentation)
/freebsd-13-stable/sys/powerpc/aim/
H A Dmmu_radix.c263 /* 1GB pages aren't currently supported. */
401 "Number of kernel page table pages allocated on bootup");
465 void mmu_radix_page_array_startup(long pages);
642 #define NUPML1E (RADIX_PGD_SIZE/sizeof(uint64_t)) /* number of userland PML1 pages */
643 #define NUPDPE (NUPML1E * NL2EPG)/* number of userland PDP pages */
1355 * drastic measures to free some pages so we can allocate
1834 vm_paddr_t ptpages, pages; local
1856 pages = ptpages;
1858 *pte = htobe64(pages | RPTE_VALID | RPTE_SHIFT);
1859 pages
6319 mmu_radix_page_array_startup(long pages) argument
[all...]
/freebsd-13-stable/sys/dev/oce/
H A Doce_mbox.c1663 /* oce_page_list will also prepare pages */
1664 num_pages = oce_page_list(rq->ring, &fwcmd->params.req.pages[0]);
1726 num_pages = oce_page_list(wq->ring, &fwcmd->params.req.pages[0]);
1776 num_pages = oce_page_list(eq->ring, &fwcmd->params.req.pages[0]);
1831 num_pages = oce_page_list(cq->ring, &fwcmd->params.req.pages[0]);
2296 /* oce_page_list will also prepare pages */
2297 num_pages = oce_page_list(rq->ring, &fwcmd->params.req.pages[0]);
/freebsd-13-stable/sys/arm/arm/
H A Dpmap-v6.c766 * it saves physical pages and KVA per a process, and it's simple change.
820 * preallocated pages for kernel L2 page tables so that vm_page
821 * structures representing these pages will be created. The vm_page
828 * Insert allocated L2 page table pages to PT2TAB and make
853 * Note that pages holding PT2s are stored in PT2TAB as pt2_entry_t
910 * Note that we have allocated NKPT2PG pages for L2 page tables in advance
991 /* Map all the pages. */
1065 /* Map all the pages. */
1204 * mapping of pages.
1212 * Local CMAP1/CMAP2 are used for zeroing and copying pages
1255 vm_offset_t pages; local
[all...]
/freebsd-13-stable/sys/dev/mthca/
H A Dmthca_cmd.c657 __be64 *pages; local
668 pages = mailbox->buf;
674 * We have to pass pages that are aligned to their
689 pages[nent * 2] = cpu_to_be64(virt);
693 pages[nent * 2 + 1] =
852 * Round up number of system pages needed in case
1564 mthca_dbg(dev, "Unmapping %d pages at %llx from ICM.\n",
1590 * Round up number of system pages needed in case
/freebsd-13-stable/sys/dev/drm2/
H A Ddrm_drv.c213 drm_free_agp(entry->memory, entry->pages);
/freebsd-13-stable/sys/dev/qlnx/qlnxr/
H A Dqlnxr_def.h148 * An S/RQ PBL contains a list a pointers to pages. Each page contains S/RQE
177 * calculation that comes from having a u16 for the number of pages i.e. 0xffff
178 * is the maximum number of pages (in single layer).
732 u64 *pages; member in struct:qlnxr_mr

Completed in 391 milliseconds

1234567