/netbsd-6-1-5-RELEASE/sys/arch/hpcsh/include/ |
H A D | kloader.h | 31 #define PG_VADDR(pg) SH3_PHYS_TO_P1SEG(VM_PAGE_TO_PHYS(pg))
|
/netbsd-6-1-5-RELEASE/external/gpl3/gcc/dist/libmudflap/testsuite/libmudflap.c/ |
H A D | pass51-frag.c | 21 unsigned pg = getpagesize (); local 24 p = mmap (NULL, 4 * pg, PROT_READ|PROT_WRITE, 29 memset (p, 0, 4*pg); 30 rc = munmap (p, pg); 32 memset (p+pg, 0, 3*pg); 33 rc = munmap (p+pg, pg); 35 memset (p+2*pg, 0, 2*pg); [all...] |
/netbsd-6-1-5-RELEASE/sys/arch/landisk/include/ |
H A D | kloader.h | 32 #define PG_VADDR(pg) SH3_PHYS_TO_P1SEG(VM_PAGE_TO_PHYS(pg))
|
/netbsd-6-1-5-RELEASE/sys/arch/dreamcast/include/ |
H A D | kloader.h | 31 #define PG_VADDR(pg) SH3_PHYS_TO_P1SEG(VM_PAGE_TO_PHYS(pg))
|
/netbsd-6-1-5-RELEASE/sys/arch/hpcmips/include/ |
H A D | kloader.h | 31 #define PG_VADDR(pg) MIPS_PHYS_TO_KSEG0(VM_PAGE_TO_PHYS(pg))
|
/netbsd-6-1-5-RELEASE/sys/arch/evbsh3/include/ |
H A D | kloader.h | 32 #define PG_VADDR(pg) SH3_PHYS_TO_P1SEG(VM_PAGE_TO_PHYS(pg))
|
/netbsd-6-1-5-RELEASE/regress/sys/uvm/pdsim/ |
H A D | pdsim.c | 61 struct vm_page *pg; local 63 pg = TAILQ_FIRST(&freeq); 64 if (pg == NULL) { 67 TAILQ_REMOVE(&freeq, pg, pageq); 68 pg->offset = idx << PAGE_SHIFT; 69 pg->uanon = NULL; 70 pg->uobject = obj; 71 pg->pqflags = 0; 72 obj->pages[idx] = pg; 76 return pg; 80 pdsim_pagefree(struct vm_page *pg) argument 112 struct vm_page *pg; local 120 pdsim_pagemarkreferenced(struct vm_page *pg) argument 127 pmap_is_referenced(struct vm_page *pg) argument 134 pmap_clear_reference(struct vm_page *pg) argument 146 struct vm_page *pg; local 165 struct vm_page *pg; local 182 struct vm_page *pg; local [all...] |
/netbsd-6-1-5-RELEASE/sys/uvm/ |
H A D | uvm_pdpolicy_clockpro.c | 148 clockpro_setq(struct vm_page *pg, int qidx) argument 153 pg->pqflags = (pg->pqflags & ~PQ_QMASK) | (qidx * PQ_QFACTOR); 157 clockpro_getq(struct vm_page *pg) argument 161 qidx = (pg->pqflags & PQ_QMASK) / PQ_QFACTOR; 236 pageq_insert_tail(pageq_t *q, struct vm_page *pg) argument 239 TAILQ_INSERT_TAIL(&q->q_q, pg, pageq.queue); 245 pageq_insert_head(pageq_t *q, struct vm_page *pg) argument 248 TAILQ_INSERT_HEAD(&q->q_q, pg, pageq.queue); 254 pageq_remove(pageq_t *q, struct vm_page *pg) argument 268 struct vm_page *pg; local 282 clockpro_insert_tail(struct clockpro_state *s, int qidx, struct vm_page *pg) argument 292 clockpro_insert_head(struct clockpro_state *s, int qidx, struct vm_page *pg) argument 495 pageobj(struct vm_page *pg) argument 514 pageidx(struct vm_page *pg) argument 522 nonresident_pagelookupremove(struct vm_page *pg) argument 542 nonresident_pagerecord(struct vm_page *pg) argument 637 clockpro_movereferencebit(struct vm_page *pg, bool locked) argument 664 clockpro_clearreferencebit(struct vm_page *pg, bool locked) argument 676 struct vm_page *pg; local 724 clockpro___enqueuetail(struct vm_page *pg) argument 744 clockpro_pageenqueue(struct vm_page *pg) argument 803 clockpro_pagequeue(struct vm_page *pg) argument 815 clockpro_pagedequeue(struct vm_page *pg) argument 836 clockpro_pagerequeue(struct vm_page *pg) argument 843 pageq_remove(clockpro_queue(s, qidx), pg); local 851 handhot_endtest(struct vm_page *pg) argument 868 struct vm_page *pg; local 975 struct vm_page *pg; local 1090 uvmpdpol_pageactivate(struct vm_page *pg) argument 1108 uvmpdpol_pagedeactivate(struct vm_page *pg) argument 1115 uvmpdpol_pagedequeue(struct vm_page *pg) argument 1126 uvmpdpol_pageenqueue(struct vm_page *pg) argument 1179 uvmpdpol_pageisqueued_p(struct vm_page *pg) argument 1198 struct vm_page *pg; local 1212 struct vm_page *pg; local 1303 struct vm_page *pg; local 1402 struct vm_page *pg; local [all...] |
H A D | uvm_page.c | 173 const struct vm_page *pg = n; local 174 const voff_t a = pg->offset; 200 * => call should have already set pg's object and offset pointers 205 uvm_pageinsert_list(struct uvm_object *uobj, struct vm_page *pg, argument 209 KASSERT(uobj == pg->uobject); 211 KASSERT((pg->flags & PG_TABLED) == 0); 231 TAILQ_INSERT_AFTER(&uobj->memq, where, pg, listq.queue); 233 TAILQ_INSERT_TAIL(&uobj->memq, pg, listq.queue); 234 pg->flags |= PG_TABLED; 240 uvm_pageinsert_tree(struct uvm_object *uobj, struct vm_page *pg) argument 250 uvm_pageinsert(struct uvm_object *uobj, struct vm_page *pg) argument 266 uvm_pageremove_list(struct uvm_object *uobj, struct vm_page *pg) argument 296 uvm_pageremove_tree(struct uvm_object *uobj, struct vm_page *pg) argument 304 uvm_pageremove(struct uvm_object *uobj, struct vm_page *pg) argument 965 uvm_vm_page_to_phys(const struct vm_page *pg) argument 981 struct vm_page *pg; local 1116 struct vm_page *pg; local 1202 struct vm_page *pg; local 1404 uvm_pagerealloc(struct vm_page *pg, struct uvm_object *newobj, voff_t newoff) argument 1432 uvm_pagezerocheck(struct vm_page *pg) argument 1473 uvm_pagefree(struct vm_page *pg) argument 1624 struct vm_page *pg; local 1666 uvm_page_own(struct vm_page *pg, const char *tag) argument 1721 struct vm_page *pg; local 1824 struct vm_page *pg; local 1843 uvm_pagewire(struct vm_page *pg) argument 1867 uvm_pageunwire(struct vm_page *pg) argument 1887 uvm_pagedeactivate(struct vm_page *pg) argument 1903 uvm_pageactivate(struct vm_page *pg) argument 1925 uvm_pagedequeue(struct vm_page *pg) argument 1941 uvm_pageenqueue(struct vm_page *pg) argument 1959 uvm_pagezero(struct vm_page *pg) argument 1996 uvm_page_lookup_freelist(struct vm_page *pg) argument 2011 uvm_page_locked_p(struct vm_page *pg) argument 2033 uvm_page_printit(struct vm_page *pg, bool full, void (*pr)(const char *, ...)) argument 2119 struct vm_page *pg; local [all...] |
H A D | uvm_anon.c | 105 struct vm_page *pg = anon->an_page; local 118 if (pg && pg->loan_count) { 120 pg = uvm_anon_lockloanpg(anon); 127 if (pg) { 136 if (pg->uobject) { 138 KASSERT(pg->loan_count > 0); 139 pg->loan_count--; 140 pg->uanon = NULL; 142 mutex_exit(pg 263 struct vm_page *pg; local 342 struct vm_page *pg; local 433 struct vm_page *pg = anon->an_page; local [all...] |
H A D | uvm_pager.c | 306 struct vm_page *pg; local 318 pg = pgs[0]; 319 swap = (pg->uanon != NULL && pg->uobject == NULL) || 320 (pg->pqflags & PQ_AOBJ) != 0; 322 uobj = pg->uobject; 329 if (pg->uobject != NULL) { 330 swslot = uao_find_swslot(pg->uobject, 331 pg->offset >> PAGE_SHIFT); 333 KASSERT(pg 528 struct vm_page *pg; local [all...] |
H A D | uvm_loan.c | 57 * loans are tracked by pg->loan_count. an O->A page will have both 61 * each loan of a page to the kernel bumps the pg->wire_count. the 76 * locking: to read pg->loan_count either the owner or the page queues 77 * must be locked. to modify pg->loan_count, both the owner of the page 78 * and the PQs must be locked. pg->flags is (as always) locked by 343 struct vm_page *pg; local 356 pg = anon->an_page; 357 if (pg && (pg->pqflags & PQ_ANON) != 0 && anon->an_ref == 1) { 358 if (pg 453 struct vm_page *pg = pgpp[i]; local 540 struct vm_page *pg = *pgpp; local 618 struct vm_page *pg; local 837 struct vm_page *pg; local 963 struct vm_page *pg; local 1046 struct vm_page *pg; local 1109 struct vm_page *pg; local 1186 struct vm_page *pg; local [all...] |
H A D | uvm_pdpolicy_clock.c | 171 struct vm_page *pg; local 181 pg = TAILQ_FIRST(&pdpol_state.s_inactiveq); 184 pg = ss->ss_nextpg; 185 if (pg != NULL && (pg->pqflags & PQ_INACTIVE) == 0) { 186 pg = TAILQ_FIRST(&pdpol_state.s_inactiveq); 189 if (pg == NULL) { 192 ss->ss_nextpg = TAILQ_NEXT(pg, pageq.queue); 201 lock = uvmpd_trylockowner(pg); 203 if (pmap_is_referenced(pg)) { 294 uvmpdpol_pagedeactivate(struct vm_page *pg) argument 316 uvmpdpol_pageactivate(struct vm_page *pg) argument 326 uvmpdpol_pagedequeue(struct vm_page *pg) argument 345 uvmpdpol_pageenqueue(struct vm_page *pg) argument 357 uvmpdpol_pageisqueued_p(struct vm_page *pg) argument [all...] |
H A D | uvm_vnode.c | 233 struct vm_page *pg; local 246 pg = uvm_pagelookup(uobj, offset); 249 if (pg == NULL) { 254 pg = uvm_pagealloc(uobj, offset, NULL, 256 if (pg == NULL) { 266 UVMHIST_LOG(ubchist, "alloced %p (color %u)", pg, 267 VM_PGCOLOR_BUCKET(pg), 0,0); 275 if ((pg->flags & PG_BUSY) != 0) { 280 pg->flags |= PG_WANTED; 281 UVMHIST_LOG(ubchist, "wait %p (color %u)", pg, [all...] |
/netbsd-6-1-5-RELEASE/external/gpl3/gcc/dist/gcc/testsuite/g++.dg/ext/arm-fp16/ |
H A D | fp16-param-1.C | 10 extern void (*pg) (__fp16 *);
|
H A D | fp16-return-1.C | 10 extern __fp16 *(*pg) (void);
|
/netbsd-6-1-5-RELEASE/external/gpl3/gcc/dist/gcc/testsuite/gcc.target/arm/ |
H A D | fp16-param-1.c | 10 extern void (*pg) (__fp16 *);
|
H A D | fp16-return-1.c | 10 extern __fp16 *(*pg) (void);
|
/netbsd-6-1-5-RELEASE/sys/arch/zaurus/include/ |
H A D | kloader.h | 45 #define PG_VADDR(pg) kloader_phystov(VM_PAGE_TO_PHYS(pg))
|
/netbsd-6-1-5-RELEASE/sys/arch/ia64/include/ |
H A D | pmap.h | 120 #define pmap_is_referenced(pg) \ 121 (((pg)->mdpage.pvh_attrs & PGA_REFERENCED) != 0) 122 #define pmap_is_modified(pg) \ 123 (((pg)->mdpage.pvh_attrs & PGA_MODIFIED) != 0) 162 #define VM_MDPAGE_INIT(pg) \ 164 TAILQ_INIT(&(pg)->mdpage.pv_list); \ 165 mutex_init(&(pg)->mdpage.pv_mutex, MUTEX_DEFAULT, IPL_NONE); \
|
/netbsd-6-1-5-RELEASE/sys/arch/xen/x86/ |
H A D | xen_bus_dma.c | 69 struct vm_page *pg, *pgnext; local 90 for (pg = mlistp->tqh_first; pg != NULL; pg = pg->pageq.queue.tqe_next) { 91 pa = VM_PAGE_TO_PHYS(pg); 128 pg = NULL; 133 for (pg = mlistp->tqh_first, i = 0; pg != NULL; pg [all...] |
/netbsd-6-1-5-RELEASE/sys/common/pmap/tlb/ |
H A D | vmpagemd.h | 86 #define VM_PAGEMD_PVLIST_LOCK(pg, list_change) \ 103 #define VM_MDPAGE_INIT(pg) \ 105 (pg)->mdpage.mdpg_first.pv_next = NULL; \ 106 (pg)->mdpage.mdpg_first.pv_pmap = NULL; \ 107 (pg)->mdpage.mdpg_first.pv_va = (pg)->phys_addr; \ 108 (pg)->mdpage.mdpg_attrs = 0; \ 109 VM_PAGEMD_PVLIST_LOCK_INIT(&(pg)->mdpage); \
|
/netbsd-6-1-5-RELEASE/external/bsd/openldap/dist/contrib/slapd-modules/acl/ |
H A D | posixgroup.c | 50 pg_t *pg; local 57 pg = ch_calloc( 1, sizeof( pg_t ) ); 59 pg->pg_style = style; 61 switch ( pg->pg_style ) { 63 rc = dnNormalize( 0, NULL, NULL, &pat, &pg->pg_pat, NULL ); 73 ber_dupbv( &pg->pg_pat, &pat ); 79 fname, lineno, style_strings[ pg->pg_style ] ); 123 *privp = (void *)pg; 127 (void)pg_dynacl_destroy( (void *)pg ); 137 pg_t *pg local 178 pg_t *pg = (pg_t *)priv; local 301 pg_t *pg = (pg_t *)priv; local [all...] |
/netbsd-6-1-5-RELEASE/sys/arch/hppa/include/ |
H A D | pmap.h | 66 #define PVF_MOD PTE_PROT(TLB_DIRTY) /* pg/mp is modified */ 67 #define PVF_REF PTE_PROT(TLB_REFTRAP) /* pg/mp (inv) is referenced */ 68 #define PVF_WRITE PTE_PROT(TLB_WRITE) /* pg/mp is writable */ 69 #define PVF_EXEC PTE_PROT(TLB_EXECUTE) /* pg/mp is executable */ 70 #define PVF_UNCACHEABLE PTE_PROT(TLB_UNCACHEABLE) /* pg/mp is uncacheable */ 136 #define pmap_clear_modify(pg) pmap_changebit(pg, 0, PTE_PROT(TLB_DIRTY)) 137 #define pmap_clear_reference(pg) \ 138 pmap_changebit(pg, PTE_PROT(TLB_REFTRAP), 0) 139 #define pmap_is_modified(pg) pmap_testbi 174 pmap_page_protect(struct vm_page *pg, vm_prot_t prot) argument [all...] |
/netbsd-6-1-5-RELEASE/lib/libc/db/btree/ |
H A D | bt_overflow.c | 86 pgno_t pg; local 90 memmove(&pg, p, sizeof(pg)); 95 if (pg == P_INVALID || sz == 0) 113 for (p = *buf;; p = (char *)p + nb, pg = h->nextpg) { 114 if ((h = mpool_get(t->bt_mp, pg, 0)) == NULL) 139 __ovfl_put(BTREE *t, const DBT *dbt, pgno_t *pg) argument 175 *pg = h->pgno; 199 pgno_t pg; local 203 (void)memmove(&pg, [all...] |