Searched refs:pg (Results 1 - 25 of 256) sorted by relevance

1234567891011

/netbsd-6-1-5-RELEASE/sys/arch/hpcsh/include/
H A Dkloader.h31 #define PG_VADDR(pg) SH3_PHYS_TO_P1SEG(VM_PAGE_TO_PHYS(pg))
/netbsd-6-1-5-RELEASE/external/gpl3/gcc/dist/libmudflap/testsuite/libmudflap.c/
H A Dpass51-frag.c21 unsigned pg = getpagesize (); local
24 p = mmap (NULL, 4 * pg, PROT_READ|PROT_WRITE,
29 memset (p, 0, 4*pg);
30 rc = munmap (p, pg);
32 memset (p+pg, 0, 3*pg);
33 rc = munmap (p+pg, pg);
35 memset (p+2*pg, 0, 2*pg);
[all...]
/netbsd-6-1-5-RELEASE/sys/arch/landisk/include/
H A Dkloader.h32 #define PG_VADDR(pg) SH3_PHYS_TO_P1SEG(VM_PAGE_TO_PHYS(pg))
/netbsd-6-1-5-RELEASE/sys/arch/dreamcast/include/
H A Dkloader.h31 #define PG_VADDR(pg) SH3_PHYS_TO_P1SEG(VM_PAGE_TO_PHYS(pg))
/netbsd-6-1-5-RELEASE/sys/arch/hpcmips/include/
H A Dkloader.h31 #define PG_VADDR(pg) MIPS_PHYS_TO_KSEG0(VM_PAGE_TO_PHYS(pg))
/netbsd-6-1-5-RELEASE/sys/arch/evbsh3/include/
H A Dkloader.h32 #define PG_VADDR(pg) SH3_PHYS_TO_P1SEG(VM_PAGE_TO_PHYS(pg))
/netbsd-6-1-5-RELEASE/regress/sys/uvm/pdsim/
H A Dpdsim.c61 struct vm_page *pg; local
63 pg = TAILQ_FIRST(&freeq);
64 if (pg == NULL) {
67 TAILQ_REMOVE(&freeq, pg, pageq);
68 pg->offset = idx << PAGE_SHIFT;
69 pg->uanon = NULL;
70 pg->uobject = obj;
71 pg->pqflags = 0;
72 obj->pages[idx] = pg;
76 return pg;
80 pdsim_pagefree(struct vm_page *pg) argument
112 struct vm_page *pg; local
120 pdsim_pagemarkreferenced(struct vm_page *pg) argument
127 pmap_is_referenced(struct vm_page *pg) argument
134 pmap_clear_reference(struct vm_page *pg) argument
146 struct vm_page *pg; local
165 struct vm_page *pg; local
182 struct vm_page *pg; local
[all...]
/netbsd-6-1-5-RELEASE/sys/uvm/
H A Duvm_pdpolicy_clockpro.c148 clockpro_setq(struct vm_page *pg, int qidx) argument
153 pg->pqflags = (pg->pqflags & ~PQ_QMASK) | (qidx * PQ_QFACTOR);
157 clockpro_getq(struct vm_page *pg) argument
161 qidx = (pg->pqflags & PQ_QMASK) / PQ_QFACTOR;
236 pageq_insert_tail(pageq_t *q, struct vm_page *pg) argument
239 TAILQ_INSERT_TAIL(&q->q_q, pg, pageq.queue);
245 pageq_insert_head(pageq_t *q, struct vm_page *pg) argument
248 TAILQ_INSERT_HEAD(&q->q_q, pg, pageq.queue);
254 pageq_remove(pageq_t *q, struct vm_page *pg) argument
268 struct vm_page *pg; local
282 clockpro_insert_tail(struct clockpro_state *s, int qidx, struct vm_page *pg) argument
292 clockpro_insert_head(struct clockpro_state *s, int qidx, struct vm_page *pg) argument
495 pageobj(struct vm_page *pg) argument
514 pageidx(struct vm_page *pg) argument
522 nonresident_pagelookupremove(struct vm_page *pg) argument
542 nonresident_pagerecord(struct vm_page *pg) argument
637 clockpro_movereferencebit(struct vm_page *pg, bool locked) argument
664 clockpro_clearreferencebit(struct vm_page *pg, bool locked) argument
676 struct vm_page *pg; local
724 clockpro___enqueuetail(struct vm_page *pg) argument
744 clockpro_pageenqueue(struct vm_page *pg) argument
803 clockpro_pagequeue(struct vm_page *pg) argument
815 clockpro_pagedequeue(struct vm_page *pg) argument
836 clockpro_pagerequeue(struct vm_page *pg) argument
843 pageq_remove(clockpro_queue(s, qidx), pg); local
851 handhot_endtest(struct vm_page *pg) argument
868 struct vm_page *pg; local
975 struct vm_page *pg; local
1090 uvmpdpol_pageactivate(struct vm_page *pg) argument
1108 uvmpdpol_pagedeactivate(struct vm_page *pg) argument
1115 uvmpdpol_pagedequeue(struct vm_page *pg) argument
1126 uvmpdpol_pageenqueue(struct vm_page *pg) argument
1179 uvmpdpol_pageisqueued_p(struct vm_page *pg) argument
1198 struct vm_page *pg; local
1212 struct vm_page *pg; local
1303 struct vm_page *pg; local
1402 struct vm_page *pg; local
[all...]
H A Duvm_page.c173 const struct vm_page *pg = n; local
174 const voff_t a = pg->offset;
200 * => call should have already set pg's object and offset pointers
205 uvm_pageinsert_list(struct uvm_object *uobj, struct vm_page *pg, argument
209 KASSERT(uobj == pg->uobject);
211 KASSERT((pg->flags & PG_TABLED) == 0);
231 TAILQ_INSERT_AFTER(&uobj->memq, where, pg, listq.queue);
233 TAILQ_INSERT_TAIL(&uobj->memq, pg, listq.queue);
234 pg->flags |= PG_TABLED;
240 uvm_pageinsert_tree(struct uvm_object *uobj, struct vm_page *pg) argument
250 uvm_pageinsert(struct uvm_object *uobj, struct vm_page *pg) argument
266 uvm_pageremove_list(struct uvm_object *uobj, struct vm_page *pg) argument
296 uvm_pageremove_tree(struct uvm_object *uobj, struct vm_page *pg) argument
304 uvm_pageremove(struct uvm_object *uobj, struct vm_page *pg) argument
965 uvm_vm_page_to_phys(const struct vm_page *pg) argument
981 struct vm_page *pg; local
1116 struct vm_page *pg; local
1202 struct vm_page *pg; local
1404 uvm_pagerealloc(struct vm_page *pg, struct uvm_object *newobj, voff_t newoff) argument
1432 uvm_pagezerocheck(struct vm_page *pg) argument
1473 uvm_pagefree(struct vm_page *pg) argument
1624 struct vm_page *pg; local
1666 uvm_page_own(struct vm_page *pg, const char *tag) argument
1721 struct vm_page *pg; local
1824 struct vm_page *pg; local
1843 uvm_pagewire(struct vm_page *pg) argument
1867 uvm_pageunwire(struct vm_page *pg) argument
1887 uvm_pagedeactivate(struct vm_page *pg) argument
1903 uvm_pageactivate(struct vm_page *pg) argument
1925 uvm_pagedequeue(struct vm_page *pg) argument
1941 uvm_pageenqueue(struct vm_page *pg) argument
1959 uvm_pagezero(struct vm_page *pg) argument
1996 uvm_page_lookup_freelist(struct vm_page *pg) argument
2011 uvm_page_locked_p(struct vm_page *pg) argument
2033 uvm_page_printit(struct vm_page *pg, bool full, void (*pr)(const char *, ...)) argument
2119 struct vm_page *pg; local
[all...]
H A Duvm_anon.c105 struct vm_page *pg = anon->an_page; local
118 if (pg && pg->loan_count) {
120 pg = uvm_anon_lockloanpg(anon);
127 if (pg) {
136 if (pg->uobject) {
138 KASSERT(pg->loan_count > 0);
139 pg->loan_count--;
140 pg->uanon = NULL;
142 mutex_exit(pg
263 struct vm_page *pg; local
342 struct vm_page *pg; local
433 struct vm_page *pg = anon->an_page; local
[all...]
H A Duvm_pager.c306 struct vm_page *pg; local
318 pg = pgs[0];
319 swap = (pg->uanon != NULL && pg->uobject == NULL) ||
320 (pg->pqflags & PQ_AOBJ) != 0;
322 uobj = pg->uobject;
329 if (pg->uobject != NULL) {
330 swslot = uao_find_swslot(pg->uobject,
331 pg->offset >> PAGE_SHIFT);
333 KASSERT(pg
528 struct vm_page *pg; local
[all...]
H A Duvm_loan.c57 * loans are tracked by pg->loan_count. an O->A page will have both
61 * each loan of a page to the kernel bumps the pg->wire_count. the
76 * locking: to read pg->loan_count either the owner or the page queues
77 * must be locked. to modify pg->loan_count, both the owner of the page
78 * and the PQs must be locked. pg->flags is (as always) locked by
343 struct vm_page *pg; local
356 pg = anon->an_page;
357 if (pg && (pg->pqflags & PQ_ANON) != 0 && anon->an_ref == 1) {
358 if (pg
453 struct vm_page *pg = pgpp[i]; local
540 struct vm_page *pg = *pgpp; local
618 struct vm_page *pg; local
837 struct vm_page *pg; local
963 struct vm_page *pg; local
1046 struct vm_page *pg; local
1109 struct vm_page *pg; local
1186 struct vm_page *pg; local
[all...]
H A Duvm_pdpolicy_clock.c171 struct vm_page *pg; local
181 pg = TAILQ_FIRST(&pdpol_state.s_inactiveq);
184 pg = ss->ss_nextpg;
185 if (pg != NULL && (pg->pqflags & PQ_INACTIVE) == 0) {
186 pg = TAILQ_FIRST(&pdpol_state.s_inactiveq);
189 if (pg == NULL) {
192 ss->ss_nextpg = TAILQ_NEXT(pg, pageq.queue);
201 lock = uvmpd_trylockowner(pg);
203 if (pmap_is_referenced(pg)) {
294 uvmpdpol_pagedeactivate(struct vm_page *pg) argument
316 uvmpdpol_pageactivate(struct vm_page *pg) argument
326 uvmpdpol_pagedequeue(struct vm_page *pg) argument
345 uvmpdpol_pageenqueue(struct vm_page *pg) argument
357 uvmpdpol_pageisqueued_p(struct vm_page *pg) argument
[all...]
H A Duvm_vnode.c233 struct vm_page *pg; local
246 pg = uvm_pagelookup(uobj, offset);
249 if (pg == NULL) {
254 pg = uvm_pagealloc(uobj, offset, NULL,
256 if (pg == NULL) {
266 UVMHIST_LOG(ubchist, "alloced %p (color %u)", pg,
267 VM_PGCOLOR_BUCKET(pg), 0,0);
275 if ((pg->flags & PG_BUSY) != 0) {
280 pg->flags |= PG_WANTED;
281 UVMHIST_LOG(ubchist, "wait %p (color %u)", pg,
[all...]
/netbsd-6-1-5-RELEASE/external/gpl3/gcc/dist/gcc/testsuite/g++.dg/ext/arm-fp16/
H A Dfp16-param-1.C10 extern void (*pg) (__fp16 *);
H A Dfp16-return-1.C10 extern __fp16 *(*pg) (void);
/netbsd-6-1-5-RELEASE/external/gpl3/gcc/dist/gcc/testsuite/gcc.target/arm/
H A Dfp16-param-1.c10 extern void (*pg) (__fp16 *);
H A Dfp16-return-1.c10 extern __fp16 *(*pg) (void);
/netbsd-6-1-5-RELEASE/sys/arch/zaurus/include/
H A Dkloader.h45 #define PG_VADDR(pg) kloader_phystov(VM_PAGE_TO_PHYS(pg))
/netbsd-6-1-5-RELEASE/sys/arch/ia64/include/
H A Dpmap.h120 #define pmap_is_referenced(pg) \
121 (((pg)->mdpage.pvh_attrs & PGA_REFERENCED) != 0)
122 #define pmap_is_modified(pg) \
123 (((pg)->mdpage.pvh_attrs & PGA_MODIFIED) != 0)
162 #define VM_MDPAGE_INIT(pg) \
164 TAILQ_INIT(&(pg)->mdpage.pv_list); \
165 mutex_init(&(pg)->mdpage.pv_mutex, MUTEX_DEFAULT, IPL_NONE); \
/netbsd-6-1-5-RELEASE/sys/arch/xen/x86/
H A Dxen_bus_dma.c69 struct vm_page *pg, *pgnext; local
90 for (pg = mlistp->tqh_first; pg != NULL; pg = pg->pageq.queue.tqe_next) {
91 pa = VM_PAGE_TO_PHYS(pg);
128 pg = NULL;
133 for (pg = mlistp->tqh_first, i = 0; pg != NULL; pg
[all...]
/netbsd-6-1-5-RELEASE/sys/common/pmap/tlb/
H A Dvmpagemd.h86 #define VM_PAGEMD_PVLIST_LOCK(pg, list_change) \
103 #define VM_MDPAGE_INIT(pg) \
105 (pg)->mdpage.mdpg_first.pv_next = NULL; \
106 (pg)->mdpage.mdpg_first.pv_pmap = NULL; \
107 (pg)->mdpage.mdpg_first.pv_va = (pg)->phys_addr; \
108 (pg)->mdpage.mdpg_attrs = 0; \
109 VM_PAGEMD_PVLIST_LOCK_INIT(&(pg)->mdpage); \
/netbsd-6-1-5-RELEASE/external/bsd/openldap/dist/contrib/slapd-modules/acl/
H A Dposixgroup.c50 pg_t *pg; local
57 pg = ch_calloc( 1, sizeof( pg_t ) );
59 pg->pg_style = style;
61 switch ( pg->pg_style ) {
63 rc = dnNormalize( 0, NULL, NULL, &pat, &pg->pg_pat, NULL );
73 ber_dupbv( &pg->pg_pat, &pat );
79 fname, lineno, style_strings[ pg->pg_style ] );
123 *privp = (void *)pg;
127 (void)pg_dynacl_destroy( (void *)pg );
137 pg_t *pg local
178 pg_t *pg = (pg_t *)priv; local
301 pg_t *pg = (pg_t *)priv; local
[all...]
/netbsd-6-1-5-RELEASE/sys/arch/hppa/include/
H A Dpmap.h66 #define PVF_MOD PTE_PROT(TLB_DIRTY) /* pg/mp is modified */
67 #define PVF_REF PTE_PROT(TLB_REFTRAP) /* pg/mp (inv) is referenced */
68 #define PVF_WRITE PTE_PROT(TLB_WRITE) /* pg/mp is writable */
69 #define PVF_EXEC PTE_PROT(TLB_EXECUTE) /* pg/mp is executable */
70 #define PVF_UNCACHEABLE PTE_PROT(TLB_UNCACHEABLE) /* pg/mp is uncacheable */
136 #define pmap_clear_modify(pg) pmap_changebit(pg, 0, PTE_PROT(TLB_DIRTY))
137 #define pmap_clear_reference(pg) \
138 pmap_changebit(pg, PTE_PROT(TLB_REFTRAP), 0)
139 #define pmap_is_modified(pg) pmap_testbi
174 pmap_page_protect(struct vm_page *pg, vm_prot_t prot) argument
[all...]
/netbsd-6-1-5-RELEASE/lib/libc/db/btree/
H A Dbt_overflow.c86 pgno_t pg; local
90 memmove(&pg, p, sizeof(pg));
95 if (pg == P_INVALID || sz == 0)
113 for (p = *buf;; p = (char *)p + nb, pg = h->nextpg) {
114 if ((h = mpool_get(t->bt_mp, pg, 0)) == NULL)
139 __ovfl_put(BTREE *t, const DBT *dbt, pgno_t *pg) argument
175 *pg = h->pgno;
199 pgno_t pg; local
203 (void)memmove(&pg,
[all...]

Completed in 301 milliseconds

1234567891011