/xnu-2422.115.4/osfmk/mach/ |
H A D | memory_object.defs | 132 * This call indicates whether the returned data is dirty and whether 146 dirty : boolean_t;
|
H A D | memory_object_types.h | 139 boolean_t dirty, 235 /* ... only dirty pages. */ 237 /* ... dirty and precious pages. */ 411 dirty:1, /* Page must be cleaned (O) */ member in struct:upl_page_info 545 * with a dirty page that hasn't yet been seen by the FS 646 (((upl)[(index)].phys_addr != 0) ? ((upl)[(index)].dirty) : FALSE)
|
/xnu-2422.115.4/osfmk/vm/ |
H A D | vm_pageout.c | 512 assert((m->dirty) || (m->precious) || 546 * Since the page is left "dirty" but "not modifed", we 553 m->dirty = FALSE; 556 if (m->dirty) { 586 * We do not re-set m->dirty ! 591 * will take care of resetting dirty. We clear the 610 * Set the dirty state according to whether or not the page was 616 * consulted if m->dirty is false. 619 m->dirty = pmap_is_modified(m->phys_page); 621 if (m->dirty) vm_pageout_cluster_dirtie 4163 boolean_t dirty; local [all...] |
H A D | vm_resident.c | 552 m->dirty = FALSE; 2878 m->dirty && m->object->internal && 3012 m->dirty && m->object->internal && 3624 (p->dirty ? "" : "!"), 4020 if (m->pmapped || m->dirty || m->precious) { 4173 * when substituing for pmapped/dirty pages 4238 !m1->dirty && 4254 m1->dirty || 4296 m2->dirty = m1->dirty; [all...] |
H A D | vm_object.c | 1127 if (p->pmapped || p->dirty || p->precious) { 1135 if (p->reference == FALSE || p->dirty == FALSE) { 1145 if (p->dirty == FALSE && p->precious == FALSE) { 1155 if (p->dirty == FALSE) 1375 if (p->wpmapped || p->dirty || p->precious) { 1391 if (p->wpmapped || p->dirty || p->precious) { 2003 if (!p->dirty && p->wpmapped) 2004 p->dirty = pmap_is_modified(p->phys_page); 2006 if ((p->dirty || p->precious) && !p->error && object->alive) { 2580 m->dirty [all...] |
H A D | vm_apple_protect.c | 99 boolean_t dirty, 274 * the pages provided by this EMM are not supposed to be dirty or dirtied 285 __unused boolean_t dirty, 599 * The pages are currently dirty because we've just been 279 apple_protect_pager_data_return( __unused memory_object_t mem_obj, __unused memory_object_offset_t offset, __unused memory_object_cluster_size_t data_cnt, __unused memory_object_offset_t *resid_offset, __unused int *io_error, __unused boolean_t dirty, __unused boolean_t kernel_copy, __unused int upl_flags) argument
|
H A D | vm_compressor_pager.c | 96 __unused boolean_t dirty, 490 __unused boolean_t dirty, 484 compressor_memory_object_data_return( __unused memory_object_t mem_obj, __unused memory_object_offset_t offset, __unused memory_object_cluster_size_t size, __unused memory_object_offset_t *resid_offset, __unused int *io_error, __unused boolean_t dirty, __unused boolean_t kernel_copy, __unused int upl_flags) argument
|
H A D | vm_page.h | 232 dirty:1, /* Page must be cleaned (O) */ member in struct:vm_page 252 was_dirty:1, /* was this page previously dirty? */ 356 * object/offset-page mapping, and may be dirty. 701 __page__->dirty = TRUE; \
|
H A D | vm_swapfile_pager.c | 99 boolean_t dirty, 261 * the pages provided by this EMM are not supposed to be dirty or dirtied 272 __unused boolean_t dirty, 439 * The pages are currently dirty because we've just been 266 swapfile_pager_data_return( __unused memory_object_t mem_obj, __unused memory_object_offset_t offset, __unused memory_object_cluster_size_t data_cnt, __unused memory_object_offset_t *resid_offset, __unused int *io_error, __unused boolean_t dirty, __unused boolean_t kernel_copy, __unused int upl_flags) argument
|
H A D | device_vm.c | 287 __unused boolean_t dirty, 281 device_pager_data_return( memory_object_t mem_obj, memory_object_offset_t offset, memory_object_cluster_size_t data_cnt, __unused memory_object_offset_t *resid_offset, __unused int *io_error, __unused boolean_t dirty, __unused boolean_t kernel_copy, __unused int upl_flags) argument
|
H A D | memory_object.c | 124 * 1. Page is dirty and should_return is not RETURN_NONE. 128 * As a side effect, m->dirty will be made consistent 135 (((m)->dirty || ((m)->dirty = pmap_is_modified((m)->phys_page))) || \ 215 * for the page to go from the clean to the dirty state 231 * Handle returning dirty or precious pages 286 * is RETURN_DIRTY and the page is dirty, or 288 * is either dirty or precious); and, 416 * are flushed, that is dirty & precious pages are written to 577 * End of a run of dirty/preciou 2179 memory_object_data_return( memory_object_t memory_object, memory_object_offset_t offset, memory_object_cluster_size_t size, memory_object_offset_t *resid_offset, int *io_error, boolean_t dirty, boolean_t kernel_copy, int upl_flags ) argument [all...] |
H A D | default_freezer.c | 685 __unused boolean_t dirty, 679 df_memory_object_data_return( __unused memory_object_t mem_obj, __unused memory_object_offset_t offset, __unused memory_object_cluster_size_t size, __unused memory_object_offset_t *resid_offset, __unused int *io_error, __unused boolean_t dirty, __unused boolean_t kernel_copy, __unused int upl_flags) argument
|
H A D | vm_map.c | 10711 (p->dirty || pmap_is_modified(p->phys_page))) 12984 if (m->dirty || pmap_is_modified(m->phys_page)) 13033 * - discard pages, write dirty or precious 13037 * - write dirty or precious pages back to 13234 * is writable (ie dirty pages may have already been sent back) 13814 /* Pull the dirty status from the pmap, since we cleared the 13816 if ((refmod & VM_MEM_MODIFIED) && !m->dirty) { 13850 unsigned int purgeable, clean, dirty, wired; local 13859 default_freezer_pack(&purgeable, &wired, &clean, &dirty, dirty_budget, &shared, entry->object.vm_object, NULL); 13864 *dirty_count += dirty; 13935 unsigned int purgeable, clean, dirty, wired; local [all...] |
H A D | bsd_vm.c | 616 __unused boolean_t dirty, 610 vnode_pager_data_return( memory_object_t mem_obj, memory_object_offset_t offset, memory_object_cluster_size_t data_cnt, memory_object_offset_t *resid_offset, int *io_error, __unused boolean_t dirty, __unused boolean_t kernel_copy, int upl_flags) argument
|
H A D | vm_fault.c | 1555 m->dirty = TRUE; 2239 * then we'll activate it, mark it dirty and keep it 2302 * dirty is protected by the object lock 3995 m->dirty = TRUE; 5346 * Copy the page, and note that it is dirty 5384 if(!dst_page->dirty){ 5401 if(!dst_page->dirty){
|
/xnu-2422.115.4/libsyscall/wrappers/libproc/ |
H A D | libproc.h | 123 int proc_set_dirty(pid_t pid, bool dirty);
|
H A D | libproc.c | 268 proc_set_dirty(pid_t pid, bool dirty) argument 270 if (__proc_info(PROC_INFO_CALL_DIRTYCONTROL, pid, PROC_DIRTYCONTROL_SET, dirty, NULL, 0) == -1) {
|
/xnu-2422.115.4/bsd/kern/ |
H A D | kern_memorystatus.c | 495 uint32_t purgeable, wired, clean, dirty; variable 504 error = task_freeze(p->task, &purgeable, &wired, &clean, &dirty, max_pages, &shared, FALSE); 749 MEMORYSTATUS_DEBUG(1, "memorystatus_schedule_idle_demotion_locked: scheduling demotion to idle band for process %d (dirty:0x%x, set_state %d, demotions %d).\n", 1021 MEMORYSTATUS_DEBUG(1, "memorystatus_update_idle_priority_locked(): pid %d dirty 0x%X\n", p->p_pid, p->p_memstat_dirty); 1033 * Processes can opt to have their state tracked by the kernel, indicating when they are busy (dirty) or idle 1035 * to their desired, higher, jetsam priority when dirty (and are therefore killed later), and demoted to the low 1045 * is marked dirty, the timer will be cancelled. Upon being subsequently marked clean, the deferment will either be 1085 MEMORYSTATUS_DEBUG(1, "memorystatus_on_track_dirty(): set idle-exit %s / deferred %s / dirty %s for process %d\n", 1142 * Process is set to be terminated and we're attempting to mark it dirty. 1155 /* Clearing the dirty shutdow 2181 uint32_t purgeable, wired, clean, dirty; local [all...] |
/xnu-2422.115.4/bsd/hfs/hfscommon/Misc/ |
H A D | VolumeAllocation.c | 246 Boolean dirty); 1233 // Update the volume's free block count, and mark the VCB as dirty. 1527 ; dirty 1533 Boolean dirty) 1538 KERNEL_DEBUG_CONSTANT(HFSDBG_RELEASE_BITMAP_BLOCK | DBG_FUNC_START, dirty, 0, 0, 0, 0); 1541 if (dirty) 1547 if (dirty) { 1877 Boolean dirty = false; local 2030 dirty = true; 4265 * dirty 1530 ReleaseBitmapBlock( ExtendedVCB *vcb, uintptr_t blockRef, Boolean dirty) argument [all...] |
/xnu-2422.115.4/bsd/nfs/ |
H A D | nfs_bio.c | 337 * Called by vnode_pager() on pageout request of non-dirty page. 362 * If there's a dirty range in the buffer, check to 363 * see if this page intersects with the dirty range. 431 * update buffer's valid/dirty info from UBC 456 /* anything beyond the end of the file is not valid or dirty */ 1018 * non-needcommit nocache buffer AND no pages are dirty. 1238 /* move to dirty list */ 2087 * Flush the buffer if it's dirty. 2133 /* there are also dirty page(s) (or range) in the read range, */ 2153 /* buffer's not dirty, s 2607 uint32_t dirty = bp->nb_dirty; local 3012 int error = 0, retv, wcred_set, flags, dirty; local [all...] |
/xnu-2422.115.4/osfmk/default_pager/ |
H A D | dp_memory_object.c | 758 __unused boolean_t dirty, 752 dp_memory_object_data_return( memory_object_t mem_obj, memory_object_offset_t offset, memory_object_cluster_size_t size, __unused memory_object_offset_t *resid_offset, __unused int *io_error, __unused boolean_t dirty, __unused boolean_t kernel_copy, __unused int upl_flags) argument
|
/xnu-2422.115.4/bsd/vfs/ |
H A D | vfs_cluster.c | 2699 * never force zero valid pages - dirty or clean 2881 * written as dirty 3073 * UPL has been automatically set to clear the dirty flags (both software and hardware) 3109 * cluster method of delaying dirty pages 5366 * - only pages that are currently dirty are returned... these are the ones we need to clean 5367 * - the hardware dirty bit is cleared when the page is gathered into the UPL... the software dirty bit is set 5368 * - if we have to abort the I/O for some reason, the software dirty bit is left set since we didn't clean the page 5369 * - when we commit the page, the software dirty bit is cleared... the hardware dirty bi 6269 vfs_drt_do_mark_pages( void **private, u_int64_t offset, u_int length, u_int *setcountp, int dirty) argument [all...] |
/xnu-2422.115.4/bsd/dev/dtrace/ |
H A D | dtrace.c | 1423 * clean the dirty dynamic variable lists on all CPUs. Dynamic variable 1432 dtrace_dynvar_t *dirty; 1442 * If the dirty list is NULL, there is no dirty work to do. 1459 * Atomically move the dirty list aside. 1462 dirty = dcpu->dtdsc_dirty; 1465 * Before we zap the dirty list, set the rinsing list. 1468 * on a hash chain, either the dirty list or the 1471 dcpu->dtdsc_rinsing = dirty; 1474 dirty, NUL [all...] |
/xnu-2422.115.4/iokit/Kernel/ |
H A D | IOMemoryDescriptor.cpp | 2095 page_list[page].dirty = 0; 2125 // Pages do not need to be marked as dirty on commit
|