/linux-master/fs/ |
H A D | buffer.c | 84 * Returns if the folio has dirty or writeback buffers. If all the buffers 89 bool *dirty, bool *writeback) 92 *dirty = false; 110 *dirty = true; 530 * as you dirty the buffers, and then use osync_inode_buffers to wait for 531 * completion. Any other dirty buffers which are not yet queued for 655 * `bblock + 1' is probably a dirty indirect block. Hunt it down and, if it's 656 * dirty, schedule it for IO. So that indirects merge nicely with their data. 691 * Add a page to the dirty page list. 696 * If the page has buffers, the uptodate buffers are set dirty, t 88 buffer_check_dirty_writeback(struct folio *folio, bool *dirty, bool *writeback) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_pages.c | 161 obj->mm.dirty = false; 310 obj->mm.dirty = false;
|
/linux-master/include/linux/ |
H A D | kvm_host.h | 1230 void kvm_release_pfn(kvm_pfn_t pfn, bool dirty); 1306 void kvm_vcpu_unmap(struct kvm_vcpu *vcpu, struct kvm_host_map *map, bool dirty); 2387 * kick the vcpu to the userspace to avoid dirty ring full. This 2392 /* Max number of entries allowed for each kvm dirty ring */
|
/linux-master/arch/m68k/fpsp040/ |
H A D | fpsp.h | 128 .set FPR_DIRTY_BITS,LV-91 | fpr dirty bits
|
/linux-master/arch/powerpc/kvm/ |
H A D | e500_mmu.c | 857 struct kvm_dirty_tlb *dirty) 856 kvm_vcpu_ioctl_dirty_tlb(struct kvm_vcpu *vcpu, struct kvm_dirty_tlb *dirty) argument
|
/linux-master/fs/ubifs/ |
H A D | recovery.c | 333 * dirty. 748 * find a dirty LEB which could be GC'd into LEB Y! Even LEB X 1175 dbg_rcvry("could not find a dirty LEB"); 1180 ubifs_assert(c, lp.free + lp.dirty >= wbuf->offs);
|
/linux-master/fs/ocfs2/dlm/ |
H A D | dlmmaster.c | 485 !list_empty(&res->dirty) || 496 !list_empty(&res->dirty) ? 'D' : ' ', 509 BUG_ON(!list_empty(&res->dirty)); 545 INIT_LIST_HEAD(&res->dirty); 2602 * if we fail after this we need to re-dirty the lockres 2731 /* re-dirty the lockres if we failed */ 2852 * try to dirty the lockres before MIGRATING is set */ 2861 mlog(0, "about to wait on migration_wq, dirty=%s\n", 3469 /* re-dirty it on the new master */
|
H A D | dlmcommon.h | 295 struct list_head dirty; member in struct:dlm_lock_resource
|
/linux-master/drivers/net/ethernet/via/ |
H A D | via-velocity.h | 1411 int dirty; member in struct:velocity_info::rx_info
|
/linux-master/drivers/infiniband/hw/mlx4/ |
H A D | mlx4_ib.h | 293 bool dirty; member in struct:mlx4_wqn_range
|
/linux-master/sound/pci/emu10k1/ |
H A D | emuproc.c | 439 voice->dirty,
|
/linux-master/include/sound/ |
H A D | soc-dapm.h | 675 struct list_head dirty; member in struct:snd_soc_dapm_widget
|
/linux-master/arch/x86/events/ |
H A D | perf_event.h | 242 unsigned long dirty[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; member in struct:cpu_hw_events
|
/linux-master/arch/powerpc/include/asm/ |
H A D | kvm_book3s.h | 250 unsigned long gpa, bool dirty);
|
/linux-master/drivers/net/ethernet/amd/xgbe/ |
H A D | xgbe.h | 475 * dirty - Tx: index of descriptor to check for transfer complete 479 unsigned int dirty; member in struct:xgbe_ring
|
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_gem.c | 888 reg->dirty = true;
|
/linux-master/sound/pci/hda/ |
H A D | hda_codec.c | 683 unsigned char dirty; /* setups should be cleared */ member in struct:hda_cvt_setup 1162 p->dirty = 0; 1164 /* make other inactive cvts with the same stream-tag dirty */ 1170 p->dirty = 1; 1235 if (p->dirty)
|
/linux-master/arch/powerpc/kernel/ |
H A D | asm-offsets.c | 430 OFFSET(VCPU_VPA_DIRTY, kvm_vcpu, arch.vpa.dirty);
|
/linux-master/fs/bcachefs/ |
H A D | bcachefs.h | 15 * Multiple cache devices is intended to give us the ability to mirror dirty 29 * "cached" data is always dirty. The end result is that we get thin 139 * We can't just invalidate any bucket - it might contain dirty data or 140 * metadata. If it once contained dirty data, other writes might overwrite it 668 bool dirty; member in struct:journal_seq_blacklist_table::journal_seq_blacklist_table_entry
|
/linux-master/virt/kvm/ |
H A D | kvm_main.c | 385 * are related to dirty logging, and many do the TLB flush out of 1447 * Allocation size is twice as large as the actual dirty bitmap size. 1696 * If dirty logging is disabled, nullify the bitmap; the old bitmap 1761 * Free the dirty bitmap as needed; the below check encompasses 2164 * kvm_get_dirty_log - get a snapshot of dirty pages 2167 * @is_dirty: set to '1' if any dirty pages were found 2178 /* Dirty ring tracking may be exclusive to dirty log tracking */ 2213 * kvm_get_dirty_log_protect - get a snapshot of dirty pages 2214 * and reenable dirty page tracking for the corresponding pages. 2219 * concurrently. So, to avoid losing track of dirty page 3141 kvm_release_pfn(kvm_pfn_t pfn, bool dirty) argument 3183 kvm_vcpu_unmap(struct kvm_vcpu *vcpu, struct kvm_host_map *map, bool dirty) argument [all...] |
/linux-master/drivers/misc/vmw_vmci/ |
H A D | vmci_queue_pair.c | 632 u64 num_pages, bool dirty) 637 if (dirty) 631 qp_release_pages(struct page **pages, u64 num_pages, bool dirty) argument
|
/linux-master/fs/jfs/ |
H A D | jfs_txnmgr.c | 548 /* write dirty metadata & forward log syncpt */ 701 /* mark the page dirty and nohomeok */ 1233 * Mark inode as not dirty. It will still be on the dirty 1235 * it gets marked dirty again 2593 void txAbort(tid_t tid, int dirty) argument 2634 * mark filesystem dirty 2636 if (dirty)
|
/linux-master/arch/sparc/mm/ |
H A D | init_64.c | 483 bool dirty = test_bit(PG_dcache_dirty, &folio->flags); local 484 if (dirty) { 2299 * When a page gets marked as dcache-dirty, we store the
|
/linux-master/drivers/perf/ |
H A D | xgene_pmu.c | 319 XGENE_PMU_EVENT_ATTR(csw-inbound-dirty, 0x16), 394 XGENE_PMU_EVENT_ATTR(dirty-eviction, 0x07),
|
/linux-master/kernel/ |
H A D | kprobes.c | 257 kprobe_opcode_t *slot, int dirty) 279 if (dirty) { 256 __free_insn_slot(struct kprobe_insn_cache *c, kprobe_opcode_t *slot, int dirty) argument
|