/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | huge_pages.c | 161 obj->mm.dirty = false; 310 obj->mm.dirty = false;
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_shmem.c | 33 bool dirty, bool backup) 49 if (dirty) 137 * Our bo are always dirty and so we require 144 * dirty pages -- unless you try over and over 321 /* Begin writeback on each dirty page */ 376 obj->mm.dirty = false; 406 obj->mm.dirty, obj->mm.madv == I915_MADV_WILLNEED); 408 obj->mm.dirty = false; 32 shmem_sg_free_table(struct sg_table *st, struct address_space *mapping, bool dirty, bool backup) argument
|
H A D | i915_gem_object_types.h | 136 * GPU access is not coherent with the CPU cache. If the cache is dirty 152 * GPU access is coherent with the CPU cache. If the cache is dirty, 154 * reading and writing through the CPU cache. GPU writes can dirty the 501 * Track if we are we dirty with writes through the CPU cache for this 512 * dirty, for example the page zeroing done by the kernel might leave 690 bool dirty:1; member in struct:drm_i915_gem_object::__anon629
|
H A D | i915_gem_phys.c | 106 if (obj->mm.dirty) { 128 obj->mm.dirty = false;
|
H A D | i915_gem_internal.c | 126 obj->mm.dirty = false;
|
/linux-master/drivers/gpu/drm/i915/display/ |
H A D | intel_fbdev.c | 263 if (helper->fb->funcs->dirty) 264 return helper->fb->funcs->dirty(helper->fb, NULL, 0, 0, clip, 1);
|
H A D | intel_dpt.c | 167 dpt->obj->mm.dirty = true;
|
/linux-master/drivers/gpu/drm/ |
H A D | drm_framebuffer.c | 702 * implement the corresponding &drm_framebuffer_funcs.dirty callback. 762 if (fb->funcs->dirty) { 763 ret = fb->funcs->dirty(fb, file_priv, flags, r->color,
|
H A D | drm_client.c | 477 * This calls &drm_framebuffer_funcs->dirty (if present) to flush buffer changes 485 if (!buffer || !buffer->fb || !buffer->fb->funcs->dirty) 496 return buffer->fb->funcs->dirty(buffer->fb, buffer->client->file, 500 return buffer->fb->funcs->dirty(buffer->fb, buffer->client->file,
|
/linux-master/include/linux/ |
H A D | buffer_head.h | 21 BH_Dirty, /* Is dirty */ 122 BUFFER_FNS(Dirty, dirty) 123 TAS_BUFFER_FNS(Dirty, dirty) 189 bool *dirty, bool *writeback);
|
H A D | kprobes.h | 307 kprobe_opcode_t *slot, int dirty); 320 static inline void free_##__name##_slot(kprobe_opcode_t *slot, int dirty)\ 322 __free_insn_slot(&kprobe_##__name##_slots, slot, dirty); \
|
/linux-master/fs/xfs/scrub/ |
H A D | quota_repair.c | 112 bool *dirty) 150 *dirty = true; 159 bool *dirty) 164 *dirty = true; 167 *dirty = true; 180 bool dirty = false; local 197 error = xrep_quota_item_bmap(sc, dq, &dirty); 205 dirty = true; 210 dirty = true; 215 dirty 109 xrep_quota_item_bmap( struct xfs_scrub *sc, struct xfs_dquot *dq, bool *dirty) argument 156 xrep_quota_item_timer( struct xfs_scrub *sc, const struct xfs_dquot_res *res, bool *dirty) argument [all...] |
/linux-master/fs/jfs/ |
H A D | jfs_txnmgr.c | 548 /* write dirty metadata & forward log syncpt */ 701 /* mark the page dirty and nohomeok */ 1233 * Mark inode as not dirty. It will still be on the dirty 1235 * it gets marked dirty again 2593 void txAbort(tid_t tid, int dirty) argument 2634 * mark filesystem dirty 2636 if (dirty)
|
/linux-master/fs/jffs2/ |
H A D | debug.c | 35 JFFS2_ERROR("free %#08x + dirty %#08x + used %#08x + wasted %#08x + unchecked %#08x != total %#08x.\n", 44 JFFS2_ERROR("free %#08x + dirty %#08x + used %#08x + erasing %#08x + bad %#08x + wasted %#08x + unchecked %#08x != total %#08x.\n", 163 uint32_t free = 0, dirty = 0, used = 0, wasted = 0, local 171 dirty += c->gcblock->dirty_size; 179 dirty += c->nextblock->dirty_size; 187 dirty += jeb->dirty_size; 195 dirty += jeb->dirty_size; 203 dirty += jeb->dirty_size; 211 dirty += jeb->dirty_size; 219 dirty 503 uint32_t dirty = 0; local 525 uint32_t dirty = 0; local 548 uint32_t dirty = 0; local [all...] |
/linux-master/fs/dlm/ |
H A D | lowcomms.c | 133 bool dirty; member in struct:writequeue_entry 785 if (e->dirty) 1120 e->dirty = true; 1185 entry->dirty = false;
|
/linux-master/fs/bcachefs/ |
H A D | btree_io.h | 21 atomic_inc(&c->btree_cache.dirty); 27 atomic_dec(&c->btree_cache.dirty);
|
/linux-master/arch/mips/include/asm/ |
H A D | pgtable.h | 455 * contains the memory attribute bits, dirty bits, and various other 496 pte_t entry, int dirty) 622 /* We don't have hardware dirty/accessed bits, generic_pmdp_establish is fine.*/ 494 ptep_set_access_flags(struct vm_area_struct *vma, unsigned long address, pte_t *ptep, pte_t entry, int dirty) argument
|
/linux-master/tools/testing/selftests/powerpc/ |
H A D | Makefile | 10 GIT_VERSION = $(shell git describe --always --long --dirty || echo "unknown")
|
/linux-master/samples/vfio-mdev/ |
H A D | mtty.c | 1303 struct iova_bitmap *dirty) 1301 mtty_log_read_and_clear(struct vfio_device *vdev, unsigned long iova, unsigned long length, struct iova_bitmap *dirty) argument
|
/linux-master/fs/ |
H A D | dax.c | 563 * the tree, and PTE writes will simply dirty the entire PMD entry. 852 * MAP_SYNC on a dax mapping guarantees dirty metadata is 866 * already in the tree, we will skip the insertion and just dirty the PMD as 876 bool dirty = write && !dax_fault_is_synchronous(iter, vmf->vma); local 879 if (dirty) 906 * tree and dirty it if necessary. 916 if (dirty) 934 * A page got tagged dirty in DAX mapping? Something is seriously 1001 * After we have flushed the cache, we can clear the dirty tag. There 1002 * cannot be new dirty dat [all...] |
/linux-master/arch/x86/events/ |
H A D | perf_event.h | 242 unsigned long dirty[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; member in struct:cpu_hw_events
|
/linux-master/arch/powerpc/include/asm/ |
H A D | kvm_book3s.h | 250 unsigned long gpa, bool dirty);
|
/linux-master/drivers/infiniband/core/ |
H A D | umem.c | 48 static void __ib_umem_release(struct ib_device *dev, struct ib_umem *umem, int dirty) argument 50 bool make_dirty = umem->writable && dirty; 54 if (dirty)
|
/linux-master/drivers/md/bcache/ |
H A D | writeback.c | 3 * background writeback - scan btree for dirty data and write it to the backing 42 * Unfortunately there is no control of global dirty data. If the 43 * user states that they want 10% dirty data in the cache, and has, 45 * backing volume uses about 2% of the cache for dirty data. 54 /* Ensure each backing dev gets at least one dirty share */ 67 * First, the error (number of sectors that are dirty beyond our 75 * attempts to write at a rate that would retire all the dirty 84 int64_t dirty = bcache_dev_sectors_dirty(&dc->disk); local 85 int64_t error = dirty - target; 92 * We need to consider the number of dirty bucket [all...] |
H A D | journal.c | 554 pr_debug("bnode %p: dirty bit cleaned by others\n", b); 710 j->cur->dirty = false; 912 if (c->journal.cur->dirty) 949 } else if (!w->dirty) { 950 w->dirty = true;
|