/linux-master/include/drm/ttm/ |
H A D | ttm_caching.h | 31 * enum ttm_caching - CPU caching and BUS snooping behavior. 53 pgprot_t ttm_prot_from_caching(enum ttm_caching caching, pgprot_t tmp);
|
H A D | ttm_pool.h | 44 * @caching: the caching type our pages have 52 enum ttm_caching caching; member in struct:ttm_pool_type 61 * struct ttm_pool - Pool for all caching and orders 67 * @caching: pools for each caching/order 78 } caching[TTM_NUM_CACHING_TYPES]; member in struct:ttm_pool
|
H A D | ttm_tt.h | 42 * struct ttm_tt - This is a structure holding the pages, caching- and aperture 109 * @caching: The current caching state of the pages, see enum 112 enum ttm_caching caching; member in struct:ttm_tt 149 * @caching: the desired caching state of the pages 158 uint32_t page_flags, enum ttm_caching caching, 161 uint32_t page_flags, enum ttm_caching caching);
|
H A D | ttm_resource.h | 185 * @caching: See enum ttm_caching 193 enum ttm_caching caching; member in struct:ttm_bus_placement
|
/linux-master/drivers/gpu/drm/ttm/ |
H A D | ttm_module.c | 57 * @caching: The ttm caching mode 62 pgprot_t ttm_prot_from_caching(enum ttm_caching caching, pgprot_t tmp) argument 65 if (caching == ttm_cached) 69 if (caching == ttm_write_combined) 78 if (caching == ttm_write_combined)
|
H A D | ttm_pool.c | 26 /* Pooling of allocated pages is necessary because changing the caching 132 /* Reset the caching and pages of size 1 << order */ 133 static void ttm_pool_free_page(struct ttm_pool *pool, enum ttm_caching caching, argument 144 if (caching != ttm_cached && !PageHighMem(p)) 163 /* Apply a new caching to an array of pages */ 165 enum ttm_caching caching) 173 switch (caching) { 260 enum ttm_caching caching, unsigned int order) 263 pt->caching = caching; 164 ttm_pool_apply_caching(struct page **first, struct page **last, enum ttm_caching caching) argument 259 ttm_pool_type_init(struct ttm_pool_type *pt, struct ttm_pool *pool, enum ttm_caching caching, unsigned int order) argument 287 ttm_pool_select_type(struct ttm_pool *pool, enum ttm_caching caching, unsigned int order) argument 392 ttm_pool_free_range(struct ttm_pool *pool, struct ttm_tt *tt, enum ttm_caching caching, pgoff_t start_page, pgoff_t end_page) argument 433 struct page **caching = tt->pages; local [all...] |
H A D | ttm_tt.c | 151 enum ttm_caching caching, 159 ttm->caching = caching; 163 uint32_t page_flags, enum ttm_caching caching, 166 ttm_tt_init_fields(ttm, bo, page_flags, caching, extra_pages); 194 uint32_t page_flags, enum ttm_caching caching) 198 ttm_tt_init_fields(ttm, bo, page_flags, caching, 0); 460 iter_tt->prot = ttm_prot_from_caching(tt->caching, PAGE_KERNEL); 148 ttm_tt_init_fields(struct ttm_tt *ttm, struct ttm_buffer_object *bo, uint32_t page_flags, enum ttm_caching caching, unsigned long extra_pages) argument 162 ttm_tt_init(struct ttm_tt *ttm, struct ttm_buffer_object *bo, uint32_t page_flags, enum ttm_caching caching, unsigned long extra_pages) argument 193 ttm_sg_tt_init(struct ttm_tt *ttm, struct ttm_buffer_object *bo, uint32_t page_flags, enum ttm_caching caching) argument
|
H A D | ttm_bo_util.c | 288 * setting up a PTE with the caching model indicated by @c_state. 294 enum ttm_caching caching; local 298 caching = bo->ttm->caching; 302 caching = res->bus.caching; 305 return ttm_prot_from_caching(caching, tmp); 323 if (mem->bus.caching == ttm_write_combined) 326 else if (mem->bus.caching == ttm_cached) 357 if (num_pages == 1 && ttm->caching [all...] |
H A D | ttm_agp_backend.c | 55 int ret, cached = ttm->caching == ttm_cached;
|
H A D | ttm_resource.c | 192 res->bus.caching = ttm_cached; 635 if (mem->bus.caching == ttm_write_combined) 639 else if (mem->bus.caching == ttm_cached)
|
/linux-master/drivers/gpu/drm/ttm/tests/ |
H A D | ttm_pool_test.c | 52 enum ttm_caching caching, 67 err = ttm_tt_init(tt, priv->mock_bo, page_flags, caching, 0); 75 enum ttm_caching caching) 83 tt = ttm_tt_kunit_init(test, 0, caching, size); 142 enum ttm_caching caching = ttm_uncached; local 147 tt = ttm_tt_kunit_init(test, 0, caching, size); 204 enum ttm_caching caching = ttm_uncached; local 215 err = ttm_sg_tt_init(tt, bo, 0, caching); 243 enum ttm_caching caching = ttm_uncached; local 248 pool = ttm_pool_pre_populated(test, size, caching); 50 ttm_tt_kunit_init(struct kunit *test, uint32_t page_flags, enum ttm_caching caching, size_t size) argument 73 ttm_pool_pre_populated(struct kunit *test, size_t size, enum ttm_caching caching) argument 305 enum ttm_caching caching = ttm_uncached; local 341 enum ttm_caching caching = ttm_uncached; local 372 enum ttm_caching caching = ttm_uncached; local 400 enum ttm_caching caching = ttm_uncached; local [all...] |
H A D | ttm_tt_test.c | 58 enum ttm_caching caching = ttm_cached; local 68 err = ttm_tt_init(tt, bo, page_flags, caching, extra_pages); 74 KUNIT_ASSERT_EQ(test, tt->caching, caching); 84 enum ttm_caching caching = ttm_cached; local 97 err = ttm_tt_init(tt, bo, 0, caching, 0); 107 enum ttm_caching caching = ttm_cached; local 115 err = ttm_tt_init(tt, bo, 0, caching, 0); 127 enum ttm_caching caching = ttm_cached; local 135 err = ttm_sg_tt_init(tt, bo, 0, caching); 148 enum ttm_caching caching = ttm_cached; local 205 enum ttm_caching caching = ttm_cached; local [all...] |
H A D | ttm_device_test.c | 179 pt = pool->caching[i].orders[j]; 181 KUNIT_EXPECT_EQ(test, pt.caching, i);
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_sgdma.c | 73 enum ttm_caching caching; local 76 caching = ttm_uncached; 78 caching = ttm_write_combined; 80 caching = ttm_cached; 86 if (ttm_sg_tt_init(&nvbe->ttm, bo, page_flags, caching)) {
|
/linux-master/drivers/acpi/ |
H A D | acpi_memhotplug.c | 49 unsigned short caching; /* memory cache attribute */ member in struct:acpi_memory_info 75 if ((info->caching == address64.info.mem.caching) && 88 new->caching = address64.info.mem.caching;
|
/linux-master/include/linux/ |
H A D | fscache.h | 2 /* General filesystem caching interface 9 * Documentation/filesystems/caching/netfs-api.rst 151 * slow-path functions for when there is actually caching available, and the 181 * fscache_acquire_volume - Register a volume as desiring caching services 187 * Register a volume as desiring caching services if they're available. The 194 * cookie pointer and can be returned if caching is refused. 209 * fscache_relinquish_volume - Cease caching a volume 214 * Indicate that a filesystem no longer desires caching services for a volume. 239 * See Documentation/filesystems/caching/netfs-api.rst for a complete 301 * See Documentation/filesystems/caching/netf 583 fscache_clear_page_bits(struct address_space *mapping, loff_t start, size_t len, bool caching) argument 614 fscache_write_to_cache(struct fscache_cookie *cookie, struct address_space *mapping, loff_t start, size_t len, loff_t i_size, netfs_io_terminated_t term_func, void *term_func_priv, bool caching) argument [all...] |
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_domain.c | 329 args->caching = I915_CACHING_CACHED; 331 args->caching = I915_CACHING_DISPLAY; 333 args->caching = I915_CACHING_NONE; 354 switch (args->caching) { 391 * The caching mode of proxy object is handled by its generator, and 400 args->caching != I915_CACHING_CACHED)
|
H A D | i915_gem_ttm.c | 273 enum ttm_caching caching; local 288 caching = i915_ttm_select_tt_caching(obj); 289 if (i915_gem_object_is_shrinkable(obj) && caching == ttm_cached) { 300 ret = ttm_tt_init(&i915_tt->ttm, bo, page_flags, caching, ccs_pages); 551 * @obj: The GEM object used for sg-table caching 678 mem->bus.caching = ttm_write_combined;
|
/linux-master/fs/netfs/ |
H A D | buffered_write.c | 38 static void netfs_folio_start_fscache(bool caching, struct folio *folio) argument 40 if (caching) 44 static void netfs_folio_start_fscache(bool caching, struct folio *folio) argument 52 * avoid it. If we're doing local caching or content crypto, we award that 89 * caching service temporarily because the backing store got 775 bool caching, 867 * writable and mark them for caching. 879 netfs_folio_start_fscache(caching, folio); 905 bool caching = netfs_is_cache_enabled(ctx); local 909 _enter(",%lx,%llx-%llx,%u", folio->index, start, end, caching); 769 netfs_extend_writeback(struct address_space *mapping, struct netfs_group *group, struct xa_state *xas, long *_count, loff_t start, loff_t max_len, bool caching, size_t *_len, size_t *_top) argument [all...] |
/linux-master/drivers/gpu/drm/qxl/ |
H A D | qxl_ttm.c | 83 mem->bus.caching = ttm_write_combined; 89 mem->bus.caching = ttm_write_combined;
|
/linux-master/drivers/gpu/drm/radeon/ |
H A D | radeon_ttm.c | 272 mem->bus.caching = ttm_write_combined; 283 mem->bus.caching = ttm_write_combined; 443 if (ttm->caching == ttm_cached) 484 enum ttm_caching caching; local 501 caching = ttm_uncached; 503 caching = ttm_write_combined; 505 caching = ttm_cached; 507 if (ttm_sg_tt_init(>t->ttm, bo, page_flags, caching)) {
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_ttm_stolen_mgr.c | 286 mem->bus.caching = ttm_write_combined; 308 mem->bus.caching = ttm_write_combined;
|
/linux-master/arch/arm/mm/ |
H A D | proc-arm740.S | 110 bic r0, r0, #0x3F000000 @ set to standard caching mode
|
/linux-master/drivers/acpi/acpica/ |
H A D | rsaddr.c | 218 {ACPI_RSC_2BITFLAG, ACPI_RS_OFFSET(data.address.info.mem.caching),
|
/linux-master/fs/ceph/ |
H A D | addr.c | 514 static void ceph_fscache_write_to_cache(struct inode *inode, u64 off, u64 len, bool caching) argument 520 ceph_fscache_write_terminated, inode, caching); 527 static inline void ceph_fscache_write_to_cache(struct inode *inode, u64 off, u64 len, bool caching) argument 656 bool caching = ceph_is_cache_enabled(inode); local 718 if (caching) 720 ceph_fscache_write_to_cache(inode, page_off, len, caching); 939 bool caching = ceph_is_cache_enabled(inode); local 1252 ceph_fscache_write_to_cache(inode, offset, len, caching); 1271 if (caching) 1275 ceph_fscache_write_to_cache(inode, offset, len, caching); [all...] |