/linux-master/fs/cachefiles/ |
H A D | Makefile | 7 cache.o \
|
H A D | io.c | 43 * Handle completion of a read from the cache. 71 * Initiate a read from the cache. 194 * Query the occupancy of the cache in a region, returning where the next chunk 213 granularity = max_t(size_t, object->volume->cache->bsize, granularity); 252 * Handle completion of a write to the cache. 269 atomic_long_sub(ki->b_writing, &object->volume->cache->b_writing); 277 * Initiate a write to the cache. 286 struct cachefiles_cache *cache; local 293 cache = object->volume->cache; 384 struct cachefiles_cache *cache; local 523 struct cachefiles_cache *cache = object->volume->cache; local 606 struct cachefiles_cache *cache = object->volume->cache; local [all...] |
/linux-master/arch/arm/mach-mvebu/ |
H A D | dove.c | 12 #include <asm/hardware/cache-tauros2.h>
|
/linux-master/arch/loongarch/mm/ |
H A D | Makefile | 6 obj-y += init.o cache.o tlb.o tlbex.o extable.o \
|
/linux-master/arch/m68k/include/asm/ |
H A D | syscalls.h | 8 asmlinkage int sys_cacheflush(unsigned long addr, int scope, int cache,
|
/linux-master/arch/powerpc/kernel/ |
H A D | firmware.c | 13 #include <linux/cache.h>
|
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_params.h | 9 #include <linux/cache.h>
|
/linux-master/net/core/ |
H A D | hotdata.c | 3 #include <linux/cache.h>
|
/linux-master/drivers/md/dm-vdo/ |
H A D | cpu.h | 9 #include <linux/cache.h> 12 * uds_prefetch_address() - Minimize cache-miss latency by attempting to move data into a CPU cache 35 * uds_prefetch_range() - Minimize cache-miss latency by attempting to move a range of addresses 36 * into a CPU cache before they are accessed. 46 * Count the number of cache lines to fetch, allowing for the address range to span an 47 * extra cache line boundary due to address alignment.
|
H A D | block-map.c | 42 * In the current (newest) era, block map pages are not written unless there is cache pressure. In 121 * For adjusting VDO page cache statistic fields which are only mutated on the logical zone thread. 158 struct vdo_page_cache *cache = info->cache; local 160 return &cache->pages[(info - cache->infos) * VDO_BLOCK_SIZE]; 180 static int initialize_info(struct vdo_page_cache *cache) argument 184 INIT_LIST_HEAD(&cache->free_list); 185 for (info = cache->infos; info < cache 219 allocate_cache_components(struct vdo_page_cache *cache) argument 244 assert_on_cache_thread(struct vdo_page_cache *cache, const char *function_name) argument 255 assert_io_allowed(struct vdo_page_cache *cache) argument 262 report_cache_pressure(struct vdo_page_cache *cache) argument 378 struct vdo_page_cache *cache = info->cache; local 424 find_free_page(struct vdo_page_cache *cache) argument 442 find_page(struct vdo_page_cache *cache, physical_block_number_t pbn) argument 462 select_lru_page(struct vdo_page_cache *cache) argument 558 set_persistent_error(struct vdo_page_cache *cache, const char *context, int result) argument 671 struct vdo_page_cache *cache = info->cache; local 696 struct vdo_page_cache *cache = info->cache; local 738 struct vdo_page_cache *cache = info->cache; local 771 struct vdo_page_cache *cache = info->cache; local 813 save_pages(struct vdo_page_cache *cache) argument 891 struct vdo_page_cache *cache = info->cache; local 938 discard_a_page(struct vdo_page_cache *cache) argument 966 struct vdo_page_cache *cache = vdo_page_comp->cache; local 978 discard_page_if_needed(struct vdo_page_cache *cache) argument 1009 struct vdo_page_cache *cache = info->cache; local 1049 struct vdo_page_cache *cache = info->cache; local 1095 struct vdo_page_cache *cache = ((struct page_info *) flush_completion->parent)->cache; local 1146 struct vdo_page_cache *cache; local 1215 struct vdo_page_cache *cache = &zone->page_cache; local 1329 vdo_invalidate_page_cache(struct vdo_page_cache *cache) argument 2823 struct vdo_page_cache *cache = &zone->page_cache; local [all...] |
/linux-master/arch/x86/entry/ |
H A D | syscall_x32.c | 6 #include <linux/cache.h>
|
/linux-master/arch/arm/boot/compressed/ |
H A D | head-xscale.S | 17 @ Data cache might be active. 18 @ Be sure to flush kernel binary out of the cache, 21 @ memory to be sure we hit the same cache.
|
/linux-master/drivers/block/ |
H A D | ps3vram.c | 83 struct ps3vram_cache cache; member in struct:ps3vram_priv 311 struct ps3vram_cache *cache = &priv->cache; local 313 if (!(cache->tags[entry].flags & CACHE_PAGE_DIRTY)) 317 cache->tags[entry].address); 318 if (ps3vram_upload(dev, CACHE_OFFSET + entry * cache->page_size, 319 cache->tags[entry].address, DMA_PAGE_SIZE, 320 cache->page_size / DMA_PAGE_SIZE) < 0) { 323 entry * cache->page_size, cache 333 struct ps3vram_cache *cache = &priv->cache; local 353 struct ps3vram_cache *cache = &priv->cache; local 367 struct ps3vram_cache *cache = &priv->cache; local [all...] |
/linux-master/net/mac80211/ |
H A D | mesh_pathtbl.c | 55 struct mesh_tx_cache *cache; local 57 cache = &sdata->u.mesh.tx_cache; 58 rhashtable_free_and_destroy(&cache->rht, 64 struct mesh_tx_cache *cache; local 66 cache = &sdata->u.mesh.tx_cache; 67 rhashtable_init(&cache->rht, &fast_tx_rht_params); 68 INIT_HLIST_HEAD(&cache->walk_head); 69 spin_lock_init(&cache->walk_lock); 425 static void mesh_fast_tx_entry_free(struct mesh_tx_cache *cache, argument 429 rhashtable_remove_fast(&cache 438 struct mesh_tx_cache *cache; local 471 struct mesh_tx_cache *cache; local 610 struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache; local 627 struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache; local 641 struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache; local 655 struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache; local [all...] |
/linux-master/drivers/net/ethernet/microchip/vcap/ |
H A D | vcap_api_debugfs_kunit.c | 96 memset(admin->cache.keystream, 0, test_cache_erase_count); 97 memset(admin->cache.maskstream, 0, test_cache_erase_count); 98 memset(admin->cache.actionstream, 0, test_cache_erase_count); 121 keystr = &admin->cache.keystream[start]; 122 mskstr = &admin->cache.maskstream[start]; 135 actstr = &admin->cache.actionstream[start]; 144 admin->cache.counter = test_hw_cache.counter; 145 admin->cache.sticky = test_hw_cache.sticky; 162 keystr = &admin->cache.keystream[start]; 163 mskstr = &admin->cache [all...] |
/linux-master/fs/netfs/ |
H A D | fscache_volume.c | 2 /* Volume-level cache cookie handling. 39 * Pin the cache behind a volume so that we can access it. 55 * fscache_begin_volume_access - Pin a cache so a volume can be accessed 60 * Attempt to pin the cache to prevent it from going away whilst we're 63 * (1) If the cache tests as not live (state is not FSCACHE_CACHE_IS_ACTIVE), 66 * (2) If the cache tests as live, then we increment the volume's n_accesses 67 * count and then recheck the cache liveness, ending the access if it 73 * (4) Whilst the cache is caching, the volume's n_accesses is kept 76 * (5) When the cache is taken offline, the state is changed to prevent new 87 if (!fscache_cache_is_live(volume->cache)) 205 struct fscache_cache *cache; local 380 struct fscache_cache *cache = volume->cache; local [all...] |
H A D | internal.h | 12 #include <linux/fscache-cache.h> 157 struct fscache_cookie *cookie = ctx->cache; 195 * fscache-cache.c 200 bool fscache_begin_cache_access(struct fscache_cache *cache, enum fscache_access_trace why); 201 void fscache_end_cache_access(struct fscache_cache *cache, enum fscache_access_trace why); 203 void fscache_put_cache(struct fscache_cache *cache, enum fscache_cache_trace where); 205 static inline enum fscache_cache_state fscache_cache_state(const struct fscache_cache *cache) argument 207 return smp_load_acquire(&cache->state); 210 static inline bool fscache_cache_is_live(const struct fscache_cache *cache) argument 212 return fscache_cache_state(cache) 215 fscache_set_cache_state(struct fscache_cache *cache, enum fscache_cache_state new_state) argument 222 fscache_set_cache_state_maybe(struct fscache_cache *cache, enum fscache_cache_state old_state, enum fscache_cache_state new_state) argument [all...] |
/linux-master/drivers/soc/tegra/cbb/ |
H A D | tegra-cbb.c | 36 void tegra_cbb_print_cache(struct seq_file *file, u32 cache) argument 40 buff_str = (cache & BIT(0)) ? "Bufferable " : ""; 41 mod_str = (cache & BIT(1)) ? "Modifiable " : ""; 42 rd_str = (cache & BIT(2)) ? "Read-Allocate " : ""; 43 wr_str = (cache & BIT(3)) ? "Write-Allocate" : ""; 45 if (cache == 0x0) 49 cache, buff_str, mod_str, rd_str, wr_str);
|
/linux-master/fs/nilfs2/ |
H A D | alloc.c | 222 * cache contents during the get block call. 264 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local 269 bhp, &cache->prev_desc, &cache->lock); 283 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local 288 &cache->prev_bitmap, &cache->lock); 299 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local 304 &cache->prev_bitmap, &cache 317 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local 332 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local 839 nilfs_palloc_setup_cache(struct inode *inode, struct nilfs_palloc_cache *cache) argument 848 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local [all...] |
/linux-master/drivers/net/xen-netback/ |
H A D | hash.c | 50 spin_lock_irqsave(&vif->hash.cache.lock, flags); 54 list_for_each_entry_rcu(entry, &vif->hash.cache.list, link, 55 lockdep_is_held(&vif->hash.cache.lock)) { 65 new->seq = atomic_inc_return(&vif->hash.cache.seq); 66 list_add_rcu(&new->link, &vif->hash.cache.list); 68 if (++vif->hash.cache.count > xenvif_hash_cache_size) { 70 vif->hash.cache.count--; 75 spin_unlock_irqrestore(&vif->hash.cache.lock, flags); 104 spin_lock_irqsave(&vif->hash.cache.lock, flags); 106 list_for_each_entry_rcu(entry, &vif->hash.cache [all...] |
/linux-master/fs/btrfs/ |
H A D | backref.h | 112 * cache which maps the logical address (bytenr) of leaves to an array 195 * So keep a small cache with the lookup results for the extent pointed 196 * by the last few file extent items. This cache is checked, with a 306 * Backref cache related structures 313 * Represent a tree block in the backref cache 340 /* 1 if no child node is in the cache */ 357 * For generic purpose backref cache, where we only care if it's a reloc 385 /* Red black tree of all backref nodes in the cache */ 406 /* List of unchecked backref edges during backref cache build */ 409 /* List of useless backref nodes during backref cache buil [all...] |
/linux-master/block/ |
H A D | bio.c | 416 static void bio_alloc_irq_cache_splice(struct bio_alloc_cache *cache) argument 420 /* cache->free_list must be empty */ 421 if (WARN_ON_ONCE(cache->free_list)) 425 cache->free_list = cache->free_list_irq; 426 cache->free_list_irq = NULL; 427 cache->nr += cache->nr_irq; 428 cache->nr_irq = 0; 436 struct bio_alloc_cache *cache; local 703 __bio_alloc_cache_prune(struct bio_alloc_cache *cache, unsigned int nr) argument 719 bio_alloc_cache_prune(struct bio_alloc_cache *cache, unsigned int nr) argument 735 struct bio_alloc_cache *cache = per_cpu_ptr(bs->cache, cpu); local 751 struct bio_alloc_cache *cache; local 762 struct bio_alloc_cache *cache; local [all...] |
/linux-master/drivers/iio/dac/ |
H A D | ti-dac7612.c | 23 uint16_t cache[2]; member in struct:dac7612 34 * DMA (thus cache coherency maintenance) may require the 35 * transfer buffers to live in their own cache lines. 48 priv->cache[channel] = val; 84 *val = priv->cache[chan->channel]; 109 if (val == priv->cache[chan->channel]) 153 iio_dev->num_channels = ARRAY_SIZE(priv->cache); 158 for (i = 0; i < ARRAY_SIZE(priv->cache); i++) {
|
/linux-master/arch/arm/mm/ |
H A D | cache-v7.S | 3 * linux/arch/arm/mm/cache-v7.S 15 #include <asm/hardware/cache-b15-rac.h> 33 * of cache lines with uninitialized data and uninitialized tags to get 43 mcr p15, 2, r0, c0, c0, 0 @ select L1 data cache in CSSELR 45 mrc p15, 1, r0, c0, c0, 0 @ read cache geometry from CCSIDR 68 mrc p15, 1, r0, c0, c0, 0 @ re-read cache geometry from CCSIDR 78 * Flush the whole I-cache. 85 ALT_SMP(mcr p15, 0, r0, c7, c1, 0) @ invalidate I-cache inner shareable 86 ALT_UP(mcr p15, 0, r0, c7, c5, 0) @ I+BTB cache invalidate 93 * Flush the D-cache u [all...] |
/linux-master/security/selinux/ss/ |
H A D | sidtab.c | 94 isid->entry.cache = NULL; 512 kfree(rcu_dereference_raw(entry->cache)); 566 struct sidtab_str_cache *cache, *victim = NULL; local 569 /* do not cache invalid contexts */ 575 cache = rcu_dereference_protected(entry->cache, 577 if (cache) { 578 /* entry in cache - just bump to the head of LRU list */ 579 list_move(&cache->lru_member, &s->cache_lru_list); 583 cache 611 struct sidtab_str_cache *cache; local [all...] |