Searched refs:cache (Results 101 - 125 of 961) sorted by relevance

1234567891011>>

/linux-master/fs/cachefiles/
H A DMakefile7 cache.o \
H A Dio.c43 * Handle completion of a read from the cache.
71 * Initiate a read from the cache.
194 * Query the occupancy of the cache in a region, returning where the next chunk
213 granularity = max_t(size_t, object->volume->cache->bsize, granularity);
252 * Handle completion of a write to the cache.
269 atomic_long_sub(ki->b_writing, &object->volume->cache->b_writing);
277 * Initiate a write to the cache.
286 struct cachefiles_cache *cache; local
293 cache = object->volume->cache;
384 struct cachefiles_cache *cache; local
523 struct cachefiles_cache *cache = object->volume->cache; local
606 struct cachefiles_cache *cache = object->volume->cache; local
[all...]
/linux-master/arch/arm/mach-mvebu/
H A Ddove.c12 #include <asm/hardware/cache-tauros2.h>
/linux-master/arch/loongarch/mm/
H A DMakefile6 obj-y += init.o cache.o tlb.o tlbex.o extable.o \
/linux-master/arch/m68k/include/asm/
H A Dsyscalls.h8 asmlinkage int sys_cacheflush(unsigned long addr, int scope, int cache,
/linux-master/arch/powerpc/kernel/
H A Dfirmware.c13 #include <linux/cache.h>
/linux-master/drivers/gpu/drm/imagination/
H A Dpvr_params.h9 #include <linux/cache.h>
/linux-master/net/core/
H A Dhotdata.c3 #include <linux/cache.h>
/linux-master/drivers/md/dm-vdo/
H A Dcpu.h9 #include <linux/cache.h>
12 * uds_prefetch_address() - Minimize cache-miss latency by attempting to move data into a CPU cache
35 * uds_prefetch_range() - Minimize cache-miss latency by attempting to move a range of addresses
36 * into a CPU cache before they are accessed.
46 * Count the number of cache lines to fetch, allowing for the address range to span an
47 * extra cache line boundary due to address alignment.
H A Dblock-map.c42 * In the current (newest) era, block map pages are not written unless there is cache pressure. In
121 * For adjusting VDO page cache statistic fields which are only mutated on the logical zone thread.
158 struct vdo_page_cache *cache = info->cache; local
160 return &cache->pages[(info - cache->infos) * VDO_BLOCK_SIZE];
180 static int initialize_info(struct vdo_page_cache *cache) argument
184 INIT_LIST_HEAD(&cache->free_list);
185 for (info = cache->infos; info < cache
219 allocate_cache_components(struct vdo_page_cache *cache) argument
244 assert_on_cache_thread(struct vdo_page_cache *cache, const char *function_name) argument
255 assert_io_allowed(struct vdo_page_cache *cache) argument
262 report_cache_pressure(struct vdo_page_cache *cache) argument
378 struct vdo_page_cache *cache = info->cache; local
424 find_free_page(struct vdo_page_cache *cache) argument
442 find_page(struct vdo_page_cache *cache, physical_block_number_t pbn) argument
462 select_lru_page(struct vdo_page_cache *cache) argument
558 set_persistent_error(struct vdo_page_cache *cache, const char *context, int result) argument
671 struct vdo_page_cache *cache = info->cache; local
696 struct vdo_page_cache *cache = info->cache; local
738 struct vdo_page_cache *cache = info->cache; local
771 struct vdo_page_cache *cache = info->cache; local
813 save_pages(struct vdo_page_cache *cache) argument
891 struct vdo_page_cache *cache = info->cache; local
938 discard_a_page(struct vdo_page_cache *cache) argument
966 struct vdo_page_cache *cache = vdo_page_comp->cache; local
978 discard_page_if_needed(struct vdo_page_cache *cache) argument
1009 struct vdo_page_cache *cache = info->cache; local
1049 struct vdo_page_cache *cache = info->cache; local
1095 struct vdo_page_cache *cache = ((struct page_info *) flush_completion->parent)->cache; local
1146 struct vdo_page_cache *cache; local
1215 struct vdo_page_cache *cache = &zone->page_cache; local
1329 vdo_invalidate_page_cache(struct vdo_page_cache *cache) argument
2823 struct vdo_page_cache *cache = &zone->page_cache; local
[all...]
/linux-master/arch/x86/entry/
H A Dsyscall_x32.c6 #include <linux/cache.h>
/linux-master/arch/arm/boot/compressed/
H A Dhead-xscale.S17 @ Data cache might be active.
18 @ Be sure to flush kernel binary out of the cache,
21 @ memory to be sure we hit the same cache.
/linux-master/drivers/block/
H A Dps3vram.c83 struct ps3vram_cache cache; member in struct:ps3vram_priv
311 struct ps3vram_cache *cache = &priv->cache; local
313 if (!(cache->tags[entry].flags & CACHE_PAGE_DIRTY))
317 cache->tags[entry].address);
318 if (ps3vram_upload(dev, CACHE_OFFSET + entry * cache->page_size,
319 cache->tags[entry].address, DMA_PAGE_SIZE,
320 cache->page_size / DMA_PAGE_SIZE) < 0) {
323 entry * cache->page_size, cache
333 struct ps3vram_cache *cache = &priv->cache; local
353 struct ps3vram_cache *cache = &priv->cache; local
367 struct ps3vram_cache *cache = &priv->cache; local
[all...]
/linux-master/net/mac80211/
H A Dmesh_pathtbl.c55 struct mesh_tx_cache *cache; local
57 cache = &sdata->u.mesh.tx_cache;
58 rhashtable_free_and_destroy(&cache->rht,
64 struct mesh_tx_cache *cache; local
66 cache = &sdata->u.mesh.tx_cache;
67 rhashtable_init(&cache->rht, &fast_tx_rht_params);
68 INIT_HLIST_HEAD(&cache->walk_head);
69 spin_lock_init(&cache->walk_lock);
425 static void mesh_fast_tx_entry_free(struct mesh_tx_cache *cache, argument
429 rhashtable_remove_fast(&cache
438 struct mesh_tx_cache *cache; local
471 struct mesh_tx_cache *cache; local
610 struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache; local
627 struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache; local
641 struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache; local
655 struct mesh_tx_cache *cache = &sdata->u.mesh.tx_cache; local
[all...]
/linux-master/drivers/net/ethernet/microchip/vcap/
H A Dvcap_api_debugfs_kunit.c96 memset(admin->cache.keystream, 0, test_cache_erase_count);
97 memset(admin->cache.maskstream, 0, test_cache_erase_count);
98 memset(admin->cache.actionstream, 0, test_cache_erase_count);
121 keystr = &admin->cache.keystream[start];
122 mskstr = &admin->cache.maskstream[start];
135 actstr = &admin->cache.actionstream[start];
144 admin->cache.counter = test_hw_cache.counter;
145 admin->cache.sticky = test_hw_cache.sticky;
162 keystr = &admin->cache.keystream[start];
163 mskstr = &admin->cache
[all...]
/linux-master/fs/netfs/
H A Dfscache_volume.c2 /* Volume-level cache cookie handling.
39 * Pin the cache behind a volume so that we can access it.
55 * fscache_begin_volume_access - Pin a cache so a volume can be accessed
60 * Attempt to pin the cache to prevent it from going away whilst we're
63 * (1) If the cache tests as not live (state is not FSCACHE_CACHE_IS_ACTIVE),
66 * (2) If the cache tests as live, then we increment the volume's n_accesses
67 * count and then recheck the cache liveness, ending the access if it
73 * (4) Whilst the cache is caching, the volume's n_accesses is kept
76 * (5) When the cache is taken offline, the state is changed to prevent new
87 if (!fscache_cache_is_live(volume->cache))
205 struct fscache_cache *cache; local
380 struct fscache_cache *cache = volume->cache; local
[all...]
H A Dinternal.h12 #include <linux/fscache-cache.h>
157 struct fscache_cookie *cookie = ctx->cache;
195 * fscache-cache.c
200 bool fscache_begin_cache_access(struct fscache_cache *cache, enum fscache_access_trace why);
201 void fscache_end_cache_access(struct fscache_cache *cache, enum fscache_access_trace why);
203 void fscache_put_cache(struct fscache_cache *cache, enum fscache_cache_trace where);
205 static inline enum fscache_cache_state fscache_cache_state(const struct fscache_cache *cache) argument
207 return smp_load_acquire(&cache->state);
210 static inline bool fscache_cache_is_live(const struct fscache_cache *cache) argument
212 return fscache_cache_state(cache)
215 fscache_set_cache_state(struct fscache_cache *cache, enum fscache_cache_state new_state) argument
222 fscache_set_cache_state_maybe(struct fscache_cache *cache, enum fscache_cache_state old_state, enum fscache_cache_state new_state) argument
[all...]
/linux-master/drivers/soc/tegra/cbb/
H A Dtegra-cbb.c36 void tegra_cbb_print_cache(struct seq_file *file, u32 cache) argument
40 buff_str = (cache & BIT(0)) ? "Bufferable " : "";
41 mod_str = (cache & BIT(1)) ? "Modifiable " : "";
42 rd_str = (cache & BIT(2)) ? "Read-Allocate " : "";
43 wr_str = (cache & BIT(3)) ? "Write-Allocate" : "";
45 if (cache == 0x0)
49 cache, buff_str, mod_str, rd_str, wr_str);
/linux-master/fs/nilfs2/
H A Dalloc.c222 * cache contents during the get block call.
264 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local
269 bhp, &cache->prev_desc, &cache->lock);
283 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local
288 &cache->prev_bitmap, &cache->lock);
299 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local
304 &cache->prev_bitmap, &cache
317 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local
332 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local
839 nilfs_palloc_setup_cache(struct inode *inode, struct nilfs_palloc_cache *cache) argument
848 struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; local
[all...]
/linux-master/drivers/net/xen-netback/
H A Dhash.c50 spin_lock_irqsave(&vif->hash.cache.lock, flags);
54 list_for_each_entry_rcu(entry, &vif->hash.cache.list, link,
55 lockdep_is_held(&vif->hash.cache.lock)) {
65 new->seq = atomic_inc_return(&vif->hash.cache.seq);
66 list_add_rcu(&new->link, &vif->hash.cache.list);
68 if (++vif->hash.cache.count > xenvif_hash_cache_size) {
70 vif->hash.cache.count--;
75 spin_unlock_irqrestore(&vif->hash.cache.lock, flags);
104 spin_lock_irqsave(&vif->hash.cache.lock, flags);
106 list_for_each_entry_rcu(entry, &vif->hash.cache
[all...]
/linux-master/fs/btrfs/
H A Dbackref.h112 * cache which maps the logical address (bytenr) of leaves to an array
195 * So keep a small cache with the lookup results for the extent pointed
196 * by the last few file extent items. This cache is checked, with a
306 * Backref cache related structures
313 * Represent a tree block in the backref cache
340 /* 1 if no child node is in the cache */
357 * For generic purpose backref cache, where we only care if it's a reloc
385 /* Red black tree of all backref nodes in the cache */
406 /* List of unchecked backref edges during backref cache build */
409 /* List of useless backref nodes during backref cache buil
[all...]
/linux-master/block/
H A Dbio.c416 static void bio_alloc_irq_cache_splice(struct bio_alloc_cache *cache) argument
420 /* cache->free_list must be empty */
421 if (WARN_ON_ONCE(cache->free_list))
425 cache->free_list = cache->free_list_irq;
426 cache->free_list_irq = NULL;
427 cache->nr += cache->nr_irq;
428 cache->nr_irq = 0;
436 struct bio_alloc_cache *cache; local
703 __bio_alloc_cache_prune(struct bio_alloc_cache *cache, unsigned int nr) argument
719 bio_alloc_cache_prune(struct bio_alloc_cache *cache, unsigned int nr) argument
735 struct bio_alloc_cache *cache = per_cpu_ptr(bs->cache, cpu); local
751 struct bio_alloc_cache *cache; local
762 struct bio_alloc_cache *cache; local
[all...]
/linux-master/drivers/iio/dac/
H A Dti-dac7612.c23 uint16_t cache[2]; member in struct:dac7612
34 * DMA (thus cache coherency maintenance) may require the
35 * transfer buffers to live in their own cache lines.
48 priv->cache[channel] = val;
84 *val = priv->cache[chan->channel];
109 if (val == priv->cache[chan->channel])
153 iio_dev->num_channels = ARRAY_SIZE(priv->cache);
158 for (i = 0; i < ARRAY_SIZE(priv->cache); i++) {
/linux-master/arch/arm/mm/
H A Dcache-v7.S3 * linux/arch/arm/mm/cache-v7.S
15 #include <asm/hardware/cache-b15-rac.h>
33 * of cache lines with uninitialized data and uninitialized tags to get
43 mcr p15, 2, r0, c0, c0, 0 @ select L1 data cache in CSSELR
45 mrc p15, 1, r0, c0, c0, 0 @ read cache geometry from CCSIDR
68 mrc p15, 1, r0, c0, c0, 0 @ re-read cache geometry from CCSIDR
78 * Flush the whole I-cache.
85 ALT_SMP(mcr p15, 0, r0, c7, c1, 0) @ invalidate I-cache inner shareable
86 ALT_UP(mcr p15, 0, r0, c7, c5, 0) @ I+BTB cache invalidate
93 * Flush the D-cache u
[all...]
/linux-master/security/selinux/ss/
H A Dsidtab.c94 isid->entry.cache = NULL;
512 kfree(rcu_dereference_raw(entry->cache));
566 struct sidtab_str_cache *cache, *victim = NULL; local
569 /* do not cache invalid contexts */
575 cache = rcu_dereference_protected(entry->cache,
577 if (cache) {
578 /* entry in cache - just bump to the head of LRU list */
579 list_move(&cache->lru_member, &s->cache_lru_list);
583 cache
611 struct sidtab_str_cache *cache; local
[all...]

Completed in 251 milliseconds

1234567891011>>