/linux-master/drivers/acpi/acpica/ |
H A D | utcache.c | 4 * Module Name: utcache - local cache allocation routines 21 * PARAMETERS: cache_name - Ascii name for the cache 23 * max_depth - Maximum depth of the cache (in objects) 24 * return_cache - Where the new cache object is returned 28 * DESCRIPTION: Create a cache object 36 struct acpi_memory_list *cache; local 44 /* Create the cache object */ 46 cache = acpi_os_allocate(sizeof(struct acpi_memory_list)); 47 if (!cache) { 51 /* Populate the cache objec 74 acpi_os_purge_cache(struct acpi_memory_list *cache) argument 120 acpi_os_delete_cache(struct acpi_memory_list *cache) argument 153 acpi_os_release_object(struct acpi_memory_list *cache, void *object) argument 208 acpi_os_acquire_object(struct acpi_memory_list *cache) argument [all...] |
/linux-master/io_uring/ |
H A D | alloc_cache.h | 5 * Don't allow the cache to grow beyond this size. 13 static inline bool io_alloc_cache_put(struct io_alloc_cache *cache, argument 16 if (cache->nr_cached < cache->max_cached) { 17 cache->nr_cached++; 18 wq_stack_add_head(&entry->node, &cache->list); 25 static inline bool io_alloc_cache_empty(struct io_alloc_cache *cache) argument 27 return !cache->list.next; 30 static inline struct io_cache_entry *io_alloc_cache_get(struct io_alloc_cache *cache) argument 32 if (cache 45 io_alloc_cache_init(struct io_alloc_cache *cache, unsigned max_nr, size_t size) argument 54 io_alloc_cache_free(struct io_alloc_cache *cache, void (*free)(struct io_cache_entry *)) argument [all...] |
/linux-master/drivers/md/ |
H A D | dm-cache-target.c | 11 #include "dm-cache-metadata.h" 24 #define DM_MSG_PREFIX "cache" 27 "A percentage of time allocated for copying to and/or from cache"); 35 * cblock: index of a cache block 36 * promotion: movement of a block from origin to cache 37 * demotion: movement of a block from cache to origin 38 * migration: movement of a block between the origin and cache device, 241 * The block size of the device holding cache data must be 256 * dirty. If you lose the cache device you will lose data. 262 * Data is written to both cache an 297 struct cache { struct 421 struct cache *cache; member in struct:dm_cache_migration 433 writethrough_mode(struct cache *cache) argument 438 writeback_mode(struct cache *cache) argument 443 passthrough_mode(struct cache *cache) argument 450 wake_deferred_bio_worker(struct cache *cache) argument 455 wake_migration_worker(struct cache *cache) argument 465 alloc_prison_cell(struct cache *cache) argument 470 free_prison_cell(struct cache *cache, struct dm_bio_prison_cell_v2 *cell) argument 475 alloc_migration(struct cache *cache) argument 491 struct cache *cache = mg->cache; local 556 defer_bio(struct cache *cache, struct bio *bio) argument 565 defer_bios(struct cache *cache, struct bio_list *bios) argument 577 bio_detain_shared(struct cache *cache, dm_oblock_t oblock, struct bio *bio) argument 608 is_dirty(struct cache *cache, dm_cblock_t b) argument 613 set_dirty(struct cache *cache, dm_cblock_t cblock) argument 625 force_set_dirty(struct cache *cache, dm_cblock_t cblock) argument 632 force_clear_dirty(struct cache *cache, dm_cblock_t cblock) argument 644 block_size_is_power_of_two(struct cache *cache) argument 656 oblocks_per_dblock(struct cache *cache) argument 668 oblock_to_dblock(struct cache *cache, dm_oblock_t oblock) argument 674 set_discard(struct cache *cache, dm_dblock_t b) argument 684 clear_discard(struct cache *cache, dm_dblock_t b) argument 691 is_discarded(struct cache *cache, dm_dblock_t b) argument 702 is_discarded_oblock(struct cache *cache, dm_oblock_t b) argument 719 remap_to_origin(struct cache *cache, struct bio *bio) argument 724 remap_to_cache(struct cache *cache, struct bio *bio, dm_cblock_t cblock) argument 741 check_if_tick_bio_needed(struct cache *cache, struct bio *bio) argument 755 remap_to_origin_clear_discard(struct cache *cache, struct bio *bio, dm_oblock_t oblock) argument 765 remap_to_cache_dirty(struct cache *cache, struct bio *bio, dm_oblock_t oblock, dm_cblock_t cblock) argument 776 get_bio_block(struct cache *cache, struct bio *bio) argument 788 accountable_bio(struct cache *cache, struct bio *bio) argument 793 accounted_begin(struct cache *cache, struct bio *bio) argument 804 accounted_complete(struct cache *cache, struct bio *bio) argument 811 accounted_request(struct cache *cache, struct bio *bio) argument 819 struct cache *cache = context; local 828 remap_to_origin_and_cache(struct cache *cache, struct bio *bio, dm_oblock_t oblock, dm_cblock_t cblock) argument 850 get_cache_mode(struct cache *cache) argument 855 cache_device_name(struct cache *cache) argument 860 notify_mode_switch(struct cache *cache, enum cache_metadata_mode mode) argument 873 set_cache_mode(struct cache *cache, enum cache_metadata_mode new_mode) argument 914 abort_transaction(struct cache *cache) argument 933 metadata_operation_failed(struct cache *cache, const char *op, int r) argument 943 load_stats(struct cache *cache) argument 954 save_stats(struct cache *cache) argument 994 inc_io_migrations(struct cache *cache) argument 999 dec_io_migrations(struct cache *cache) argument 1009 calc_discard_block_range(struct cache *cache, struct bio *bio, dm_dblock_t *b, dm_dblock_t *e) argument 1025 prevent_background_work(struct cache *cache) argument 1032 allow_background_work(struct cache *cache) argument 1039 background_work_begin(struct cache *cache) argument 1050 background_work_end(struct cache *cache) argument 1059 bio_writes_complete_block(struct cache *cache, struct bio *bio) argument 1065 optimisable_bio(struct cache *cache, struct bio *bio, dm_oblock_t block) argument 1098 struct cache *cache = mg->cache; local 1114 bio_drop_shared_lock(struct cache *cache, struct bio *bio) argument 1126 struct cache *cache = mg->cache; local 1172 struct cache *cache = mg->cache; local 1241 struct cache *cache = mg->cache; local 1343 struct cache *cache = mg->cache; local 1397 struct cache *cache = mg->cache; local 1428 mg_start(struct cache *cache, struct policy_work *op, struct bio *bio) argument 1457 struct cache *cache = mg->cache; local 1479 invalidate_cblock(struct cache *cache, dm_cblock_t cblock) argument 1508 struct cache *cache = mg->cache; local 1527 struct cache *cache = mg->cache; local 1559 invalidate_start(struct cache *cache, dm_cblock_t cblock, dm_oblock_t oblock, struct bio *bio) argument 1587 spare_migration_bandwidth(struct cache *cache) argument 1599 inc_hit_counter(struct cache *cache, struct bio *bio) argument 1605 inc_miss_counter(struct cache *cache, struct bio *bio) argument 1613 map_bio(struct cache *cache, struct bio *bio, dm_oblock_t block, bool *commit_needed) argument 1728 process_bio(struct cache *cache, struct bio *bio) argument 1741 commit(struct cache *cache, bool clean_shutdown) argument 1761 struct cache *cache = context; local 1771 process_flush_bio(struct cache *cache, struct bio *bio) argument 1784 process_discard_bio(struct cache *cache, struct bio *bio) argument 1810 struct cache *cache = container_of(ws, struct cache, deferred_bio_worker); local 1844 requeue_deferred_bios(struct cache *cache) argument 1866 struct cache *cache = container_of(to_delayed_work(ws), struct cache, waker); local 1878 struct cache *cache = container_of(ws, struct cache, migration_worker); local 1912 destroy(struct cache *cache) argument 1960 struct cache *cache = ti->private; local 2268 process_config_option(struct cache *cache, const char *key, const char *value) argument 2283 set_config_value(struct cache *cache, const char *key, const char *value) argument 2296 set_config_values(struct cache *cache, int argc, const char **argv) argument 2317 create_cache_policy(struct cache *cache, struct cache_args *ca, char **error) argument 2360 set_cache_size(struct cache *cache, dm_cblock_t size) argument 2379 struct cache *cache; local 2572 copy_ctr_args(struct cache *cache, int argc, const char **argv) argument 2600 struct cache *cache = NULL; local 2633 struct cache *cache = ti->private; local 2665 struct cache *cache = ti->private; local 2683 write_dirty_bitset(struct cache *cache) argument 2697 write_discard_bitset(struct cache *cache) argument 2724 write_hints(struct cache *cache) argument 2743 sync_metadata(struct cache *cache) argument 2775 struct cache *cache = ti->private; local 2797 struct cache *cache = context; local 2815 struct cache *cache; member in struct:discard_load_info 2825 discard_load_info_init(struct cache *cache, struct discard_load_info *li) argument 2892 get_cache_dev_size(struct cache *cache) argument 2899 can_resize(struct cache *cache, dm_cblock_t new_size) argument 2925 resize_cache_dev(struct cache *cache, dm_cblock_t new_size) argument 2944 struct cache *cache = ti->private; local 3005 struct cache *cache = ti->private; local 3012 emit_flags(struct cache *cache, char *result, unsigned int maxlen, ssize_t *sz_ptr) argument 3065 struct cache *cache = ti->private; local 3196 parse_cblock_range(struct cache *cache, const char *str, struct cblock_range *result) argument 3233 validate_cblock_range(struct cache *cache, struct cblock_range *range) argument 3265 request_invalidation(struct cache *cache, struct cblock_range *range) argument 3287 process_invalidate_cblocks_message(struct cache *cache, unsigned int count, const char **cblock_ranges) argument 3331 struct cache *cache = ti->private; local 3355 struct cache *cache = ti->private; local 3368 disable_passdown_if_not_supported(struct cache *cache) argument 3390 set_discard_limits(struct cache *cache, struct queue_limits *limits) argument 3416 struct cache *cache = ti->private; local [all...] |
/linux-master/arch/powerpc/kernel/ |
H A D | cacheinfo.c | 3 * Processor cache information made available to userspace via sysfs; 27 * - a "cache" kobject for the top-level directory 28 * - a list of "index" objects representing the cpu's local cache hierarchy 31 struct kobject *kobj; /* bare (not embedded) kobject for cache 36 /* "index" object: each cpu's cache directory has an index 37 * subdirectory corresponding to a cache object associated with the 43 struct cache *cache; member in struct:cache_index_dir 47 * cache type */ 52 /* Allow for both [di]-cache 117 struct cache { struct 139 cache_type_string(const struct cache *cache) argument 144 cache_init(struct cache *cache, int type, int level, struct device_node *ofnode, int group_id) argument 158 struct cache *cache; local 167 release_cache_debugcheck(struct cache *cache) argument 180 release_cache(struct cache *cache) argument 194 cache_cpu_set(struct cache *cache, int cpu) argument 208 cache_size(const struct cache *cache, unsigned int *ret) argument 223 cache_size_kb(const struct cache *cache, unsigned int *ret) argument 235 cache_get_line_size(const struct cache *cache, unsigned int *ret) argument 258 cache_nr_sets(const struct cache *cache, unsigned int *ret) argument 273 cache_associativity(const struct cache *cache, unsigned int *ret) argument 305 cache_find_first_sibling(struct cache *cache) argument 326 struct cache *cache = NULL; local 394 struct cache *cache; local 408 struct cache *cache; local 442 do_subsidiary_caches_debugcheck(struct cache *cache) argument 477 do_subsidiary_caches(struct cache *cache, unsigned int cpu_id) argument 592 struct cache *cache; local 609 struct cache *cache; local 625 struct cache *cache; local 641 struct cache *cache; local 656 struct cache *cache; local 669 struct cache *cache; local 684 struct cache *cache; local 747 struct cache *cache; local 785 cacheinfo_create_index_dir(struct cache *cache, int index, struct cache_dir *cache_dir) argument 814 struct cache *cache; local 831 struct cache *cache; local 848 struct cache *cache; local 890 cache_cpu_clear(struct cache *cache, int cpu) argument 914 struct cache *cache; local [all...] |
/linux-master/fs/netfs/ |
H A D | fscache_cache.c | 2 /* FS-Cache cache handling 22 * Allocate a cache cookie. 26 struct fscache_cache *cache; local 28 cache = kzalloc(sizeof(*cache), GFP_KERNEL); 29 if (cache) { 31 cache->name = kstrdup(name, GFP_KERNEL); 32 if (!cache->name) { 33 kfree(cache); 37 refcount_set(&cache 44 fscache_get_cache_maybe(struct fscache_cache *cache, enum fscache_cache_trace where) argument 61 struct fscache_cache *candidate, *cache, *unnamed = NULL; local 151 struct fscache_cache *cache; local 179 fscache_put_cache(struct fscache_cache *cache, enum fscache_cache_trace where) argument 209 fscache_relinquish_cache(struct fscache_cache *cache) argument 234 fscache_add_cache(struct fscache_cache *cache, const struct fscache_cache_ops *ops, void *cache_priv) argument 289 fscache_begin_cache_access(struct fscache_cache *cache, enum fscache_access_trace why) argument 315 fscache_end_cache_access(struct fscache_cache *cache, enum fscache_access_trace why) argument 337 fscache_io_error(struct fscache_cache *cache) argument 355 fscache_withdraw_cache(struct fscache_cache *cache) argument 382 struct fscache_cache *cache; local [all...] |
/linux-master/fs/cachefiles/ |
H A D | cache.c | 2 /* Manage high-level VFS aspects of a cache. 14 * Bring a cache online. 16 int cachefiles_add_cache(struct cachefiles_cache *cache) argument 27 cache_cookie = fscache_acquire_cache(cache->tag); 32 ret = cachefiles_get_security_ID(cache); 36 cachefiles_begin_secure(cache, &saved_cred); 38 /* look up the directory at the root of the cache */ 39 ret = kern_path(cache->rootdirname, LOOKUP_DIRECTORY, &path); 43 cache->mnt = path.mnt; 48 pr_warn("File cache o 184 cachefiles_has_space(struct cachefiles_cache *cache, unsigned fnr, unsigned bnr, enum cachefiles_has_space_for reason) argument 287 cachefiles_withdraw_objects(struct cachefiles_cache *cache) argument 317 cachefiles_withdraw_volumes(struct cachefiles_cache *cache) argument 343 cachefiles_sync_cache(struct cachefiles_cache *cache) argument 367 cachefiles_withdraw_cache(struct cachefiles_cache *cache) argument [all...] |
H A D | daemon.c | 61 int (*handler)(struct cachefiles_cache *cache, char *args); 87 * Prepare a cache for caching. 91 struct cachefiles_cache *cache; local 103 /* allocate a cache record */ 104 cache = kzalloc(sizeof(struct cachefiles_cache), GFP_KERNEL); 105 if (!cache) { 110 mutex_init(&cache->daemon_mutex); 111 init_waitqueue_head(&cache->daemon_pollwq); 112 INIT_LIST_HEAD(&cache->volumes); 113 INIT_LIST_HEAD(&cache 136 cachefiles_flush_reqs(struct cachefiles_cache *cache) argument 169 cachefiles_put_unbind_pincount(struct cachefiles_cache *cache) argument 178 cachefiles_get_unbind_pincount(struct cachefiles_cache *cache) argument 188 struct cachefiles_cache *cache = file->private_data; local 209 cachefiles_do_daemon_read(struct cachefiles_cache *cache, char __user *_buffer, size_t buflen) argument 260 struct cachefiles_cache *cache = file->private_data; local 282 struct cachefiles_cache *cache = file->private_data; local 358 struct cachefiles_cache *cache = file->private_data; local 392 cachefiles_daemon_range_error(struct cachefiles_cache *cache, char *args) argument 404 cachefiles_daemon_frun(struct cachefiles_cache *cache, char *args) argument 428 cachefiles_daemon_fcull(struct cachefiles_cache *cache, char *args) argument 452 cachefiles_daemon_fstop(struct cachefiles_cache *cache, char *args) argument 476 cachefiles_daemon_brun(struct cachefiles_cache *cache, char *args) argument 500 cachefiles_daemon_bcull(struct cachefiles_cache *cache, char *args) argument 524 cachefiles_daemon_bstop(struct cachefiles_cache *cache, char *args) argument 548 cachefiles_daemon_dir(struct cachefiles_cache *cache, char *args) argument 576 cachefiles_daemon_secctx(struct cachefiles_cache *cache, char *args) argument 604 cachefiles_daemon_tag(struct cachefiles_cache *cache, char *args) argument 630 cachefiles_daemon_cull(struct cachefiles_cache *cache, char *args) argument 678 cachefiles_daemon_debug(struct cachefiles_cache *cache, char *args) argument 701 cachefiles_daemon_inuse(struct cachefiles_cache *cache, char *args) argument 748 cachefiles_daemon_bind(struct cachefiles_cache *cache, char *args) argument 809 cachefiles_daemon_unbind(struct cachefiles_cache *cache) argument [all...] |
H A D | security.c | 13 * determine the security context within which we access the cache from within 16 int cachefiles_get_security_ID(struct cachefiles_cache *cache) argument 21 _enter("{%s}", cache->secctx); 29 if (cache->secctx) { 30 ret = set_security_override_from_ctx(new, cache->secctx); 39 cache->cache_cred = new; 49 static int cachefiles_check_cache_dir(struct cachefiles_cache *cache, argument 70 * check the security details of the on-disk cache 75 int cachefiles_determine_cache_security(struct cachefiles_cache *cache, argument 84 /* duplicate the cache cred [all...] |
/linux-master/mm/ |
H A D | swap_slots.c | 3 * Manage cache of swap slots to be used for and returned from 25 * The swap slots cache is protected by a mutex instead of 43 /* Serialize swap slots cache enable/disable operations */ 106 /* if global pool of slot caches too low, deactivate cache */ 115 struct swap_slots_cache *cache; local 136 cache = &per_cpu(swp_slots, cpu); 137 if (cache->slots || cache->slots_ret) { 138 /* cache already allocated */ 147 if (!cache 171 struct swap_slots_cache *cache; local 259 refill_swap_slots_cache(struct swap_slots_cache *cache) argument 274 struct swap_slots_cache *cache; local 308 struct swap_slots_cache *cache; local [all...] |
H A D | mm_slot.h | 24 static inline void *mm_slot_alloc(struct kmem_cache *cache) argument 26 if (!cache) /* initialization failed */ 28 return kmem_cache_zalloc(cache, GFP_KERNEL); 31 static inline void mm_slot_free(struct kmem_cache *cache, void *objp) argument 33 kmem_cache_free(cache, objp);
|
/linux-master/fs/btrfs/ |
H A D | lru_cache.c | 8 * Initialize a cache object. 10 * @cache: The cache. 11 * @max_size: Maximum size (number of entries) for the cache. 13 * trim the cache in that case. 15 void btrfs_lru_cache_init(struct btrfs_lru_cache *cache, unsigned int max_size) argument 17 INIT_LIST_HEAD(&cache->lru_list); 18 mt_init(&cache->entries); 19 cache->size = 0; 20 cache 45 btrfs_lru_cache_lookup(struct btrfs_lru_cache *cache, u64 key, u64 gen) argument 70 btrfs_lru_cache_remove(struct btrfs_lru_cache *cache, struct btrfs_lru_cache_entry *entry) argument 106 btrfs_lru_cache_store(struct btrfs_lru_cache *cache, struct btrfs_lru_cache_entry *new_entry, gfp_t gfp) argument 156 btrfs_lru_cache_clear(struct btrfs_lru_cache *cache) argument [all...] |
H A D | lru_cache.h | 12 * A cache entry. This is meant to be embedded in a structure of a user of 46 /* Number of entries stored in the cache. */ 48 /* Maximum number of entries the cache can have. */ 52 #define btrfs_lru_cache_for_each_entry_safe(cache, entry, tmp) \ 53 list_for_each_entry_safe_reverse((entry), (tmp), &(cache)->lru_list, lru_list) 56 struct btrfs_lru_cache *cache) 58 return list_first_entry_or_null(&cache->lru_list, 62 void btrfs_lru_cache_init(struct btrfs_lru_cache *cache, unsigned int max_size); 63 struct btrfs_lru_cache_entry *btrfs_lru_cache_lookup(struct btrfs_lru_cache *cache, 65 int btrfs_lru_cache_store(struct btrfs_lru_cache *cache, 55 btrfs_lru_cache_lru_entry( struct btrfs_lru_cache *cache) argument [all...] |
/linux-master/fs/ |
H A D | mbcache.c | 16 * Ext2 and ext4 use this cache for deduplication of extended attribute blocks. 21 * identifies a cache entry. 33 /* Maximum entries in cache to avoid degrading hash too much */ 38 /* Number of entries in cache */ 41 /* Work for shrinking when the cache has too many entries */ 47 static unsigned long mb_cache_shrink(struct mb_cache *cache, 50 static inline struct hlist_bl_head *mb_cache_entry_head(struct mb_cache *cache, argument 53 return &cache->c_hash[hash_32(key, cache->c_bucket_bits)]; 58 * in cache 74 mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key, u64 value, bool reusable) argument 127 __mb_cache_entry_free(struct mb_cache *cache, struct mb_cache_entry *entry) argument 152 __entry_find(struct mb_cache *cache, struct mb_cache_entry *entry, u32 key) argument 192 mb_cache_entry_find_first(struct mb_cache *cache, u32 key) argument 209 mb_cache_entry_find_next(struct mb_cache *cache, struct mb_cache_entry *entry) argument 222 mb_cache_entry_get(struct mb_cache *cache, u32 key, u64 value) argument 254 mb_cache_entry_delete_or_get(struct mb_cache *cache, u32 key, u64 value) argument 286 mb_cache_entry_touch(struct mb_cache *cache, struct mb_cache_entry *entry) argument 296 struct mb_cache *cache = shrink->private_data; local 302 mb_cache_shrink(struct mb_cache *cache, unsigned long nr_to_scan) argument 335 struct mb_cache *cache = shrink->private_data; local 344 struct mb_cache *cache = container_of(work, struct mb_cache, local 357 struct mb_cache *cache; local 407 mb_cache_destroy(struct mb_cache *cache) argument [all...] |
/linux-master/fs/squashfs/ |
H A D | cache.c | 8 * cache.c 15 * This file implements a generic cache implementation used for both caches, 16 * plus functions layered ontop of the generic cache implementation to 19 * To avoid out of memory and fragmentation issues with vmalloc the cache 22 * It should be noted that the cache is not used for file datablocks, these 23 * are decompressed and cached in the page-cache in the normal way. The 24 * cache is only used to temporarily cache fragment and metadata blocks 49 * Look-up block in cache, and increment usage count. If not in cache, rea 52 squashfs_cache_get(struct super_block *sb, struct squashfs_cache *cache, u64 block, int length) argument 175 struct squashfs_cache *cache = entry->cache; local 197 squashfs_cache_delete(struct squashfs_cache *cache) argument 227 struct squashfs_cache *cache = kzalloc(sizeof(*cache), GFP_KERNEL); local [all...] |
/linux-master/drivers/base/regmap/ |
H A D | regcache-flat.c | 3 // Register cache access API - flat caching support 24 unsigned int *cache; local 29 map->cache = kcalloc(regcache_flat_get_index(map, map->max_register) 31 if (!map->cache) 34 cache = map->cache; 40 cache[index] = map->reg_defaults[i].def; 48 kfree(map->cache); 49 map->cache = NULL; 57 unsigned int *cache local 68 unsigned int *cache = map->cache; local [all...] |
/linux-master/fs/btrfs/tests/ |
H A D | free-space-tests.c | 10 #include "../free-space-cache.h" 20 static int test_extents(struct btrfs_block_group *cache) argument 27 ret = btrfs_add_free_space(cache, 0, SZ_4M); 33 ret = btrfs_remove_free_space(cache, 0, SZ_4M); 39 if (test_check_exists(cache, 0, SZ_4M)) { 45 ret = btrfs_add_free_space(cache, 0, SZ_4M); 51 ret = btrfs_remove_free_space(cache, 3 * SZ_1M, SZ_1M); 57 ret = btrfs_remove_free_space(cache, 0, SZ_1M); 63 ret = btrfs_remove_free_space(cache, SZ_2M, 4096); 69 if (test_check_exists(cache, 90 test_bitmaps(struct btrfs_block_group *cache, u32 sectorsize) argument 158 test_bitmaps_and_extents(struct btrfs_block_group *cache, u32 sectorsize) argument 333 check_num_extents_and_bitmaps(const struct btrfs_block_group *cache, const int num_extents, const int num_bitmaps) argument 353 check_cache_empty(struct btrfs_block_group *cache) argument 395 test_steal_space_from_bitmap_to_extent(struct btrfs_block_group *cache, u32 sectorsize) argument 833 test_bytes_index(struct btrfs_block_group *cache, u32 sectorsize) argument 1008 struct btrfs_block_group *cache; local [all...] |
H A D | free-space-tree-tests.c | 22 struct btrfs_block_group *cache, 35 info = search_free_space_info(trans, cache, path, 0); 52 end = cache->start + cache->length; 60 bit = free_space_test_bit(cache, path, offset); 109 struct btrfs_block_group *cache, 118 info = search_free_space_info(trans, cache, path, 0); 127 ret = __check_free_space_extents(trans, fs_info, cache, path, extents, 134 ret = convert_free_space_to_extents(trans, cache, path); 140 ret = convert_free_space_to_bitmaps(trans, cache, pat 20 __check_free_space_extents(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, const struct free_space_extent * const extents, unsigned int num_extents) argument 107 check_free_space_extents(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, const struct free_space_extent * const extents, unsigned int num_extents) argument 150 test_empty_block_group(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) argument 164 test_remove_all(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) argument 185 test_remove_beginning(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) argument 208 test_remove_end(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) argument 231 test_remove_middle(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) argument 255 test_merge_left(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) argument 292 test_merge_right(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) argument 330 test_merge_both(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) argument 373 test_merge_none(struct btrfs_trans_handle *trans, struct btrfs_fs_info *fs_info, struct btrfs_block_group *cache, struct btrfs_path *path, u32 alignment) argument 429 struct btrfs_block_group *cache = NULL; local [all...] |
/linux-master/arch/xtensa/include/asm/ |
H A D | cachetype.h | 5 #include <asm/cache.h>
|
/linux-master/arch/nios2/include/asm/ |
H A D | cachetype.h | 6 #include <asm/cache.h>
|
/linux-master/include/linux/ |
H A D | mbcache.h | 20 /* List of entries in cache - protected by cache->c_list_lock */ 41 void mb_cache_destroy(struct mb_cache *cache); 43 int mb_cache_entry_create(struct mb_cache *cache, gfp_t mask, u32 key, 45 void __mb_cache_entry_free(struct mb_cache *cache, 48 static inline void mb_cache_entry_put(struct mb_cache *cache, argument 58 __mb_cache_entry_free(cache, entry); 61 struct mb_cache_entry *mb_cache_entry_delete_or_get(struct mb_cache *cache, 63 struct mb_cache_entry *mb_cache_entry_get(struct mb_cache *cache, u32 key, 65 struct mb_cache_entry *mb_cache_entry_find_first(struct mb_cache *cache, [all...] |
/linux-master/arch/csky/mm/ |
H A D | syscache.c | 12 int, cache) 14 switch (cache) { 19 if (cache != BCACHE)
|
/linux-master/drivers/md/dm-vdo/indexer/ |
H A D | sparse-cache.h | 13 * The sparse cache is a cache of entire chapter indexes from sparse chapters used for searching 16 * the regular page cache in the volume. 18 * The most important property of this cache is the absence of synchronization for read operations. 19 * Safe concurrent access to the cache by the zone threads is controlled by the triage queue and 33 void uds_free_sparse_cache(struct sparse_cache *cache); 35 bool uds_sparse_cache_contains(struct sparse_cache *cache, u64 virtual_chapter, 40 void uds_invalidate_sparse_cache(struct sparse_cache *cache);
|
H A D | sparse-cache.c | 6 #include "sparse-cache.h" 8 #include <linux/cache.h> 21 * Since the cache is small, it is implemented as a simple array of cache entries. Searching for a 22 * specific virtual chapter is implemented as a linear search. The cache replacement policy is 23 * least-recently-used (LRU). Again, the small size of the cache allows the LRU order to be 26 * Changing the contents of the cache requires the coordinated participation of all zone threads 28 * thread. The critical invariant for coordination is that the cache membership must not change 32 * chapter because it has had too many cache misses" are represented separately from the cache 234 make_search_list(struct sparse_cache *cache, struct search_list **list_ptr) argument 263 struct sparse_cache *cache; local 323 score_search_miss(struct sparse_cache *cache, struct cached_chapter_index *chapter) argument 345 uds_free_sparse_cache(struct sparse_cache *cache) argument 389 uds_sparse_cache_contains(struct sparse_cache *cache, u64 virtual_chapter, unsigned int zone_number) argument 424 purge_search_list(struct search_list *search_list, struct sparse_cache *cache, u64 oldest_virtual_chapter) argument 496 struct sparse_cache *cache = index->volume->sparse_cache; local 537 uds_invalidate_sparse_cache(struct sparse_cache *cache) argument 580 struct sparse_cache *cache = volume->sparse_cache; local [all...] |
/linux-master/arch/openrisc/mm/ |
H A D | Makefile | 6 obj-y := fault.o cache.o tlb.o init.o ioremap.o
|
/linux-master/arch/sh/mm/ |
H A D | cache-debugfs.c | 2 * debugfs ops for the L1 cache 16 #include <asm/cache.h> 28 struct cache_info *cache; local 49 cache = ¤t_cpu_data.dcache; 52 cache = ¤t_cpu_data.icache; 55 waysize = cache->sets; 64 waysize <<= cache->entry_shift; 66 for (way = 0; way < cache->ways; way++) { 76 addr += cache->linesz, line++) { 83 /* U: Dirty, cache ta [all...] |