1/* SPDX-License-Identifier: GPL-2.0 */ 2#ifndef __LINUX_CACHE_H 3#define __LINUX_CACHE_H 4 5#include <uapi/linux/kernel.h> 6#include <asm/cache.h> 7 8#ifndef L1_CACHE_ALIGN 9#define L1_CACHE_ALIGN(x) __ALIGN_KERNEL(x, L1_CACHE_BYTES) 10#endif 11 12#ifndef SMP_CACHE_BYTES 13#define SMP_CACHE_BYTES L1_CACHE_BYTES 14#endif 15 16/* 17 * __read_mostly is used to keep rarely changing variables out of frequently 18 * updated cachelines. Its use should be reserved for data that is used 19 * frequently in hot paths. Performance traces can help decide when to use 20 * this. You want __read_mostly data to be tightly packed, so that in the 21 * best case multiple frequently read variables for a hot path will be next 22 * to each other in order to reduce the number of cachelines needed to 23 * execute a critical path. We should be mindful and selective of its use. 24 * ie: if you're going to use it please supply a *good* justification in your 25 * commit log 26 */ 27#ifndef __read_mostly 28#define __read_mostly 29#endif 30 31/* 32 * __ro_after_init is used to mark things that are read-only after init (i.e. 33 * after mark_rodata_ro() has been called). These are effectively read-only, 34 * but may get written to during init, so can't live in .rodata (via "const"). 35 */ 36#ifndef __ro_after_init 37#define __ro_after_init __section(".data..ro_after_init") 38#endif 39 40#ifndef ____cacheline_aligned 41#define ____cacheline_aligned __attribute__((__aligned__(SMP_CACHE_BYTES))) 42#endif 43 44#ifndef ____cacheline_aligned_in_smp 45#ifdef CONFIG_SMP 46#define ____cacheline_aligned_in_smp ____cacheline_aligned 47#else 48#define ____cacheline_aligned_in_smp 49#endif /* CONFIG_SMP */ 50#endif 51 52#ifndef __cacheline_aligned 53#define __cacheline_aligned \ 54 __attribute__((__aligned__(SMP_CACHE_BYTES), \ 55 __section__(".data..cacheline_aligned"))) 56#endif /* __cacheline_aligned */ 57 58#ifndef __cacheline_aligned_in_smp 59#ifdef CONFIG_SMP 60#define __cacheline_aligned_in_smp __cacheline_aligned 61#else 62#define __cacheline_aligned_in_smp 63#endif /* CONFIG_SMP */ 64#endif 65 66/* 67 * The maximum alignment needed for some critical structures 68 * These could be inter-node cacheline sizes/L3 cacheline 69 * size etc. Define this in asm/cache.h for your arch 70 */ 71#ifndef INTERNODE_CACHE_SHIFT 72#define INTERNODE_CACHE_SHIFT L1_CACHE_SHIFT 73#endif 74 75#if !defined(____cacheline_internodealigned_in_smp) 76#if defined(CONFIG_SMP) 77#define ____cacheline_internodealigned_in_smp \ 78 __attribute__((__aligned__(1 << (INTERNODE_CACHE_SHIFT)))) 79#else 80#define ____cacheline_internodealigned_in_smp 81#endif 82#endif 83 84#ifndef CONFIG_ARCH_HAS_CACHE_LINE_SIZE 85#define cache_line_size() L1_CACHE_BYTES 86#endif 87 88#ifndef __cacheline_group_begin 89#define __cacheline_group_begin(GROUP) \ 90 __u8 __cacheline_group_begin__##GROUP[0] 91#endif 92 93#ifndef __cacheline_group_end 94#define __cacheline_group_end(GROUP) \ 95 __u8 __cacheline_group_end__##GROUP[0] 96#endif 97 98#ifndef CACHELINE_ASSERT_GROUP_MEMBER 99#define CACHELINE_ASSERT_GROUP_MEMBER(TYPE, GROUP, MEMBER) \ 100 BUILD_BUG_ON(!(offsetof(TYPE, MEMBER) >= \ 101 offsetofend(TYPE, __cacheline_group_begin__##GROUP) && \ 102 offsetofend(TYPE, MEMBER) <= \ 103 offsetof(TYPE, __cacheline_group_end__##GROUP))) 104#endif 105 106#ifndef CACHELINE_ASSERT_GROUP_SIZE 107#define CACHELINE_ASSERT_GROUP_SIZE(TYPE, GROUP, SIZE) \ 108 BUILD_BUG_ON(offsetof(TYPE, __cacheline_group_end__##GROUP) - \ 109 offsetofend(TYPE, __cacheline_group_begin__##GROUP) > \ 110 SIZE) 111#endif 112 113/* 114 * Helper to add padding within a struct to ensure data fall into separate 115 * cachelines. 116 */ 117#if defined(CONFIG_SMP) 118struct cacheline_padding { 119 char x[0]; 120} ____cacheline_internodealigned_in_smp; 121#define CACHELINE_PADDING(name) struct cacheline_padding name 122#else 123#define CACHELINE_PADDING(name) 124#endif 125 126#ifdef ARCH_DMA_MINALIGN 127#define ARCH_HAS_DMA_MINALIGN 128#else 129#define ARCH_DMA_MINALIGN __alignof__(unsigned long long) 130#endif 131 132#endif /* __LINUX_CACHE_H */ 133