1/* SPDX-License-Identifier: GPL-2.0 */ 2#ifndef _RESCTRL_H 3#define _RESCTRL_H 4 5#include <linux/kernel.h> 6#include <linux/list.h> 7#include <linux/pid.h> 8 9/* CLOSID, RMID value used by the default control group */ 10#define RESCTRL_RESERVED_CLOSID 0 11#define RESCTRL_RESERVED_RMID 0 12 13#define RESCTRL_PICK_ANY_CPU -1 14 15#ifdef CONFIG_PROC_CPU_RESCTRL 16 17int proc_resctrl_show(struct seq_file *m, 18 struct pid_namespace *ns, 19 struct pid *pid, 20 struct task_struct *tsk); 21 22#endif 23 24/* max value for struct rdt_domain's mbps_val */ 25#define MBA_MAX_MBPS U32_MAX 26 27/** 28 * enum resctrl_conf_type - The type of configuration. 29 * @CDP_NONE: No prioritisation, both code and data are controlled or monitored. 30 * @CDP_CODE: Configuration applies to instruction fetches. 31 * @CDP_DATA: Configuration applies to reads and writes. 32 */ 33enum resctrl_conf_type { 34 CDP_NONE, 35 CDP_CODE, 36 CDP_DATA, 37}; 38 39#define CDP_NUM_TYPES (CDP_DATA + 1) 40 41/* 42 * Event IDs, the values match those used to program IA32_QM_EVTSEL before 43 * reading IA32_QM_CTR on RDT systems. 44 */ 45enum resctrl_event_id { 46 QOS_L3_OCCUP_EVENT_ID = 0x01, 47 QOS_L3_MBM_TOTAL_EVENT_ID = 0x02, 48 QOS_L3_MBM_LOCAL_EVENT_ID = 0x03, 49}; 50 51/** 52 * struct resctrl_staged_config - parsed configuration to be applied 53 * @new_ctrl: new ctrl value to be loaded 54 * @have_new_ctrl: whether the user provided new_ctrl is valid 55 */ 56struct resctrl_staged_config { 57 u32 new_ctrl; 58 bool have_new_ctrl; 59}; 60 61/** 62 * struct rdt_domain - group of CPUs sharing a resctrl resource 63 * @list: all instances of this resource 64 * @id: unique id for this instance 65 * @cpu_mask: which CPUs share this resource 66 * @rmid_busy_llc: bitmap of which limbo RMIDs are above threshold 67 * @mbm_total: saved state for MBM total bandwidth 68 * @mbm_local: saved state for MBM local bandwidth 69 * @mbm_over: worker to periodically read MBM h/w counters 70 * @cqm_limbo: worker to periodically read CQM h/w counters 71 * @mbm_work_cpu: worker CPU for MBM h/w counters 72 * @cqm_work_cpu: worker CPU for CQM h/w counters 73 * @plr: pseudo-locked region (if any) associated with domain 74 * @staged_config: parsed configuration to be applied 75 * @mbps_val: When mba_sc is enabled, this holds the array of user 76 * specified control values for mba_sc in MBps, indexed 77 * by closid 78 */ 79struct rdt_domain { 80 struct list_head list; 81 int id; 82 struct cpumask cpu_mask; 83 unsigned long *rmid_busy_llc; 84 struct mbm_state *mbm_total; 85 struct mbm_state *mbm_local; 86 struct delayed_work mbm_over; 87 struct delayed_work cqm_limbo; 88 int mbm_work_cpu; 89 int cqm_work_cpu; 90 struct pseudo_lock_region *plr; 91 struct resctrl_staged_config staged_config[CDP_NUM_TYPES]; 92 u32 *mbps_val; 93}; 94 95/** 96 * struct resctrl_cache - Cache allocation related data 97 * @cbm_len: Length of the cache bit mask 98 * @min_cbm_bits: Minimum number of consecutive bits to be set. 99 * The value 0 means the architecture can support 100 * zero CBM. 101 * @shareable_bits: Bitmask of shareable resource with other 102 * executing entities 103 * @arch_has_sparse_bitmasks: True if a bitmask like f00f is valid. 104 * @arch_has_per_cpu_cfg: True if QOS_CFG register for this cache 105 * level has CPU scope. 106 */ 107struct resctrl_cache { 108 unsigned int cbm_len; 109 unsigned int min_cbm_bits; 110 unsigned int shareable_bits; 111 bool arch_has_sparse_bitmasks; 112 bool arch_has_per_cpu_cfg; 113}; 114 115/** 116 * enum membw_throttle_mode - System's memory bandwidth throttling mode 117 * @THREAD_THROTTLE_UNDEFINED: Not relevant to the system 118 * @THREAD_THROTTLE_MAX: Memory bandwidth is throttled at the core 119 * always using smallest bandwidth percentage 120 * assigned to threads, aka "max throttling" 121 * @THREAD_THROTTLE_PER_THREAD: Memory bandwidth is throttled at the thread 122 */ 123enum membw_throttle_mode { 124 THREAD_THROTTLE_UNDEFINED = 0, 125 THREAD_THROTTLE_MAX, 126 THREAD_THROTTLE_PER_THREAD, 127}; 128 129/** 130 * struct resctrl_membw - Memory bandwidth allocation related data 131 * @min_bw: Minimum memory bandwidth percentage user can request 132 * @bw_gran: Granularity at which the memory bandwidth is allocated 133 * @delay_linear: True if memory B/W delay is in linear scale 134 * @arch_needs_linear: True if we can't configure non-linear resources 135 * @throttle_mode: Bandwidth throttling mode when threads request 136 * different memory bandwidths 137 * @mba_sc: True if MBA software controller(mba_sc) is enabled 138 * @mb_map: Mapping of memory B/W percentage to memory B/W delay 139 */ 140struct resctrl_membw { 141 u32 min_bw; 142 u32 bw_gran; 143 u32 delay_linear; 144 bool arch_needs_linear; 145 enum membw_throttle_mode throttle_mode; 146 bool mba_sc; 147 u32 *mb_map; 148}; 149 150struct rdt_parse_data; 151struct resctrl_schema; 152 153/** 154 * struct rdt_resource - attributes of a resctrl resource 155 * @rid: The index of the resource 156 * @alloc_capable: Is allocation available on this machine 157 * @mon_capable: Is monitor feature available on this machine 158 * @num_rmid: Number of RMIDs available 159 * @cache_level: Which cache level defines scope of this resource 160 * @cache: Cache allocation related data 161 * @membw: If the component has bandwidth controls, their properties. 162 * @domains: RCU list of all domains for this resource 163 * @name: Name to use in "schemata" file. 164 * @data_width: Character width of data when displaying 165 * @default_ctrl: Specifies default cache cbm or memory B/W percent. 166 * @format_str: Per resource format string to show domain value 167 * @parse_ctrlval: Per resource function pointer to parse control values 168 * @evt_list: List of monitoring events 169 * @fflags: flags to choose base and info files 170 * @cdp_capable: Is the CDP feature available on this resource 171 */ 172struct rdt_resource { 173 int rid; 174 bool alloc_capable; 175 bool mon_capable; 176 int num_rmid; 177 int cache_level; 178 struct resctrl_cache cache; 179 struct resctrl_membw membw; 180 struct list_head domains; 181 char *name; 182 int data_width; 183 u32 default_ctrl; 184 const char *format_str; 185 int (*parse_ctrlval)(struct rdt_parse_data *data, 186 struct resctrl_schema *s, 187 struct rdt_domain *d); 188 struct list_head evt_list; 189 unsigned long fflags; 190 bool cdp_capable; 191}; 192 193/** 194 * struct resctrl_schema - configuration abilities of a resource presented to 195 * user-space 196 * @list: Member of resctrl_schema_all. 197 * @name: The name to use in the "schemata" file. 198 * @conf_type: Whether this schema is specific to code/data. 199 * @res: The resource structure exported by the architecture to describe 200 * the hardware that is configured by this schema. 201 * @num_closid: The number of closid that can be used with this schema. When 202 * features like CDP are enabled, this will be lower than the 203 * hardware supports for the resource. 204 */ 205struct resctrl_schema { 206 struct list_head list; 207 char name[8]; 208 enum resctrl_conf_type conf_type; 209 struct rdt_resource *res; 210 u32 num_closid; 211}; 212 213/* The number of closid supported by this resource regardless of CDP */ 214u32 resctrl_arch_get_num_closid(struct rdt_resource *r); 215int resctrl_arch_update_domains(struct rdt_resource *r, u32 closid); 216 217/* 218 * Update the ctrl_val and apply this config right now. 219 * Must be called on one of the domain's CPUs. 220 */ 221int resctrl_arch_update_one(struct rdt_resource *r, struct rdt_domain *d, 222 u32 closid, enum resctrl_conf_type t, u32 cfg_val); 223 224u32 resctrl_arch_get_config(struct rdt_resource *r, struct rdt_domain *d, 225 u32 closid, enum resctrl_conf_type type); 226int resctrl_online_domain(struct rdt_resource *r, struct rdt_domain *d); 227void resctrl_offline_domain(struct rdt_resource *r, struct rdt_domain *d); 228void resctrl_online_cpu(unsigned int cpu); 229void resctrl_offline_cpu(unsigned int cpu); 230 231/** 232 * resctrl_arch_rmid_read() - Read the eventid counter corresponding to rmid 233 * for this resource and domain. 234 * @r: resource that the counter should be read from. 235 * @d: domain that the counter should be read from. 236 * @closid: closid that matches the rmid. Depending on the architecture, the 237 * counter may match traffic of both @closid and @rmid, or @rmid 238 * only. 239 * @rmid: rmid of the counter to read. 240 * @eventid: eventid to read, e.g. L3 occupancy. 241 * @val: result of the counter read in bytes. 242 * @arch_mon_ctx: An architecture specific value from 243 * resctrl_arch_mon_ctx_alloc(), for MPAM this identifies 244 * the hardware monitor allocated for this read request. 245 * 246 * Some architectures need to sleep when first programming some of the counters. 247 * (specifically: arm64's MPAM cache occupancy counters can return 'not ready' 248 * for a short period of time). Call from a non-migrateable process context on 249 * a CPU that belongs to domain @d. e.g. use smp_call_on_cpu() or 250 * schedule_work_on(). This function can be called with interrupts masked, 251 * e.g. using smp_call_function_any(), but may consistently return an error. 252 * 253 * Return: 254 * 0 on success, or -EIO, -EINVAL etc on error. 255 */ 256int resctrl_arch_rmid_read(struct rdt_resource *r, struct rdt_domain *d, 257 u32 closid, u32 rmid, enum resctrl_event_id eventid, 258 u64 *val, void *arch_mon_ctx); 259 260/** 261 * resctrl_arch_rmid_read_context_check() - warn about invalid contexts 262 * 263 * When built with CONFIG_DEBUG_ATOMIC_SLEEP generate a warning when 264 * resctrl_arch_rmid_read() is called with preemption disabled. 265 * 266 * The contract with resctrl_arch_rmid_read() is that if interrupts 267 * are unmasked, it can sleep. This allows NOHZ_FULL systems to use an 268 * IPI, (and fail if the call needed to sleep), while most of the time 269 * the work is scheduled, allowing the call to sleep. 270 */ 271static inline void resctrl_arch_rmid_read_context_check(void) 272{ 273 if (!irqs_disabled()) 274 might_sleep(); 275} 276 277/** 278 * resctrl_arch_reset_rmid() - Reset any private state associated with rmid 279 * and eventid. 280 * @r: The domain's resource. 281 * @d: The rmid's domain. 282 * @closid: closid that matches the rmid. Depending on the architecture, the 283 * counter may match traffic of both @closid and @rmid, or @rmid only. 284 * @rmid: The rmid whose counter values should be reset. 285 * @eventid: The eventid whose counter values should be reset. 286 * 287 * This can be called from any CPU. 288 */ 289void resctrl_arch_reset_rmid(struct rdt_resource *r, struct rdt_domain *d, 290 u32 closid, u32 rmid, 291 enum resctrl_event_id eventid); 292 293/** 294 * resctrl_arch_reset_rmid_all() - Reset all private state associated with 295 * all rmids and eventids. 296 * @r: The resctrl resource. 297 * @d: The domain for which all architectural counter state will 298 * be cleared. 299 * 300 * This can be called from any CPU. 301 */ 302void resctrl_arch_reset_rmid_all(struct rdt_resource *r, struct rdt_domain *d); 303 304extern unsigned int resctrl_rmid_realloc_threshold; 305extern unsigned int resctrl_rmid_realloc_limit; 306 307#endif /* _RESCTRL_H */ 308