1/* SPDX-License-Identifier: GPL-2.0 */ 2#ifndef _LINUX_SCHED_NUMA_BALANCING_H 3#define _LINUX_SCHED_NUMA_BALANCING_H 4 5/* 6 * This is the interface between the scheduler and the MM that 7 * implements memory access pattern based NUMA-balancing: 8 */ 9 10#include <linux/sched.h> 11 12#define TNF_MIGRATED 0x01 13#define TNF_NO_GROUP 0x02 14#define TNF_SHARED 0x04 15#define TNF_FAULT_LOCAL 0x08 16#define TNF_MIGRATE_FAIL 0x10 17 18enum numa_vmaskip_reason { 19 NUMAB_SKIP_UNSUITABLE, 20 NUMAB_SKIP_SHARED_RO, 21 NUMAB_SKIP_INACCESSIBLE, 22 NUMAB_SKIP_SCAN_DELAY, 23 NUMAB_SKIP_PID_INACTIVE, 24 NUMAB_SKIP_IGNORE_PID, 25 NUMAB_SKIP_SEQ_COMPLETED, 26}; 27 28#ifdef CONFIG_NUMA_BALANCING 29extern void task_numa_fault(int last_node, int node, int pages, int flags); 30extern pid_t task_numa_group_id(struct task_struct *p); 31extern void set_numabalancing_state(bool enabled); 32extern void task_numa_free(struct task_struct *p, bool final); 33bool should_numa_migrate_memory(struct task_struct *p, struct folio *folio, 34 int src_nid, int dst_cpu); 35#else 36static inline void task_numa_fault(int last_node, int node, int pages, 37 int flags) 38{ 39} 40static inline pid_t task_numa_group_id(struct task_struct *p) 41{ 42 return 0; 43} 44static inline void set_numabalancing_state(bool enabled) 45{ 46} 47static inline void task_numa_free(struct task_struct *p, bool final) 48{ 49} 50static inline bool should_numa_migrate_memory(struct task_struct *p, 51 struct folio *folio, int src_nid, int dst_cpu) 52{ 53 return true; 54} 55#endif 56 57#endif /* _LINUX_SCHED_NUMA_BALANCING_H */ 58