1#ifndef _ASM_POWERPC_PERCPU_H_ 2#define _ASM_POWERPC_PERCPU_H_ 3#ifdef __powerpc64__ 4#include <linux/compiler.h> 5 6/* 7 * Same as asm-generic/percpu.h, except that we store the per cpu offset 8 * in the paca. Based on the x86-64 implementation. 9 */ 10 11#ifdef CONFIG_SMP 12 13#include <asm/paca.h> 14 15#define __per_cpu_offset(cpu) (paca[cpu].data_offset) 16#define __my_cpu_offset() get_paca()->data_offset 17#define per_cpu_offset(x) (__per_cpu_offset(x)) 18 19/* Separate out the type, so (int[3], foo) works. */ 20#define DEFINE_PER_CPU(type, name) \ 21 __attribute__((__section__(".data.percpu"))) __typeof__(type) per_cpu__##name 22 23/* var is in discarded region: offset to particular copy we want */ 24#define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset(cpu))) 25#define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __my_cpu_offset())) 26#define __raw_get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __my_cpu_offset())) 27 28/* A macro to avoid #include hell... */ 29#define percpu_modcopy(pcpudst, src, size) \ 30do { \ 31 unsigned int __i; \ 32 for_each_possible_cpu(__i) \ 33 memcpy((pcpudst)+__per_cpu_offset(__i), \ 34 (src), (size)); \ 35} while (0) 36 37extern void setup_per_cpu_areas(void); 38 39#else /* ! SMP */ 40 41#define DEFINE_PER_CPU(type, name) \ 42 __typeof__(type) per_cpu__##name 43 44#define per_cpu(var, cpu) (*((void)(cpu), &per_cpu__##var)) 45#define __get_cpu_var(var) per_cpu__##var 46#define __raw_get_cpu_var(var) per_cpu__##var 47 48#endif /* SMP */ 49 50#define DECLARE_PER_CPU(type, name) extern __typeof__(type) per_cpu__##name 51 52#define EXPORT_PER_CPU_SYMBOL(var) EXPORT_SYMBOL(per_cpu__##var) 53#define EXPORT_PER_CPU_SYMBOL_GPL(var) EXPORT_SYMBOL_GPL(per_cpu__##var) 54 55#else 56#include <asm-generic/percpu.h> 57#endif 58 59#endif /* _ASM_POWERPC_PERCPU_H_ */ 60