1/*- 2 * Copyright (c) Peter Wemm 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 *
| 1/*- 2 * Copyright (c) Peter Wemm 3 * All rights reserved. 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 *
|
26 * $FreeBSD: stable/11/sys/i386/include/pcpu.h 334152 2018-05-24 13:17:24Z kib $
| 26 * $FreeBSD: stable/11/sys/i386/include/pcpu.h 347568 2019-05-14 17:05:02Z kib $
|
27 */ 28 29#ifndef _MACHINE_PCPU_H_ 30#define _MACHINE_PCPU_H_ 31 32#ifndef _SYS_CDEFS_H_ 33#error "sys/cdefs.h is a prerequisite for this file" 34#endif 35 36#include <machine/segments.h> 37#include <machine/tss.h> 38 39#include <sys/_lock.h> 40#include <sys/_mutex.h> 41 42/* 43 * The SMP parts are setup in pmap.c and locore.s for the BSP, and 44 * mp_machdep.c sets up the data for the AP's to "see" when they awake. 45 * The reason for doing it via a struct is so that an array of pointers 46 * to each CPU's data can be set up for things like "check curproc on all 47 * other processors" 48 */ 49 50#define PCPU_MD_FIELDS \ 51 char pc_monitorbuf[128] __aligned(128); /* cache line */ \ 52 struct pcpu *pc_prvspace; /* Self-reference */ \ 53 struct pmap *pc_curpmap; \ 54 struct i386tss pc_common_tss; \ 55 struct segment_descriptor pc_common_tssd; \ 56 struct segment_descriptor *pc_tss_gdt; \ 57 struct segment_descriptor *pc_fsgs_gdt; \ 58 int pc_currentldt; \ 59 u_int pc_acpi_id; /* ACPI CPU id */ \ 60 u_int pc_apic_id; \ 61 int pc_private_tss; /* Flag indicating private tss*/\ 62 u_int pc_cmci_mask; /* MCx banks for CMCI */ \ 63 u_int pc_vcpu_id; /* Xen vCPU ID */ \ 64 struct mtx pc_cmap_lock; \ 65 void *pc_cmap_pte1; \ 66 void *pc_cmap_pte2; \ 67 caddr_t pc_cmap_addr1; \ 68 caddr_t pc_cmap_addr2; \ 69 vm_offset_t pc_qmap_addr; /* KVA for temporary mappings */\ 70 uint32_t pc_smp_tlb_done; /* TLB op acknowledgement */ \ 71 uint32_t pc_ibpb_set; \
| 27 */ 28 29#ifndef _MACHINE_PCPU_H_ 30#define _MACHINE_PCPU_H_ 31 32#ifndef _SYS_CDEFS_H_ 33#error "sys/cdefs.h is a prerequisite for this file" 34#endif 35 36#include <machine/segments.h> 37#include <machine/tss.h> 38 39#include <sys/_lock.h> 40#include <sys/_mutex.h> 41 42/* 43 * The SMP parts are setup in pmap.c and locore.s for the BSP, and 44 * mp_machdep.c sets up the data for the AP's to "see" when they awake. 45 * The reason for doing it via a struct is so that an array of pointers 46 * to each CPU's data can be set up for things like "check curproc on all 47 * other processors" 48 */ 49 50#define PCPU_MD_FIELDS \ 51 char pc_monitorbuf[128] __aligned(128); /* cache line */ \ 52 struct pcpu *pc_prvspace; /* Self-reference */ \ 53 struct pmap *pc_curpmap; \ 54 struct i386tss pc_common_tss; \ 55 struct segment_descriptor pc_common_tssd; \ 56 struct segment_descriptor *pc_tss_gdt; \ 57 struct segment_descriptor *pc_fsgs_gdt; \ 58 int pc_currentldt; \ 59 u_int pc_acpi_id; /* ACPI CPU id */ \ 60 u_int pc_apic_id; \ 61 int pc_private_tss; /* Flag indicating private tss*/\ 62 u_int pc_cmci_mask; /* MCx banks for CMCI */ \ 63 u_int pc_vcpu_id; /* Xen vCPU ID */ \ 64 struct mtx pc_cmap_lock; \ 65 void *pc_cmap_pte1; \ 66 void *pc_cmap_pte2; \ 67 caddr_t pc_cmap_addr1; \ 68 caddr_t pc_cmap_addr2; \ 69 vm_offset_t pc_qmap_addr; /* KVA for temporary mappings */\ 70 uint32_t pc_smp_tlb_done; /* TLB op acknowledgement */ \ 71 uint32_t pc_ibpb_set; \
|
72 char __pad[185]
| 72 void *pc_mds_buf; \ 73 void *pc_mds_buf64; \ 74 uint32_t pc_pad[12]; \ 75 uint8_t pc_mds_tmp[64]; \ 76 char __pad[153]
|
73 74#ifdef _KERNEL 75 76#ifdef lint 77 78extern struct pcpu *pcpup; 79 80#define get_pcpu() (pcpup) 81#define PCPU_GET(member) (pcpup->pc_ ## member) 82#define PCPU_ADD(member, val) (pcpup->pc_ ## member += (val)) 83#define PCPU_INC(member) PCPU_ADD(member, 1) 84#define PCPU_PTR(member) (&pcpup->pc_ ## member) 85#define PCPU_SET(member, val) (pcpup->pc_ ## member = (val)) 86 87#elif defined(__GNUCLIKE_ASM) && defined(__GNUCLIKE___TYPEOF) 88 89/* 90 * Evaluates to the byte offset of the per-cpu variable name. 91 */ 92#define __pcpu_offset(name) \ 93 __offsetof(struct pcpu, name) 94 95/* 96 * Evaluates to the type of the per-cpu variable name. 97 */ 98#define __pcpu_type(name) \ 99 __typeof(((struct pcpu *)0)->name) 100 101/* 102 * Evaluates to the address of the per-cpu variable name. 103 */ 104#define __PCPU_PTR(name) __extension__ ({ \ 105 __pcpu_type(name) *__p; \ 106 \ 107 __asm __volatile("movl %%fs:%1,%0; addl %2,%0" \ 108 : "=r" (__p) \ 109 : "m" (*(struct pcpu *)(__pcpu_offset(pc_prvspace))), \ 110 "i" (__pcpu_offset(name))); \ 111 \ 112 __p; \ 113}) 114 115/* 116 * Evaluates to the value of the per-cpu variable name. 117 */ 118#define __PCPU_GET(name) __extension__ ({ \ 119 __pcpu_type(name) __res; \ 120 struct __s { \ 121 u_char __b[MIN(sizeof(__res), 4)]; \ 122 } __s; \ 123 \ 124 if (sizeof(__res) == 1 || sizeof(__res) == 2 || \ 125 sizeof(__res) == 4) { \ 126 __asm __volatile("mov %%fs:%1,%0" \ 127 : "=r" (__s) \ 128 : "m" (*(struct __s *)(__pcpu_offset(name)))); \ 129 *(struct __s *)(void *)&__res = __s; \ 130 } else { \ 131 __res = *__PCPU_PTR(name); \ 132 } \ 133 __res; \ 134}) 135 136/* 137 * Adds a value of the per-cpu counter name. The implementation 138 * must be atomic with respect to interrupts. 139 */ 140#define __PCPU_ADD(name, val) do { \ 141 __pcpu_type(name) __val; \ 142 struct __s { \ 143 u_char __b[MIN(sizeof(__val), 4)]; \ 144 } __s; \ 145 \ 146 __val = (val); \ 147 if (sizeof(__val) == 1 || sizeof(__val) == 2 || \ 148 sizeof(__val) == 4) { \ 149 __s = *(struct __s *)(void *)&__val; \ 150 __asm __volatile("add %1,%%fs:%0" \ 151 : "=m" (*(struct __s *)(__pcpu_offset(name))) \ 152 : "r" (__s)); \ 153 } else \ 154 *__PCPU_PTR(name) += __val; \ 155} while (0) 156 157/* 158 * Increments the value of the per-cpu counter name. The implementation 159 * must be atomic with respect to interrupts. 160 */ 161#define __PCPU_INC(name) do { \ 162 CTASSERT(sizeof(__pcpu_type(name)) == 1 || \ 163 sizeof(__pcpu_type(name)) == 2 || \ 164 sizeof(__pcpu_type(name)) == 4); \ 165 if (sizeof(__pcpu_type(name)) == 1) { \ 166 __asm __volatile("incb %%fs:%0" \ 167 : "=m" (*(__pcpu_type(name) *)(__pcpu_offset(name)))\ 168 : "m" (*(__pcpu_type(name) *)(__pcpu_offset(name))));\ 169 } else if (sizeof(__pcpu_type(name)) == 2) { \ 170 __asm __volatile("incw %%fs:%0" \ 171 : "=m" (*(__pcpu_type(name) *)(__pcpu_offset(name)))\ 172 : "m" (*(__pcpu_type(name) *)(__pcpu_offset(name))));\ 173 } else if (sizeof(__pcpu_type(name)) == 4) { \ 174 __asm __volatile("incl %%fs:%0" \ 175 : "=m" (*(__pcpu_type(name) *)(__pcpu_offset(name)))\ 176 : "m" (*(__pcpu_type(name) *)(__pcpu_offset(name))));\ 177 } \ 178} while (0) 179 180/* 181 * Sets the value of the per-cpu variable name to value val. 182 */ 183#define __PCPU_SET(name, val) do { \ 184 __pcpu_type(name) __val; \ 185 struct __s { \ 186 u_char __b[MIN(sizeof(__val), 4)]; \ 187 } __s; \ 188 \ 189 __val = (val); \ 190 if (sizeof(__val) == 1 || sizeof(__val) == 2 || \ 191 sizeof(__val) == 4) { \ 192 __s = *(struct __s *)(void *)&__val; \ 193 __asm __volatile("mov %1,%%fs:%0" \ 194 : "=m" (*(struct __s *)(__pcpu_offset(name))) \ 195 : "r" (__s)); \ 196 } else { \ 197 *__PCPU_PTR(name) = __val; \ 198 } \ 199} while (0) 200 201#define get_pcpu() __extension__ ({ \ 202 struct pcpu *__pc; \ 203 \ 204 __asm __volatile("movl %%fs:%1,%0" \ 205 : "=r" (__pc) \ 206 : "m" (*(struct pcpu *)(__pcpu_offset(pc_prvspace)))); \ 207 __pc; \ 208}) 209 210#define PCPU_GET(member) __PCPU_GET(pc_ ## member) 211#define PCPU_ADD(member, val) __PCPU_ADD(pc_ ## member, val) 212#define PCPU_INC(member) __PCPU_INC(pc_ ## member) 213#define PCPU_PTR(member) __PCPU_PTR(pc_ ## member) 214#define PCPU_SET(member, val) __PCPU_SET(pc_ ## member, val) 215 216#define OFFSETOF_CURTHREAD 0 217#ifdef __clang__ 218#pragma clang diagnostic push 219#pragma clang diagnostic ignored "-Wnull-dereference" 220#endif 221static __inline __pure2 struct thread * 222__curthread(void) 223{ 224 struct thread *td; 225 226 __asm("movl %%fs:%1,%0" : "=r" (td) 227 : "m" (*(char *)OFFSETOF_CURTHREAD)); 228 return (td); 229} 230#ifdef __clang__ 231#pragma clang diagnostic pop 232#endif 233#define curthread (__curthread()) 234 235#define OFFSETOF_CURPCB 16 236static __inline __pure2 struct pcb * 237__curpcb(void) 238{ 239 struct pcb *pcb; 240 241 __asm("movl %%fs:%1,%0" : "=r" (pcb) : "m" (*(char *)OFFSETOF_CURPCB)); 242 return (pcb); 243} 244#define curpcb (__curpcb()) 245 246#define IS_BSP() (PCPU_GET(cpuid) == 0) 247 248#else /* !lint || defined(__GNUCLIKE_ASM) && defined(__GNUCLIKE___TYPEOF) */ 249 250#error "this file needs to be ported to your compiler" 251 252#endif /* lint, etc. */ 253 254#endif /* _KERNEL */ 255 256#endif /* !_MACHINE_PCPU_H_ */
| 77 78#ifdef _KERNEL 79 80#ifdef lint 81 82extern struct pcpu *pcpup; 83 84#define get_pcpu() (pcpup) 85#define PCPU_GET(member) (pcpup->pc_ ## member) 86#define PCPU_ADD(member, val) (pcpup->pc_ ## member += (val)) 87#define PCPU_INC(member) PCPU_ADD(member, 1) 88#define PCPU_PTR(member) (&pcpup->pc_ ## member) 89#define PCPU_SET(member, val) (pcpup->pc_ ## member = (val)) 90 91#elif defined(__GNUCLIKE_ASM) && defined(__GNUCLIKE___TYPEOF) 92 93/* 94 * Evaluates to the byte offset of the per-cpu variable name. 95 */ 96#define __pcpu_offset(name) \ 97 __offsetof(struct pcpu, name) 98 99/* 100 * Evaluates to the type of the per-cpu variable name. 101 */ 102#define __pcpu_type(name) \ 103 __typeof(((struct pcpu *)0)->name) 104 105/* 106 * Evaluates to the address of the per-cpu variable name. 107 */ 108#define __PCPU_PTR(name) __extension__ ({ \ 109 __pcpu_type(name) *__p; \ 110 \ 111 __asm __volatile("movl %%fs:%1,%0; addl %2,%0" \ 112 : "=r" (__p) \ 113 : "m" (*(struct pcpu *)(__pcpu_offset(pc_prvspace))), \ 114 "i" (__pcpu_offset(name))); \ 115 \ 116 __p; \ 117}) 118 119/* 120 * Evaluates to the value of the per-cpu variable name. 121 */ 122#define __PCPU_GET(name) __extension__ ({ \ 123 __pcpu_type(name) __res; \ 124 struct __s { \ 125 u_char __b[MIN(sizeof(__res), 4)]; \ 126 } __s; \ 127 \ 128 if (sizeof(__res) == 1 || sizeof(__res) == 2 || \ 129 sizeof(__res) == 4) { \ 130 __asm __volatile("mov %%fs:%1,%0" \ 131 : "=r" (__s) \ 132 : "m" (*(struct __s *)(__pcpu_offset(name)))); \ 133 *(struct __s *)(void *)&__res = __s; \ 134 } else { \ 135 __res = *__PCPU_PTR(name); \ 136 } \ 137 __res; \ 138}) 139 140/* 141 * Adds a value of the per-cpu counter name. The implementation 142 * must be atomic with respect to interrupts. 143 */ 144#define __PCPU_ADD(name, val) do { \ 145 __pcpu_type(name) __val; \ 146 struct __s { \ 147 u_char __b[MIN(sizeof(__val), 4)]; \ 148 } __s; \ 149 \ 150 __val = (val); \ 151 if (sizeof(__val) == 1 || sizeof(__val) == 2 || \ 152 sizeof(__val) == 4) { \ 153 __s = *(struct __s *)(void *)&__val; \ 154 __asm __volatile("add %1,%%fs:%0" \ 155 : "=m" (*(struct __s *)(__pcpu_offset(name))) \ 156 : "r" (__s)); \ 157 } else \ 158 *__PCPU_PTR(name) += __val; \ 159} while (0) 160 161/* 162 * Increments the value of the per-cpu counter name. The implementation 163 * must be atomic with respect to interrupts. 164 */ 165#define __PCPU_INC(name) do { \ 166 CTASSERT(sizeof(__pcpu_type(name)) == 1 || \ 167 sizeof(__pcpu_type(name)) == 2 || \ 168 sizeof(__pcpu_type(name)) == 4); \ 169 if (sizeof(__pcpu_type(name)) == 1) { \ 170 __asm __volatile("incb %%fs:%0" \ 171 : "=m" (*(__pcpu_type(name) *)(__pcpu_offset(name)))\ 172 : "m" (*(__pcpu_type(name) *)(__pcpu_offset(name))));\ 173 } else if (sizeof(__pcpu_type(name)) == 2) { \ 174 __asm __volatile("incw %%fs:%0" \ 175 : "=m" (*(__pcpu_type(name) *)(__pcpu_offset(name)))\ 176 : "m" (*(__pcpu_type(name) *)(__pcpu_offset(name))));\ 177 } else if (sizeof(__pcpu_type(name)) == 4) { \ 178 __asm __volatile("incl %%fs:%0" \ 179 : "=m" (*(__pcpu_type(name) *)(__pcpu_offset(name)))\ 180 : "m" (*(__pcpu_type(name) *)(__pcpu_offset(name))));\ 181 } \ 182} while (0) 183 184/* 185 * Sets the value of the per-cpu variable name to value val. 186 */ 187#define __PCPU_SET(name, val) do { \ 188 __pcpu_type(name) __val; \ 189 struct __s { \ 190 u_char __b[MIN(sizeof(__val), 4)]; \ 191 } __s; \ 192 \ 193 __val = (val); \ 194 if (sizeof(__val) == 1 || sizeof(__val) == 2 || \ 195 sizeof(__val) == 4) { \ 196 __s = *(struct __s *)(void *)&__val; \ 197 __asm __volatile("mov %1,%%fs:%0" \ 198 : "=m" (*(struct __s *)(__pcpu_offset(name))) \ 199 : "r" (__s)); \ 200 } else { \ 201 *__PCPU_PTR(name) = __val; \ 202 } \ 203} while (0) 204 205#define get_pcpu() __extension__ ({ \ 206 struct pcpu *__pc; \ 207 \ 208 __asm __volatile("movl %%fs:%1,%0" \ 209 : "=r" (__pc) \ 210 : "m" (*(struct pcpu *)(__pcpu_offset(pc_prvspace)))); \ 211 __pc; \ 212}) 213 214#define PCPU_GET(member) __PCPU_GET(pc_ ## member) 215#define PCPU_ADD(member, val) __PCPU_ADD(pc_ ## member, val) 216#define PCPU_INC(member) __PCPU_INC(pc_ ## member) 217#define PCPU_PTR(member) __PCPU_PTR(pc_ ## member) 218#define PCPU_SET(member, val) __PCPU_SET(pc_ ## member, val) 219 220#define OFFSETOF_CURTHREAD 0 221#ifdef __clang__ 222#pragma clang diagnostic push 223#pragma clang diagnostic ignored "-Wnull-dereference" 224#endif 225static __inline __pure2 struct thread * 226__curthread(void) 227{ 228 struct thread *td; 229 230 __asm("movl %%fs:%1,%0" : "=r" (td) 231 : "m" (*(char *)OFFSETOF_CURTHREAD)); 232 return (td); 233} 234#ifdef __clang__ 235#pragma clang diagnostic pop 236#endif 237#define curthread (__curthread()) 238 239#define OFFSETOF_CURPCB 16 240static __inline __pure2 struct pcb * 241__curpcb(void) 242{ 243 struct pcb *pcb; 244 245 __asm("movl %%fs:%1,%0" : "=r" (pcb) : "m" (*(char *)OFFSETOF_CURPCB)); 246 return (pcb); 247} 248#define curpcb (__curpcb()) 249 250#define IS_BSP() (PCPU_GET(cpuid) == 0) 251 252#else /* !lint || defined(__GNUCLIKE_ASM) && defined(__GNUCLIKE___TYPEOF) */ 253 254#error "this file needs to be ported to your compiler" 255 256#endif /* lint, etc. */ 257 258#endif /* _KERNEL */ 259 260#endif /* !_MACHINE_PCPU_H_ */
|