1/******************************************************************************
2 * sysctl.h
3 *
4 * System management operations. For use by node control stack.
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to
8 * deal in the Software without restriction, including without limitation the
9 * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
10 * sell copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
19 * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
21 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
22 * DEALINGS IN THE SOFTWARE.
23 *
24 * Copyright (c) 2002-2006, K Fraser
25 */
26
27#ifndef __XEN_PUBLIC_SYSCTL_H__
28#define __XEN_PUBLIC_SYSCTL_H__
29
30#if !defined(__XEN__) && !defined(__XEN_TOOLS__)
31#error "sysctl operations are intended for use by node control tools only"
32#endif
33
34#include "xen.h"
35#include "domctl.h"
36#include "physdev.h"
37
38#define XEN_SYSCTL_INTERFACE_VERSION 0x00000014
39
40/*
41 * Read console content from Xen buffer ring.
42 */
43/* XEN_SYSCTL_readconsole */
44struct xen_sysctl_readconsole {
45    /* IN: Non-zero -> clear after reading. */
46    uint8_t clear;
47    /* IN: Non-zero -> start index specified by @index field. */
48    uint8_t incremental;
49    uint8_t pad0, pad1;
50    /*
51     * IN:  Start index for consuming from ring buffer (if @incremental);
52     * OUT: End index after consuming from ring buffer.
53     */
54    uint32_t index;
55    /* IN: Virtual address to write console data. */
56    XEN_GUEST_HANDLE_64(char) buffer;
57    /* IN: Size of buffer; OUT: Bytes written to buffer. */
58    uint32_t count;
59};
60
61/* Get trace buffers machine base address */
62/* XEN_SYSCTL_tbuf_op */
63struct xen_sysctl_tbuf_op {
64    /* IN variables */
65#define XEN_SYSCTL_TBUFOP_get_info     0
66#define XEN_SYSCTL_TBUFOP_set_cpu_mask 1
67#define XEN_SYSCTL_TBUFOP_set_evt_mask 2
68#define XEN_SYSCTL_TBUFOP_set_size     3
69#define XEN_SYSCTL_TBUFOP_enable       4
70#define XEN_SYSCTL_TBUFOP_disable      5
71    uint32_t cmd;
72    /* IN/OUT variables */
73    struct xenctl_bitmap cpu_mask;
74    uint32_t             evt_mask;
75    /* OUT variables */
76    uint64_aligned_t buffer_mfn;
77    uint32_t size;  /* Also an IN variable! */
78};
79
80/*
81 * Get physical information about the host machine
82 */
83/* XEN_SYSCTL_physinfo */
84 /* The platform supports HVM guests. */
85#define _XEN_SYSCTL_PHYSCAP_hvm          0
86#define XEN_SYSCTL_PHYSCAP_hvm           (1u<<_XEN_SYSCTL_PHYSCAP_hvm)
87 /* The platform supports PV guests. */
88#define _XEN_SYSCTL_PHYSCAP_pv           1
89#define XEN_SYSCTL_PHYSCAP_pv            (1u<<_XEN_SYSCTL_PHYSCAP_pv)
90 /* The platform supports direct access to I/O devices with IOMMU. */
91#define _XEN_SYSCTL_PHYSCAP_directio     2
92#define XEN_SYSCTL_PHYSCAP_directio  (1u<<_XEN_SYSCTL_PHYSCAP_directio)
93/* The platform supports Hardware Assisted Paging. */
94#define _XEN_SYSCTL_PHYSCAP_hap          3
95#define XEN_SYSCTL_PHYSCAP_hap           (1u<<_XEN_SYSCTL_PHYSCAP_hap)
96/* The platform supports software paging. */
97#define _XEN_SYSCTL_PHYSCAP_shadow       4
98#define XEN_SYSCTL_PHYSCAP_shadow        (1u<<_XEN_SYSCTL_PHYSCAP_shadow)
99/* The platform supports sharing of HAP page tables with the IOMMU. */
100#define _XEN_SYSCTL_PHYSCAP_iommu_hap_pt_share 5
101#define XEN_SYSCTL_PHYSCAP_iommu_hap_pt_share  \
102    (1u << _XEN_SYSCTL_PHYSCAP_iommu_hap_pt_share)
103#define XEN_SYSCTL_PHYSCAP_vmtrace       (1u << 6)
104/* The platform supports vPMU. */
105#define XEN_SYSCTL_PHYSCAP_vpmu          (1u << 7)
106
107/* Xen supports the Grant v1 and/or v2 ABIs. */
108#define XEN_SYSCTL_PHYSCAP_gnttab_v1     (1u << 8)
109#define XEN_SYSCTL_PHYSCAP_gnttab_v2     (1u << 9)
110
111/* Max XEN_SYSCTL_PHYSCAP_* constant.  Used for ABI checking. */
112#define XEN_SYSCTL_PHYSCAP_MAX XEN_SYSCTL_PHYSCAP_gnttab_v2
113
114struct xen_sysctl_physinfo {
115    uint32_t threads_per_core;
116    uint32_t cores_per_socket;
117    uint32_t nr_cpus;     /* # CPUs currently online */
118    uint32_t max_cpu_id;  /* Largest possible CPU ID on this host */
119    uint32_t nr_nodes;    /* # nodes currently online */
120    uint32_t max_node_id; /* Largest possible node ID on this host */
121    uint32_t cpu_khz;
122    uint32_t capabilities;/* XEN_SYSCTL_PHYSCAP_??? */
123    uint64_aligned_t total_pages;
124    uint64_aligned_t free_pages;
125    uint64_aligned_t scrub_pages;
126    uint64_aligned_t outstanding_pages;
127    uint64_aligned_t max_mfn; /* Largest possible MFN on this host */
128    uint32_t hw_cap[8];
129};
130
131/*
132 * Get the ID of the current scheduler.
133 */
134/* XEN_SYSCTL_sched_id */
135struct xen_sysctl_sched_id {
136    /* OUT variable */
137    uint32_t sched_id;
138};
139
140/* Interface for controlling Xen software performance counters. */
141/* XEN_SYSCTL_perfc_op */
142/* Sub-operations: */
143#define XEN_SYSCTL_PERFCOP_reset 1   /* Reset all counters to zero. */
144#define XEN_SYSCTL_PERFCOP_query 2   /* Get perfctr information. */
145struct xen_sysctl_perfc_desc {
146    char         name[80];             /* name of perf counter */
147    uint32_t     nr_vals;              /* number of values for this counter */
148};
149typedef struct xen_sysctl_perfc_desc xen_sysctl_perfc_desc_t;
150DEFINE_XEN_GUEST_HANDLE(xen_sysctl_perfc_desc_t);
151typedef uint32_t xen_sysctl_perfc_val_t;
152DEFINE_XEN_GUEST_HANDLE(xen_sysctl_perfc_val_t);
153
154struct xen_sysctl_perfc_op {
155    /* IN variables. */
156    uint32_t       cmd;                /*  XEN_SYSCTL_PERFCOP_??? */
157    /* OUT variables. */
158    uint32_t       nr_counters;       /*  number of counters description  */
159    uint32_t       nr_vals;           /*  number of values  */
160    /* counter information (or NULL) */
161    XEN_GUEST_HANDLE_64(xen_sysctl_perfc_desc_t) desc;
162    /* counter values (or NULL) */
163    XEN_GUEST_HANDLE_64(xen_sysctl_perfc_val_t) val;
164};
165
166/* XEN_SYSCTL_getdomaininfolist */
167struct xen_sysctl_getdomaininfolist {
168    /* IN variables. */
169    domid_t               first_domain;
170    uint32_t              max_domains;
171    XEN_GUEST_HANDLE_64(xen_domctl_getdomaininfo_t) buffer;
172    /* OUT variables. */
173    uint32_t              num_domains;
174};
175
176/* Inject debug keys into Xen. */
177/* XEN_SYSCTL_debug_keys */
178struct xen_sysctl_debug_keys {
179    /* IN variables. */
180    XEN_GUEST_HANDLE_64(const_char) keys;
181    uint32_t nr_keys;
182};
183
184/* Get physical CPU information. */
185/* XEN_SYSCTL_getcpuinfo */
186struct xen_sysctl_cpuinfo {
187    uint64_aligned_t idletime;
188};
189typedef struct xen_sysctl_cpuinfo xen_sysctl_cpuinfo_t;
190DEFINE_XEN_GUEST_HANDLE(xen_sysctl_cpuinfo_t);
191struct xen_sysctl_getcpuinfo {
192    /* IN variables. */
193    uint32_t max_cpus;
194    XEN_GUEST_HANDLE_64(xen_sysctl_cpuinfo_t) info;
195    /* OUT variables. */
196    uint32_t nr_cpus;
197};
198
199/* XEN_SYSCTL_availheap */
200struct xen_sysctl_availheap {
201    /* IN variables. */
202    uint32_t min_bitwidth;  /* Smallest address width (zero if don't care). */
203    uint32_t max_bitwidth;  /* Largest address width (zero if don't care). */
204    int32_t  node;          /* NUMA node of interest (-1 for all nodes). */
205    /* OUT variables. */
206    uint64_aligned_t avail_bytes;/* Bytes available in the specified region. */
207};
208
209/* XEN_SYSCTL_get_pmstat */
210struct pm_px_val {
211    uint64_aligned_t freq;        /* Px core frequency */
212    uint64_aligned_t residency;   /* Px residency time */
213    uint64_aligned_t count;       /* Px transition count */
214};
215typedef struct pm_px_val pm_px_val_t;
216DEFINE_XEN_GUEST_HANDLE(pm_px_val_t);
217
218struct pm_px_stat {
219    uint8_t total;        /* total Px states */
220    uint8_t usable;       /* usable Px states */
221    uint8_t last;         /* last Px state */
222    uint8_t cur;          /* current Px state */
223    XEN_GUEST_HANDLE_64(uint64) trans_pt;   /* Px transition table */
224    XEN_GUEST_HANDLE_64(pm_px_val_t) pt;
225};
226
227struct pm_cx_stat {
228    uint32_t nr;    /* entry nr in triggers & residencies, including C0 */
229    uint32_t last;  /* last Cx state */
230    uint64_aligned_t idle_time;                 /* idle time from boot */
231    XEN_GUEST_HANDLE_64(uint64) triggers;    /* Cx trigger counts */
232    XEN_GUEST_HANDLE_64(uint64) residencies; /* Cx residencies */
233    uint32_t nr_pc;                          /* entry nr in pc[] */
234    uint32_t nr_cc;                          /* entry nr in cc[] */
235    /*
236     * These two arrays may (and generally will) have unused slots; slots not
237     * having a corresponding hardware register will not be written by the
238     * hypervisor. It is therefore up to the caller to put a suitable sentinel
239     * into all slots before invoking the function.
240     * Indexing is 1-biased (PC1/CC1 being at index 0).
241     */
242    XEN_GUEST_HANDLE_64(uint64) pc;
243    XEN_GUEST_HANDLE_64(uint64) cc;
244};
245
246struct xen_sysctl_get_pmstat {
247#define PMSTAT_CATEGORY_MASK 0xf0
248#define PMSTAT_PX            0x10
249#define PMSTAT_CX            0x20
250#define PMSTAT_get_max_px    (PMSTAT_PX | 0x1)
251#define PMSTAT_get_pxstat    (PMSTAT_PX | 0x2)
252#define PMSTAT_reset_pxstat  (PMSTAT_PX | 0x3)
253#define PMSTAT_get_max_cx    (PMSTAT_CX | 0x1)
254#define PMSTAT_get_cxstat    (PMSTAT_CX | 0x2)
255#define PMSTAT_reset_cxstat  (PMSTAT_CX | 0x3)
256    uint32_t type;
257    uint32_t cpuid;
258    union {
259        struct pm_px_stat getpx;
260        struct pm_cx_stat getcx;
261        /* other struct for tx, etc */
262    } u;
263};
264
265/* XEN_SYSCTL_cpu_hotplug */
266struct xen_sysctl_cpu_hotplug {
267    /* IN variables */
268    uint32_t cpu;   /* Physical cpu. */
269
270    /* Single CPU enable/disable. */
271#define XEN_SYSCTL_CPU_HOTPLUG_ONLINE  0
272#define XEN_SYSCTL_CPU_HOTPLUG_OFFLINE 1
273
274    /*
275     * SMT enable/disable.
276     *
277     * These two ops loop over all present CPUs, and either online or offline
278     * every non-primary sibling thread (those with a thread id which is not
279     * 0).  This behaviour is chosen to simplify the implementation.
280     *
281     * They are intended as a shorthand for identifying and feeding the cpu
282     * numbers individually to HOTPLUG_{ON,OFF}LINE.
283     *
284     * These are not expected to be used in conjunction with debugging options
285     * such as `maxcpus=` or when other manual configuration of offline cpus
286     * is in use.
287     */
288#define XEN_SYSCTL_CPU_HOTPLUG_SMT_ENABLE  2
289#define XEN_SYSCTL_CPU_HOTPLUG_SMT_DISABLE 3
290    uint32_t op;    /* hotplug opcode */
291};
292
293/*
294 * Get/set xen power management, include
295 * 1. cpufreq governors and related parameters
296 */
297/* XEN_SYSCTL_pm_op */
298struct xen_userspace {
299    uint32_t scaling_setspeed;
300};
301
302struct xen_ondemand {
303    uint32_t sampling_rate_max;
304    uint32_t sampling_rate_min;
305
306    uint32_t sampling_rate;
307    uint32_t up_threshold;
308};
309
310/*
311 * cpufreq para name of this structure named
312 * same as sysfs file name of native linux
313 */
314#define CPUFREQ_NAME_LEN 16
315struct xen_get_cpufreq_para {
316    /* IN/OUT variable */
317    uint32_t cpu_num;
318    uint32_t freq_num;
319    uint32_t gov_num;
320
321    /* for all governors */
322    /* OUT variable */
323    XEN_GUEST_HANDLE_64(uint32) affected_cpus;
324    XEN_GUEST_HANDLE_64(uint32) scaling_available_frequencies;
325    XEN_GUEST_HANDLE_64(char)   scaling_available_governors;
326    char scaling_driver[CPUFREQ_NAME_LEN];
327
328    uint32_t cpuinfo_cur_freq;
329    uint32_t cpuinfo_max_freq;
330    uint32_t cpuinfo_min_freq;
331    uint32_t scaling_cur_freq;
332
333    char scaling_governor[CPUFREQ_NAME_LEN];
334    uint32_t scaling_max_freq;
335    uint32_t scaling_min_freq;
336
337    /* for specific governor */
338    union {
339        struct  xen_userspace userspace;
340        struct  xen_ondemand ondemand;
341    } u;
342
343    int32_t turbo_enabled;
344};
345
346struct xen_set_cpufreq_gov {
347    char scaling_governor[CPUFREQ_NAME_LEN];
348};
349
350struct xen_set_cpufreq_para {
351    #define SCALING_MAX_FREQ           1
352    #define SCALING_MIN_FREQ           2
353    #define SCALING_SETSPEED           3
354    #define SAMPLING_RATE              4
355    #define UP_THRESHOLD               5
356
357    uint32_t ctrl_type;
358    uint32_t ctrl_value;
359};
360
361struct xen_sysctl_pm_op {
362    #define PM_PARA_CATEGORY_MASK      0xf0
363    #define CPUFREQ_PARA               0x10
364
365    /* cpufreq command type */
366    #define GET_CPUFREQ_PARA           (CPUFREQ_PARA | 0x01)
367    #define SET_CPUFREQ_GOV            (CPUFREQ_PARA | 0x02)
368    #define SET_CPUFREQ_PARA           (CPUFREQ_PARA | 0x03)
369    #define GET_CPUFREQ_AVGFREQ        (CPUFREQ_PARA | 0x04)
370
371    /* set/reset scheduler power saving option */
372    #define XEN_SYSCTL_pm_op_set_sched_opt_smt    0x21
373
374    /*
375     * cpuidle max C-state and max C-sub-state access command:
376     * Set cpuid to 0 for max C-state.
377     * Set cpuid to 1 for max C-sub-state.
378     */
379    #define XEN_SYSCTL_pm_op_get_max_cstate       0x22
380    #define XEN_SYSCTL_pm_op_set_max_cstate       0x23
381
382    /* set scheduler migration cost value */
383    #define XEN_SYSCTL_pm_op_set_vcpu_migration_delay   0x24
384    #define XEN_SYSCTL_pm_op_get_vcpu_migration_delay   0x25
385
386    /* enable/disable turbo mode when in dbs governor */
387    #define XEN_SYSCTL_pm_op_enable_turbo               0x26
388    #define XEN_SYSCTL_pm_op_disable_turbo              0x27
389
390    uint32_t cmd;
391    uint32_t cpuid;
392    union {
393        struct xen_get_cpufreq_para get_para;
394        struct xen_set_cpufreq_gov  set_gov;
395        struct xen_set_cpufreq_para set_para;
396        uint64_aligned_t get_avgfreq;
397        uint32_t                    set_sched_opt_smt;
398#define XEN_SYSCTL_CX_UNLIMITED 0xffffffff
399        uint32_t                    get_max_cstate;
400        uint32_t                    set_max_cstate;
401    } u;
402};
403
404/* XEN_SYSCTL_page_offline_op */
405struct xen_sysctl_page_offline_op {
406    /* IN: range of page to be offlined */
407#define sysctl_page_offline     1
408#define sysctl_page_online      2
409#define sysctl_query_page_offline  3
410    uint32_t cmd;
411    uint32_t start;
412    uint32_t end;
413    /* OUT: result of page offline request */
414    /*
415     * bit 0~15: result flags
416     * bit 16~31: owner
417     */
418    XEN_GUEST_HANDLE(uint32) status;
419};
420
421#define PG_OFFLINE_STATUS_MASK    (0xFFUL)
422
423/* The result is invalid, i.e. HV does not handle it */
424#define PG_OFFLINE_INVALID   (0x1UL << 0)
425
426#define PG_OFFLINE_OFFLINED  (0x1UL << 1)
427#define PG_OFFLINE_PENDING   (0x1UL << 2)
428#define PG_OFFLINE_FAILED    (0x1UL << 3)
429#define PG_OFFLINE_AGAIN     (0x1UL << 4)
430
431#define PG_ONLINE_FAILED     PG_OFFLINE_FAILED
432#define PG_ONLINE_ONLINED    PG_OFFLINE_OFFLINED
433
434#define PG_OFFLINE_STATUS_OFFLINED              (0x1UL << 1)
435#define PG_OFFLINE_STATUS_ONLINE                (0x1UL << 2)
436#define PG_OFFLINE_STATUS_OFFLINE_PENDING       (0x1UL << 3)
437#define PG_OFFLINE_STATUS_BROKEN                (0x1UL << 4)
438
439#define PG_OFFLINE_MISC_MASK    (0xFFUL << 4)
440
441/* valid when PG_OFFLINE_FAILED or PG_OFFLINE_PENDING */
442#define PG_OFFLINE_XENPAGE   (0x1UL << 8)
443#define PG_OFFLINE_DOM0PAGE  (0x1UL << 9)
444#define PG_OFFLINE_ANONYMOUS (0x1UL << 10)
445#define PG_OFFLINE_NOT_CONV_RAM   (0x1UL << 11)
446#define PG_OFFLINE_OWNED     (0x1UL << 12)
447
448#define PG_OFFLINE_BROKEN    (0x1UL << 13)
449#define PG_ONLINE_BROKEN     PG_OFFLINE_BROKEN
450
451#define PG_OFFLINE_OWNER_SHIFT 16
452
453/* XEN_SYSCTL_lockprof_op */
454/* Sub-operations: */
455#define XEN_SYSCTL_LOCKPROF_reset 1   /* Reset all profile data to zero. */
456#define XEN_SYSCTL_LOCKPROF_query 2   /* Get lock profile information. */
457/* Record-type: */
458#define LOCKPROF_TYPE_GLOBAL      0   /* global lock, idx meaningless */
459#define LOCKPROF_TYPE_PERDOM      1   /* per-domain lock, idx is domid */
460#define LOCKPROF_TYPE_N           2   /* number of types */
461struct xen_sysctl_lockprof_data {
462    char     name[40];     /* lock name (may include up to 2 %d specifiers) */
463    int32_t  type;         /* LOCKPROF_TYPE_??? */
464    int32_t  idx;          /* index (e.g. domain id) */
465    uint64_aligned_t lock_cnt;     /* # of locking succeeded */
466    uint64_aligned_t block_cnt;    /* # of wait for lock */
467    uint64_aligned_t lock_time;    /* nsecs lock held */
468    uint64_aligned_t block_time;   /* nsecs waited for lock */
469};
470typedef struct xen_sysctl_lockprof_data xen_sysctl_lockprof_data_t;
471DEFINE_XEN_GUEST_HANDLE(xen_sysctl_lockprof_data_t);
472struct xen_sysctl_lockprof_op {
473    /* IN variables. */
474    uint32_t       cmd;               /* XEN_SYSCTL_LOCKPROF_??? */
475    uint32_t       max_elem;          /* size of output buffer */
476    /* OUT variables (query only). */
477    uint32_t       nr_elem;           /* number of elements available */
478    uint64_aligned_t time;            /* nsecs of profile measurement */
479    /* profile information (or NULL) */
480    XEN_GUEST_HANDLE_64(xen_sysctl_lockprof_data_t) data;
481};
482
483/* XEN_SYSCTL_cputopoinfo */
484#define XEN_INVALID_CORE_ID     (~0U)
485#define XEN_INVALID_SOCKET_ID   (~0U)
486#define XEN_INVALID_NODE_ID     (~0U)
487
488struct xen_sysctl_cputopo {
489    uint32_t core;
490    uint32_t socket;
491    uint32_t node;
492};
493typedef struct xen_sysctl_cputopo xen_sysctl_cputopo_t;
494DEFINE_XEN_GUEST_HANDLE(xen_sysctl_cputopo_t);
495
496/*
497 * IN:
498 *  - a NULL 'cputopo' handle is a request for maximun 'num_cpus'.
499 *  - otherwise it's the number of entries in 'cputopo'
500 *
501 * OUT:
502 *  - If 'num_cpus' is less than the number Xen wants to write but the handle
503 *    handle is not a NULL one, partial data gets returned and 'num_cpus' gets
504 *    updated to reflect the intended number.
505 *  - Otherwise, 'num_cpus' shall indicate the number of entries written, which
506 *    may be less than the input value.
507 */
508struct xen_sysctl_cputopoinfo {
509    uint32_t num_cpus;
510    XEN_GUEST_HANDLE_64(xen_sysctl_cputopo_t) cputopo;
511};
512
513/* XEN_SYSCTL_numainfo */
514#define XEN_INVALID_MEM_SZ     (~0U)
515#define XEN_INVALID_NODE_DIST  (~0U)
516
517struct xen_sysctl_meminfo {
518    uint64_t memsize;
519    uint64_t memfree;
520};
521typedef struct xen_sysctl_meminfo xen_sysctl_meminfo_t;
522DEFINE_XEN_GUEST_HANDLE(xen_sysctl_meminfo_t);
523
524/*
525 * IN:
526 *  - Both 'meminfo' and 'distance' handles being null is a request
527 *    for maximum value of 'num_nodes'.
528 *  - Otherwise it's the number of entries in 'meminfo' and square root
529 *    of number of entries in 'distance' (when corresponding handle is
530 *    non-null)
531 *
532 * OUT:
533 *  - If 'num_nodes' is less than the number Xen wants to write but either
534 *    handle is not a NULL one, partial data gets returned and 'num_nodes'
535 *    gets updated to reflect the intended number.
536 *  - Otherwise, 'num_nodes' shall indicate the number of entries written, which
537 *    may be less than the input value.
538 */
539
540struct xen_sysctl_numainfo {
541    uint32_t num_nodes;
542
543    XEN_GUEST_HANDLE_64(xen_sysctl_meminfo_t) meminfo;
544
545    /*
546     * Distance between nodes 'i' and 'j' is stored in index 'i*N + j',
547     * where N is the number of nodes that will be returned in 'num_nodes'
548     * (i.e. not 'num_nodes' provided by the caller)
549     */
550    XEN_GUEST_HANDLE_64(uint32) distance;
551};
552
553/* XEN_SYSCTL_cpupool_op */
554#define XEN_SYSCTL_CPUPOOL_OP_CREATE                1  /* C */
555#define XEN_SYSCTL_CPUPOOL_OP_DESTROY               2  /* D */
556#define XEN_SYSCTL_CPUPOOL_OP_INFO                  3  /* I */
557#define XEN_SYSCTL_CPUPOOL_OP_ADDCPU                4  /* A */
558#define XEN_SYSCTL_CPUPOOL_OP_RMCPU                 5  /* R */
559#define XEN_SYSCTL_CPUPOOL_OP_MOVEDOMAIN            6  /* M */
560#define XEN_SYSCTL_CPUPOOL_OP_FREEINFO              7  /* F */
561#define XEN_SYSCTL_CPUPOOL_PAR_ANY     0xFFFFFFFF
562struct xen_sysctl_cpupool_op {
563    uint32_t op;          /* IN */
564    uint32_t cpupool_id;  /* IN: CDIARM OUT: CI */
565    uint32_t sched_id;    /* IN: C      OUT: I  */
566    uint32_t domid;       /* IN: M              */
567    uint32_t cpu;         /* IN: AR             */
568    uint32_t n_dom;       /*            OUT: I  */
569    struct xenctl_bitmap cpumap; /*     OUT: IF */
570};
571
572/*
573 * Error return values of cpupool operations:
574 *
575 * -EADDRINUSE:
576 *  XEN_SYSCTL_CPUPOOL_OP_RMCPU: A vcpu is temporarily pinned to the cpu
577 *    which is to be removed from a cpupool.
578 * -EADDRNOTAVAIL:
579 *  XEN_SYSCTL_CPUPOOL_OP_ADDCPU, XEN_SYSCTL_CPUPOOL_OP_RMCPU: A previous
580 *    request to remove a cpu from a cpupool was terminated with -EAGAIN
581 *    and has not been retried using the same parameters.
582 * -EAGAIN:
583 *  XEN_SYSCTL_CPUPOOL_OP_RMCPU: The cpu can't be removed from the cpupool
584 *    as it is active in the hypervisor. A retry will succeed soon.
585 * -EBUSY:
586 *  XEN_SYSCTL_CPUPOOL_OP_DESTROY, XEN_SYSCTL_CPUPOOL_OP_RMCPU: A cpupool
587 *    can't be destroyed or the last cpu can't be removed as there is still
588 *    a running domain in that cpupool.
589 * -EEXIST:
590 *  XEN_SYSCTL_CPUPOOL_OP_CREATE: A cpupool_id was specified and is already
591 *    existing.
592 * -EINVAL:
593 *  XEN_SYSCTL_CPUPOOL_OP_ADDCPU, XEN_SYSCTL_CPUPOOL_OP_RMCPU: An illegal
594 *    cpu was specified (cpu does not exist).
595 *  XEN_SYSCTL_CPUPOOL_OP_MOVEDOMAIN: An illegal domain was specified
596 *    (domain id illegal or not suitable for operation).
597 * -ENODEV:
598 *  XEN_SYSCTL_CPUPOOL_OP_ADDCPU, XEN_SYSCTL_CPUPOOL_OP_RMCPU: The specified
599 *    cpu is either not free (add) or not member of the specified cpupool
600 *    (remove).
601 * -ENOENT:
602 *  all: The cpupool with the specified cpupool_id doesn't exist.
603 *
604 * Some common error return values like -ENOMEM and -EFAULT are possible for
605 * all the operations.
606 */
607
608#define ARINC653_MAX_DOMAINS_PER_SCHEDULE   64
609/*
610 * This structure is used to pass a new ARINC653 schedule from a
611 * privileged domain (ie dom0) to Xen.
612 */
613struct xen_sysctl_arinc653_schedule {
614    /* major_frame holds the time for the new schedule's major frame
615     * in nanoseconds. */
616    uint64_aligned_t     major_frame;
617    /* num_sched_entries holds how many of the entries in the
618     * sched_entries[] array are valid. */
619    uint8_t     num_sched_entries;
620    /* The sched_entries array holds the actual schedule entries. */
621    struct {
622        /* dom_handle must match a domain's UUID */
623        xen_domain_handle_t dom_handle;
624        /* If a domain has multiple VCPUs, vcpu_id specifies which one
625         * this schedule entry applies to. It should be set to 0 if
626         * there is only one VCPU for the domain. */
627        unsigned int vcpu_id;
628        /* runtime specifies the amount of time that should be allocated
629         * to this VCPU per major frame. It is specified in nanoseconds */
630        uint64_aligned_t runtime;
631    } sched_entries[ARINC653_MAX_DOMAINS_PER_SCHEDULE];
632};
633typedef struct xen_sysctl_arinc653_schedule xen_sysctl_arinc653_schedule_t;
634DEFINE_XEN_GUEST_HANDLE(xen_sysctl_arinc653_schedule_t);
635
636/*
637 * Valid range for context switch rate limit (in microseconds).
638 * Applicable to Credit and Credit2 schedulers.
639 */
640#define XEN_SYSCTL_SCHED_RATELIMIT_MAX 500000
641#define XEN_SYSCTL_SCHED_RATELIMIT_MIN 100
642
643struct xen_sysctl_credit_schedule {
644    /* Length of timeslice in milliseconds */
645#define XEN_SYSCTL_CSCHED_TSLICE_MAX 1000
646#define XEN_SYSCTL_CSCHED_TSLICE_MIN 1
647    unsigned tslice_ms;
648    unsigned ratelimit_us;
649    /*
650     * How long we consider a vCPU to be cache-hot on the
651     * CPU where it has run (max 100ms, in microseconds)
652    */
653#define XEN_SYSCTL_CSCHED_MGR_DLY_MAX_US (100 * 1000)
654    unsigned vcpu_migr_delay_us;
655};
656
657struct xen_sysctl_credit2_schedule {
658    unsigned ratelimit_us;
659};
660
661/* XEN_SYSCTL_scheduler_op */
662/* Set or get info? */
663#define XEN_SYSCTL_SCHEDOP_putinfo 0
664#define XEN_SYSCTL_SCHEDOP_getinfo 1
665struct xen_sysctl_scheduler_op {
666    uint32_t cpupool_id; /* Cpupool whose scheduler is to be targetted. */
667    uint32_t sched_id;   /* XEN_SCHEDULER_* (domctl.h) */
668    uint32_t cmd;        /* XEN_SYSCTL_SCHEDOP_* */
669    union {
670        struct xen_sysctl_sched_arinc653 {
671            XEN_GUEST_HANDLE_64(xen_sysctl_arinc653_schedule_t) schedule;
672        } sched_arinc653;
673        struct xen_sysctl_credit_schedule sched_credit;
674        struct xen_sysctl_credit2_schedule sched_credit2;
675    } u;
676};
677
678/*
679 * Output format of gcov data:
680 *
681 * XEN_GCOV_FORMAT_MAGIC XEN_GCOV_RECORD ... XEN_GCOV_RECORD
682 *
683 * That is, one magic number followed by 0 or more record.
684 *
685 * The magic number is stored as an uint32_t field.
686 *
687 * The record is packed and variable in length. It has the form:
688 *
689 *  filename: a NULL terminated path name extracted from gcov, used to
690 *            create the name of gcda file.
691 *  size:     a uint32_t field indicating the size of the payload, the
692 *            unit is byte.
693 *  payload:  the actual payload, length is `size' bytes.
694 *
695 * Userspace tool will split the record to different files.
696 */
697
698#define XEN_GCOV_FORMAT_MAGIC    0x58434f56 /* XCOV */
699
700/*
701 * Ouput format of LLVM coverage data is just a raw stream, as would be
702 * written by the compiler_rt run time library into a .profraw file. There
703 * are no special Xen tags or delimiters because none are needed.
704 */
705
706#define XEN_SYSCTL_COVERAGE_get_size 0 /* Get total size of output data */
707#define XEN_SYSCTL_COVERAGE_read     1 /* Read output data */
708#define XEN_SYSCTL_COVERAGE_reset    2 /* Reset all counters */
709
710struct xen_sysctl_coverage_op {
711    uint32_t cmd;
712    uint32_t size; /* IN/OUT: size of the buffer  */
713    XEN_GUEST_HANDLE_64(char) buffer; /* OUT */
714};
715
716#define XEN_SYSCTL_PSR_CMT_get_total_rmid            0
717#define XEN_SYSCTL_PSR_CMT_get_l3_upscaling_factor   1
718/* The L3 cache size is returned in KB unit */
719#define XEN_SYSCTL_PSR_CMT_get_l3_cache_size         2
720#define XEN_SYSCTL_PSR_CMT_enabled                   3
721#define XEN_SYSCTL_PSR_CMT_get_l3_event_mask         4
722struct xen_sysctl_psr_cmt_op {
723    uint32_t cmd;       /* IN: XEN_SYSCTL_PSR_CMT_* */
724    uint32_t flags;     /* padding variable, may be extended for future use */
725    union {
726        uint64_t data;  /* OUT */
727        struct {
728            uint32_t cpu;   /* IN */
729            uint32_t rsvd;
730        } l3_cache;
731    } u;
732};
733
734/* XEN_SYSCTL_pcitopoinfo */
735#define XEN_INVALID_DEV (XEN_INVALID_NODE_ID - 1)
736struct xen_sysctl_pcitopoinfo {
737    /*
738     * IN: Number of elements in 'devs' and 'nodes' arrays.
739     * OUT: Number of processed elements of those arrays.
740     */
741    uint32_t num_devs;
742
743    /* IN: list of devices for which node IDs are requested. */
744    XEN_GUEST_HANDLE_64(physdev_pci_device_t) devs;
745
746    /*
747     * OUT: node identifier for each device.
748     * If information for a particular device is not available then
749     * corresponding entry will be set to XEN_INVALID_NODE_ID. If
750     * device is not known to the hypervisor then XEN_INVALID_DEV
751     * will be provided.
752     */
753    XEN_GUEST_HANDLE_64(uint32) nodes;
754};
755
756#define XEN_SYSCTL_PSR_get_l3_info               0
757#define XEN_SYSCTL_PSR_get_l2_info               1
758#define XEN_SYSCTL_PSR_get_mba_info              2
759struct xen_sysctl_psr_alloc {
760    uint32_t cmd;       /* IN: XEN_SYSCTL_PSR_* */
761    uint32_t target;    /* IN */
762    union {
763        struct {
764            uint32_t cbm_len;   /* OUT: CBM length */
765            uint32_t cos_max;   /* OUT: Maximum COS */
766#define XEN_SYSCTL_PSR_CAT_L3_CDP       (1u << 0)
767            uint32_t flags;     /* OUT: CAT flags */
768        } cat_info;
769
770        struct {
771            uint32_t thrtl_max; /* OUT: Maximum throttle */
772            uint32_t cos_max;   /* OUT: Maximum COS */
773#define XEN_SYSCTL_PSR_MBA_LINEAR      (1u << 0)
774            uint32_t flags;     /* OUT: MBA flags */
775        } mba_info;
776    } u;
777};
778
779/*
780 * XEN_SYSCTL_get_cpu_levelling_caps (x86 specific)
781 *
782 * Return hardware capabilities concerning masking or faulting of the cpuid
783 * instruction for PV guests.
784 */
785struct xen_sysctl_cpu_levelling_caps {
786#define XEN_SYSCTL_CPU_LEVELCAP_faulting    (1ul <<  0) /* CPUID faulting    */
787#define XEN_SYSCTL_CPU_LEVELCAP_ecx         (1ul <<  1) /* 0x00000001.ecx    */
788#define XEN_SYSCTL_CPU_LEVELCAP_edx         (1ul <<  2) /* 0x00000001.edx    */
789#define XEN_SYSCTL_CPU_LEVELCAP_extd_ecx    (1ul <<  3) /* 0x80000001.ecx    */
790#define XEN_SYSCTL_CPU_LEVELCAP_extd_edx    (1ul <<  4) /* 0x80000001.edx    */
791#define XEN_SYSCTL_CPU_LEVELCAP_xsave_eax   (1ul <<  5) /* 0x0000000D:1.eax  */
792#define XEN_SYSCTL_CPU_LEVELCAP_thermal_ecx (1ul <<  6) /* 0x00000006.ecx    */
793#define XEN_SYSCTL_CPU_LEVELCAP_l7s0_eax    (1ul <<  7) /* 0x00000007:0.eax  */
794#define XEN_SYSCTL_CPU_LEVELCAP_l7s0_ebx    (1ul <<  8) /* 0x00000007:0.ebx  */
795    uint32_t caps;
796};
797
798/*
799 * XEN_SYSCTL_get_cpu_featureset (x86 specific)
800 *
801 * Return information about featuresets available on this host.
802 *  -  Raw: The real cpuid values.
803 *  - Host: The values Xen is using, (after command line overrides, etc).
804 *  -   PV: Maximum set of features which can be given to a PV guest.
805 *  -  HVM: Maximum set of features which can be given to a HVM guest.
806 * May fail with -EOPNOTSUPP if querying for PV or HVM data when support is
807 * compiled out of Xen.
808 */
809struct xen_sysctl_cpu_featureset {
810#define XEN_SYSCTL_cpu_featureset_raw      0
811#define XEN_SYSCTL_cpu_featureset_host     1
812#define XEN_SYSCTL_cpu_featureset_pv       2
813#define XEN_SYSCTL_cpu_featureset_hvm      3
814    uint32_t index;       /* IN: Which featureset to query? */
815    uint32_t nr_features; /* IN/OUT: Number of entries in/written to
816                           * 'features', or the maximum number of features if
817                           * the guest handle is NULL.  NB. All featuresets
818                           * come from the same numberspace, so have the same
819                           * maximum length. */
820    XEN_GUEST_HANDLE_64(uint32) features; /* OUT: */
821};
822
823/*
824 * XEN_SYSCTL_LIVEPATCH_op
825 *
826 * Refer to the docs/unstable/misc/livepatch.markdown
827 * for the design details of this hypercall.
828 *
829 * There are four sub-ops:
830 *  XEN_SYSCTL_LIVEPATCH_UPLOAD (0)
831 *  XEN_SYSCTL_LIVEPATCH_GET (1)
832 *  XEN_SYSCTL_LIVEPATCH_LIST (2)
833 *  XEN_SYSCTL_LIVEPATCH_ACTION (3)
834 *
835 * The normal sequence of sub-ops is to:
836 *  1) XEN_SYSCTL_LIVEPATCH_UPLOAD to upload the payload. If errors STOP.
837 *  2) XEN_SYSCTL_LIVEPATCH_GET to check the `->rc`. If -XEN_EAGAIN spin.
838 *     If zero go to next step.
839 *  3) XEN_SYSCTL_LIVEPATCH_ACTION with LIVEPATCH_ACTION_APPLY to apply the patch.
840 *  4) XEN_SYSCTL_LIVEPATCH_GET to check the `->rc`. If in -XEN_EAGAIN spin.
841 *     If zero exit with success.
842 */
843
844#define LIVEPATCH_PAYLOAD_VERSION 2
845/*
846 * .livepatch.funcs structure layout defined in the `Payload format`
847 * section in the Live Patch design document.
848 *
849 * We guard this with __XEN__ as toolstacks SHOULD not use it.
850 */
851#ifdef __XEN__
852#define LIVEPATCH_OPAQUE_SIZE 31
853
854struct livepatch_expectation {
855    uint8_t enabled : 1;
856    uint8_t len : 5;        /* Length of data up to LIVEPATCH_OPAQUE_SIZE
857                               (5 bits is enough for now) */
858    uint8_t rsv : 2;        /* Reserved. Zero value */
859    uint8_t data[LIVEPATCH_OPAQUE_SIZE]; /* Same size as opaque[] buffer of
860                                            struct livepatch_func. This is the
861                                            max number of bytes to be patched */
862};
863typedef struct livepatch_expectation livepatch_expectation_t;
864
865typedef enum livepatch_func_state {
866    LIVEPATCH_FUNC_NOT_APPLIED,
867    LIVEPATCH_FUNC_APPLIED
868} livepatch_func_state_t;
869
870struct livepatch_func {
871    const char *name;       /* Name of function to be patched. */
872    void *new_addr;
873    void *old_addr;
874    uint32_t new_size;
875    uint32_t old_size;
876    uint8_t version;        /* MUST be LIVEPATCH_PAYLOAD_VERSION. */
877    uint8_t opaque[LIVEPATCH_OPAQUE_SIZE];
878    uint8_t applied;
879    uint8_t _pad[7];
880    livepatch_expectation_t expect;
881};
882typedef struct livepatch_func livepatch_func_t;
883#endif
884
885/*
886 * Structure describing an ELF payload. Uniquely identifies the
887 * payload. Should be human readable.
888 * Recommended length is upto XEN_LIVEPATCH_NAME_SIZE.
889 * Includes the NUL terminator.
890 */
891#define XEN_LIVEPATCH_NAME_SIZE 128
892struct xen_livepatch_name {
893    XEN_GUEST_HANDLE_64(char) name;         /* IN: pointer to name. */
894    uint16_t size;                          /* IN: size of name. May be upto
895                                               XEN_LIVEPATCH_NAME_SIZE. */
896    uint16_t pad[3];                        /* IN: MUST be zero. */
897};
898
899/*
900 * Upload a payload to the hypervisor. The payload is verified
901 * against basic checks and if there are any issues the proper return code
902 * will be returned. The payload is not applied at this time - that is
903 * controlled by XEN_SYSCTL_LIVEPATCH_ACTION.
904 *
905 * The return value is zero if the payload was succesfully uploaded.
906 * Otherwise an EXX return value is provided. Duplicate `name` are not
907 * supported.
908 *
909 * The payload at this point is verified against basic checks.
910 *
911 * The `payload` is the ELF payload as mentioned in the `Payload format`
912 * section in the Live Patch design document.
913 */
914#define XEN_SYSCTL_LIVEPATCH_UPLOAD 0
915struct xen_sysctl_livepatch_upload {
916    struct xen_livepatch_name name;         /* IN, name of the patch. */
917    uint64_t size;                          /* IN, size of the ELF file. */
918    XEN_GUEST_HANDLE_64(uint8) payload;     /* IN, the ELF file. */
919};
920
921/*
922 * Retrieve an status of an specific payload.
923 *
924 * Upon completion the `struct xen_livepatch_status` is updated.
925 *
926 * The return value is zero on success and XEN_EXX on failure. This operation
927 * is synchronous and does not require preemption.
928 */
929#define XEN_SYSCTL_LIVEPATCH_GET 1
930
931struct xen_livepatch_status {
932#define LIVEPATCH_STATE_CHECKED      1
933#define LIVEPATCH_STATE_APPLIED      2
934    uint32_t state;                /* OUT: LIVEPATCH_STATE_*. */
935    int32_t rc;                    /* OUT: 0 if no error, otherwise -XEN_EXX. */
936};
937typedef struct xen_livepatch_status xen_livepatch_status_t;
938DEFINE_XEN_GUEST_HANDLE(xen_livepatch_status_t);
939
940struct xen_sysctl_livepatch_get {
941    struct xen_livepatch_name name;         /* IN, name of the payload. */
942    struct xen_livepatch_status status;     /* IN/OUT, state of it. */
943};
944
945/*
946 * Retrieve an array of abbreviated status, names and metadata of payloads that
947 * are loaded in the hypervisor.
948 *
949 * If the hypercall returns an positive number, it is the number (up to `nr`)
950 * of the payloads returned, along with `nr` updated with the number of remaining
951 * payloads, `version` updated (it may be the same across hypercalls. If it varies
952 * the data is stale and further calls could fail), `name_total_size` and
953 * `metadata_total_size` containing total sizes of transferred data for both the
954 * arrays.
955 * The `status`, `name`, `len`, `metadata` and `metadata_len` are updated at their
956 * designed index value (`idx`) with the returned value of data.
957 *
958 * If the hypercall returns E2BIG the `nr` is too big and should be
959 * lowered. The upper limit of `nr` is left to the implemention.
960 *
961 * Note that due to the asynchronous nature of hypercalls the domain might have
962 * added or removed the number of payloads making this information stale. It is
963 * the responsibility of the toolstack to use the `version` field to check
964 * between each invocation. if the version differs it should discard the stale
965 * data and start from scratch. It is OK for the toolstack to use the new
966 * `version` field.
967 */
968#define XEN_SYSCTL_LIVEPATCH_LIST 2
969struct xen_sysctl_livepatch_list {
970    uint32_t version;                       /* OUT: Hypervisor stamps value.
971                                               If varies between calls, we are
972                                             * getting stale data. */
973    uint32_t idx;                           /* IN: Index into hypervisor list. */
974    uint32_t nr;                            /* IN: How many status, name, and len
975                                               should fill out. Can be zero to get
976                                               amount of payloads and version.
977                                               OUT: How many payloads left. */
978    uint32_t pad;                           /* IN: Must be zero. */
979    uint32_t name_total_size;               /* OUT: Total size of all transfer names */
980    uint32_t metadata_total_size;           /* OUT: Total size of all transfer metadata */
981    XEN_GUEST_HANDLE_64(xen_livepatch_status_t) status;  /* OUT. Must have enough
982                                               space allocate for nr of them. */
983    XEN_GUEST_HANDLE_64(char) name;         /* OUT: Array of names. Each member
984                                               may have an arbitrary length up to
985                                               XEN_LIVEPATCH_NAME_SIZE bytes. Must have
986                                               nr of them. */
987    XEN_GUEST_HANDLE_64(uint32) len;        /* OUT: Array of lengths of name's.
988                                               Must have nr of them. */
989    XEN_GUEST_HANDLE_64(char) metadata;     /* OUT: Array of metadata strings. Each
990                                               member may have an arbitrary length.
991                                               Must have nr of them. */
992    XEN_GUEST_HANDLE_64(uint32) metadata_len;  /* OUT: Array of lengths of metadata's.
993                                                  Must have nr of them. */
994};
995
996/*
997 * Perform an operation on the payload structure referenced by the `name` field.
998 * The operation request is asynchronous and the status should be retrieved
999 * by using either XEN_SYSCTL_LIVEPATCH_GET or XEN_SYSCTL_LIVEPATCH_LIST hypercall.
1000 */
1001#define XEN_SYSCTL_LIVEPATCH_ACTION 3
1002struct xen_sysctl_livepatch_action {
1003    struct xen_livepatch_name name;         /* IN, name of the patch. */
1004#define LIVEPATCH_ACTION_UNLOAD       1
1005#define LIVEPATCH_ACTION_REVERT       2
1006#define LIVEPATCH_ACTION_APPLY        3
1007#define LIVEPATCH_ACTION_REPLACE      4
1008    uint32_t cmd;                           /* IN: LIVEPATCH_ACTION_*. */
1009    uint32_t timeout;                       /* IN: If zero then uses */
1010                                            /* hypervisor default. */
1011                                            /* Or upper bound of time (ns) */
1012                                            /* for operation to take. */
1013
1014/*
1015 * Override default inter-module buildid dependency chain enforcement.
1016 * Check only if module is built for given hypervisor by comparing buildid.
1017 */
1018#define LIVEPATCH_ACTION_APPLY_NODEPS (1 << 0)
1019    uint32_t flags;                         /* IN: action flags. */
1020                                            /* Provide additional parameters */
1021                                            /* for an action. */
1022    uint32_t pad;                           /* IN: Always zero. */
1023};
1024
1025struct xen_sysctl_livepatch_op {
1026    uint32_t cmd;                           /* IN: XEN_SYSCTL_LIVEPATCH_*. */
1027    uint32_t pad;                           /* IN: Always zero. */
1028    union {
1029        struct xen_sysctl_livepatch_upload upload;
1030        struct xen_sysctl_livepatch_list list;
1031        struct xen_sysctl_livepatch_get get;
1032        struct xen_sysctl_livepatch_action action;
1033    } u;
1034};
1035
1036#if defined(__i386__) || defined(__x86_64__)
1037/*
1038 * XEN_SYSCTL_get_cpu_policy (x86 specific)
1039 *
1040 * Return information about CPUID and MSR policies available on this host.
1041 *  -       Raw: The real H/W values.
1042 *  -      Host: The values Xen is using, (after command line overrides, etc).
1043 *  -     Max_*: Maximum set of features a PV or HVM guest can use.  Includes
1044 *               experimental features outside of security support.
1045 *  - Default_*: Default set of features a PV or HVM guest can use.  This is
1046 *               the security supported set.
1047 * May fail with -EOPNOTSUPP if querying for PV or HVM data when support is
1048 * compiled out of Xen.
1049 */
1050struct xen_sysctl_cpu_policy {
1051#define XEN_SYSCTL_cpu_policy_raw          0
1052#define XEN_SYSCTL_cpu_policy_host         1
1053#define XEN_SYSCTL_cpu_policy_pv_max       2
1054#define XEN_SYSCTL_cpu_policy_hvm_max      3
1055#define XEN_SYSCTL_cpu_policy_pv_default   4
1056#define XEN_SYSCTL_cpu_policy_hvm_default  5
1057    uint32_t index;       /* IN: Which policy to query? */
1058    uint32_t nr_leaves;   /* IN/OUT: Number of leaves in/written to
1059                           * 'cpuid_policy', or the maximum number of leaves
1060                           * if the guest handle is NULL. */
1061    uint32_t nr_msrs;     /* IN/OUT: Number of MSRs in/written to
1062                           * 'msr_policy', or the maximum number of MSRs if
1063                           * the guest handle is NULL. */
1064    uint32_t _rsvd;       /* Must be zero. */
1065    XEN_GUEST_HANDLE_64(xen_cpuid_leaf_t) cpuid_policy; /* OUT */
1066    XEN_GUEST_HANDLE_64(xen_msr_entry_t) msr_policy;    /* OUT */
1067};
1068typedef struct xen_sysctl_cpu_policy xen_sysctl_cpu_policy_t;
1069DEFINE_XEN_GUEST_HANDLE(xen_sysctl_cpu_policy_t);
1070#endif
1071
1072struct xen_sysctl {
1073    uint32_t cmd;
1074#define XEN_SYSCTL_readconsole                    1
1075#define XEN_SYSCTL_tbuf_op                        2
1076#define XEN_SYSCTL_physinfo                       3
1077#define XEN_SYSCTL_sched_id                       4
1078#define XEN_SYSCTL_perfc_op                       5
1079#define XEN_SYSCTL_getdomaininfolist              6
1080#define XEN_SYSCTL_debug_keys                     7
1081#define XEN_SYSCTL_getcpuinfo                     8
1082#define XEN_SYSCTL_availheap                      9
1083#define XEN_SYSCTL_get_pmstat                    10
1084#define XEN_SYSCTL_cpu_hotplug                   11
1085#define XEN_SYSCTL_pm_op                         12
1086#define XEN_SYSCTL_page_offline_op               14
1087#define XEN_SYSCTL_lockprof_op                   15
1088#define XEN_SYSCTL_cputopoinfo                   16
1089#define XEN_SYSCTL_numainfo                      17
1090#define XEN_SYSCTL_cpupool_op                    18
1091#define XEN_SYSCTL_scheduler_op                  19
1092#define XEN_SYSCTL_coverage_op                   20
1093#define XEN_SYSCTL_psr_cmt_op                    21
1094#define XEN_SYSCTL_pcitopoinfo                   22
1095#define XEN_SYSCTL_psr_alloc                     23
1096/* #define XEN_SYSCTL_tmem_op                       24 */
1097#define XEN_SYSCTL_get_cpu_levelling_caps        25
1098#define XEN_SYSCTL_get_cpu_featureset            26
1099#define XEN_SYSCTL_livepatch_op                  27
1100/* #define XEN_SYSCTL_set_parameter              28 */
1101#define XEN_SYSCTL_get_cpu_policy                29
1102    uint32_t interface_version; /* XEN_SYSCTL_INTERFACE_VERSION */
1103    union {
1104        struct xen_sysctl_readconsole       readconsole;
1105        struct xen_sysctl_tbuf_op           tbuf_op;
1106        struct xen_sysctl_physinfo          physinfo;
1107        struct xen_sysctl_cputopoinfo       cputopoinfo;
1108        struct xen_sysctl_pcitopoinfo       pcitopoinfo;
1109        struct xen_sysctl_numainfo          numainfo;
1110        struct xen_sysctl_sched_id          sched_id;
1111        struct xen_sysctl_perfc_op          perfc_op;
1112        struct xen_sysctl_getdomaininfolist getdomaininfolist;
1113        struct xen_sysctl_debug_keys        debug_keys;
1114        struct xen_sysctl_getcpuinfo        getcpuinfo;
1115        struct xen_sysctl_availheap         availheap;
1116        struct xen_sysctl_get_pmstat        get_pmstat;
1117        struct xen_sysctl_cpu_hotplug       cpu_hotplug;
1118        struct xen_sysctl_pm_op             pm_op;
1119        struct xen_sysctl_page_offline_op   page_offline;
1120        struct xen_sysctl_lockprof_op       lockprof_op;
1121        struct xen_sysctl_cpupool_op        cpupool_op;
1122        struct xen_sysctl_scheduler_op      scheduler_op;
1123        struct xen_sysctl_coverage_op       coverage_op;
1124        struct xen_sysctl_psr_cmt_op        psr_cmt_op;
1125        struct xen_sysctl_psr_alloc         psr_alloc;
1126        struct xen_sysctl_cpu_levelling_caps cpu_levelling_caps;
1127        struct xen_sysctl_cpu_featureset    cpu_featureset;
1128        struct xen_sysctl_livepatch_op      livepatch;
1129#if defined(__i386__) || defined(__x86_64__)
1130        struct xen_sysctl_cpu_policy        cpu_policy;
1131#endif
1132        uint8_t                             pad[128];
1133    } u;
1134};
1135typedef struct xen_sysctl xen_sysctl_t;
1136DEFINE_XEN_GUEST_HANDLE(xen_sysctl_t);
1137
1138#endif /* __XEN_PUBLIC_SYSCTL_H__ */
1139
1140/*
1141 * Local variables:
1142 * mode: C
1143 * c-file-style: "BSD"
1144 * c-basic-offset: 4
1145 * tab-width: 4
1146 * indent-tabs-mode: nil
1147 * End:
1148 */
1149