acpi_cpu.c revision 209059
1/*-
2 * Copyright (c) 2003-2005 Nate Lawson (SDG)
3 * Copyright (c) 2001 Michael Smith
4 * All rights reserved.
5 *
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
8 * are met:
9 * 1. Redistributions of source code must retain the above copyright
10 *    notice, this list of conditions and the following disclaimer.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 *    notice, this list of conditions and the following disclaimer in the
13 *    documentation and/or other materials provided with the distribution.
14 *
15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
16 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18 * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
19 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
21 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
25 * SUCH DAMAGE.
26 */
27
28#include <sys/cdefs.h>
29__FBSDID("$FreeBSD: head/sys/dev/acpica/acpi_cpu.c 209059 2010-06-11 18:46:34Z jhb $");
30
31#include "opt_acpi.h"
32#include <sys/param.h>
33#include <sys/bus.h>
34#include <sys/cpu.h>
35#include <sys/kernel.h>
36#include <sys/malloc.h>
37#include <sys/module.h>
38#include <sys/pcpu.h>
39#include <sys/power.h>
40#include <sys/proc.h>
41#include <sys/sbuf.h>
42#include <sys/smp.h>
43
44#include <dev/pci/pcivar.h>
45#include <machine/atomic.h>
46#include <machine/bus.h>
47#include <sys/rman.h>
48
49#include <contrib/dev/acpica/include/acpi.h>
50#include <contrib/dev/acpica/include/accommon.h>
51
52#include <dev/acpica/acpivar.h>
53
54/*
55 * Support for ACPI Processor devices, including C[1-3] sleep states.
56 */
57
58/* Hooks for the ACPI CA debugging infrastructure */
59#define _COMPONENT	ACPI_PROCESSOR
60ACPI_MODULE_NAME("PROCESSOR")
61
62struct acpi_cx {
63    struct resource	*p_lvlx;	/* Register to read to enter state. */
64    uint32_t		 type;		/* C1-3 (C4 and up treated as C3). */
65    uint32_t		 trans_lat;	/* Transition latency (usec). */
66    uint32_t		 power;		/* Power consumed (mW). */
67    int			 res_type;	/* Resource type for p_lvlx. */
68};
69#define MAX_CX_STATES	 8
70
71struct acpi_cpu_softc {
72    device_t		 cpu_dev;
73    ACPI_HANDLE		 cpu_handle;
74    struct pcpu		*cpu_pcpu;
75    uint32_t		 cpu_acpi_id;	/* ACPI processor id */
76    uint32_t		 cpu_p_blk;	/* ACPI P_BLK location */
77    uint32_t		 cpu_p_blk_len;	/* P_BLK length (must be 6). */
78    struct acpi_cx	 cpu_cx_states[MAX_CX_STATES];
79    int			 cpu_cx_count;	/* Number of valid Cx states. */
80    int			 cpu_prev_sleep;/* Last idle sleep duration. */
81    int			 cpu_features;	/* Child driver supported features. */
82    /* Runtime state. */
83    int			 cpu_non_c3;	/* Index of lowest non-C3 state. */
84    u_int		 cpu_cx_stats[MAX_CX_STATES];/* Cx usage history. */
85    /* Values for sysctl. */
86    struct sysctl_ctx_list cpu_sysctl_ctx;
87    struct sysctl_oid	*cpu_sysctl_tree;
88    int			 cpu_cx_lowest;
89    char 		 cpu_cx_supported[64];
90    int			 cpu_rid;
91};
92
93struct acpi_cpu_device {
94    struct resource_list	ad_rl;
95};
96
97#define CPU_GET_REG(reg, width) 					\
98    (bus_space_read_ ## width(rman_get_bustag((reg)), 			\
99		      rman_get_bushandle((reg)), 0))
100#define CPU_SET_REG(reg, width, val)					\
101    (bus_space_write_ ## width(rman_get_bustag((reg)), 			\
102		       rman_get_bushandle((reg)), 0, (val)))
103
104#define PM_USEC(x)	 ((x) >> 2)	/* ~4 clocks per usec (3.57955 Mhz) */
105
106#define ACPI_NOTIFY_CX_STATES	0x81	/* _CST changed. */
107
108#define CPU_QUIRK_NO_C3		(1<<0)	/* C3-type states are not usable. */
109#define CPU_QUIRK_NO_BM_CTRL	(1<<2)	/* No bus mastering control. */
110
111#define PCI_VENDOR_INTEL	0x8086
112#define PCI_DEVICE_82371AB_3	0x7113	/* PIIX4 chipset for quirks. */
113#define PCI_REVISION_A_STEP	0
114#define PCI_REVISION_B_STEP	1
115#define PCI_REVISION_4E		2
116#define PCI_REVISION_4M		3
117#define PIIX4_DEVACTB_REG	0x58
118#define PIIX4_BRLD_EN_IRQ0	(1<<0)
119#define PIIX4_BRLD_EN_IRQ	(1<<1)
120#define PIIX4_BRLD_EN_IRQ8	(1<<5)
121#define PIIX4_STOP_BREAK_MASK	(PIIX4_BRLD_EN_IRQ0 | PIIX4_BRLD_EN_IRQ | PIIX4_BRLD_EN_IRQ8)
122#define PIIX4_PCNTRL_BST_EN	(1<<10)
123
124/* Platform hardware resource information. */
125static uint32_t		 cpu_smi_cmd;	/* Value to write to SMI_CMD. */
126static uint8_t		 cpu_cst_cnt;	/* Indicate we are _CST aware. */
127static int		 cpu_quirks;	/* Indicate any hardware bugs. */
128
129/* Runtime state. */
130static int		 cpu_disable_idle; /* Disable entry to idle function */
131static int		 cpu_cx_count;	/* Number of valid Cx states */
132
133/* Values for sysctl. */
134static struct sysctl_ctx_list cpu_sysctl_ctx;
135static struct sysctl_oid *cpu_sysctl_tree;
136static int		 cpu_cx_generic;
137static int		 cpu_cx_lowest;
138
139static device_t		*cpu_devices;
140static int		 cpu_ndevices;
141static struct acpi_cpu_softc **cpu_softc;
142ACPI_SERIAL_DECL(cpu, "ACPI CPU");
143
144static int	acpi_cpu_probe(device_t dev);
145static int	acpi_cpu_attach(device_t dev);
146static int	acpi_cpu_suspend(device_t dev);
147static int	acpi_cpu_resume(device_t dev);
148static int	acpi_pcpu_get_id(uint32_t idx, uint32_t *acpi_id,
149		    uint32_t *cpu_id);
150static struct resource_list *acpi_cpu_get_rlist(device_t dev, device_t child);
151static device_t	acpi_cpu_add_child(device_t dev, int order, const char *name,
152		    int unit);
153static int	acpi_cpu_read_ivar(device_t dev, device_t child, int index,
154		    uintptr_t *result);
155static int	acpi_cpu_shutdown(device_t dev);
156static void	acpi_cpu_cx_probe(struct acpi_cpu_softc *sc);
157static void	acpi_cpu_generic_cx_probe(struct acpi_cpu_softc *sc);
158static int	acpi_cpu_cx_cst(struct acpi_cpu_softc *sc);
159static void	acpi_cpu_startup(void *arg);
160static void	acpi_cpu_startup_cx(struct acpi_cpu_softc *sc);
161static void	acpi_cpu_cx_list(struct acpi_cpu_softc *sc);
162static void	acpi_cpu_idle(void);
163static void	acpi_cpu_notify(ACPI_HANDLE h, UINT32 notify, void *context);
164static int	acpi_cpu_quirks(void);
165static int	acpi_cpu_usage_sysctl(SYSCTL_HANDLER_ARGS);
166static int	acpi_cpu_set_cx_lowest(struct acpi_cpu_softc *sc, int val);
167static int	acpi_cpu_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS);
168static int	acpi_cpu_global_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS);
169
170static device_method_t acpi_cpu_methods[] = {
171    /* Device interface */
172    DEVMETHOD(device_probe,	acpi_cpu_probe),
173    DEVMETHOD(device_attach,	acpi_cpu_attach),
174    DEVMETHOD(device_detach,	bus_generic_detach),
175    DEVMETHOD(device_shutdown,	acpi_cpu_shutdown),
176    DEVMETHOD(device_suspend,	acpi_cpu_suspend),
177    DEVMETHOD(device_resume,	acpi_cpu_resume),
178
179    /* Bus interface */
180    DEVMETHOD(bus_add_child,	acpi_cpu_add_child),
181    DEVMETHOD(bus_read_ivar,	acpi_cpu_read_ivar),
182    DEVMETHOD(bus_get_resource_list, acpi_cpu_get_rlist),
183    DEVMETHOD(bus_get_resource,	bus_generic_rl_get_resource),
184    DEVMETHOD(bus_set_resource,	bus_generic_rl_set_resource),
185    DEVMETHOD(bus_alloc_resource, bus_generic_rl_alloc_resource),
186    DEVMETHOD(bus_release_resource, bus_generic_rl_release_resource),
187    DEVMETHOD(bus_driver_added,	bus_generic_driver_added),
188    DEVMETHOD(bus_activate_resource, bus_generic_activate_resource),
189    DEVMETHOD(bus_deactivate_resource, bus_generic_deactivate_resource),
190    DEVMETHOD(bus_setup_intr,	bus_generic_setup_intr),
191    DEVMETHOD(bus_teardown_intr, bus_generic_teardown_intr),
192
193    {0, 0}
194};
195
196static driver_t acpi_cpu_driver = {
197    "cpu",
198    acpi_cpu_methods,
199    sizeof(struct acpi_cpu_softc),
200};
201
202static devclass_t acpi_cpu_devclass;
203DRIVER_MODULE(cpu, acpi, acpi_cpu_driver, acpi_cpu_devclass, 0, 0);
204MODULE_DEPEND(cpu, acpi, 1, 1, 1);
205
206static int
207acpi_cpu_probe(device_t dev)
208{
209    int			   acpi_id, cpu_id;
210    ACPI_BUFFER		   buf;
211    ACPI_HANDLE		   handle;
212    ACPI_OBJECT		   *obj;
213    ACPI_STATUS		   status;
214
215    if (acpi_disabled("cpu") || acpi_get_type(dev) != ACPI_TYPE_PROCESSOR)
216	return (ENXIO);
217
218    handle = acpi_get_handle(dev);
219    if (cpu_softc == NULL)
220	cpu_softc = malloc(sizeof(struct acpi_cpu_softc *) *
221	    (mp_maxid + 1), M_TEMP /* XXX */, M_WAITOK | M_ZERO);
222
223    /* Get our Processor object. */
224    buf.Pointer = NULL;
225    buf.Length = ACPI_ALLOCATE_BUFFER;
226    status = AcpiEvaluateObject(handle, NULL, NULL, &buf);
227    if (ACPI_FAILURE(status)) {
228	device_printf(dev, "probe failed to get Processor obj - %s\n",
229		      AcpiFormatException(status));
230	return (ENXIO);
231    }
232    obj = (ACPI_OBJECT *)buf.Pointer;
233    if (obj->Type != ACPI_TYPE_PROCESSOR) {
234	device_printf(dev, "Processor object has bad type %d\n", obj->Type);
235	AcpiOsFree(obj);
236	return (ENXIO);
237    }
238
239    /*
240     * Find the processor associated with our unit.  We could use the
241     * ProcId as a key, however, some boxes do not have the same values
242     * in their Processor object as the ProcId values in the MADT.
243     */
244    acpi_id = obj->Processor.ProcId;
245    AcpiOsFree(obj);
246    if (acpi_pcpu_get_id(device_get_unit(dev), &acpi_id, &cpu_id) != 0)
247	return (ENXIO);
248
249    /*
250     * Check if we already probed this processor.  We scan the bus twice
251     * so it's possible we've already seen this one.
252     */
253    if (cpu_softc[cpu_id] != NULL)
254	return (ENXIO);
255
256    /* Mark this processor as in-use and save our derived id for attach. */
257    cpu_softc[cpu_id] = (void *)1;
258    acpi_set_private(dev, (void*)(intptr_t)cpu_id);
259    device_set_desc(dev, "ACPI CPU");
260
261    return (0);
262}
263
264static int
265acpi_cpu_attach(device_t dev)
266{
267    ACPI_BUFFER		   buf;
268    ACPI_OBJECT		   arg[4], *obj;
269    ACPI_OBJECT_LIST	   arglist;
270    struct pcpu		   *pcpu_data;
271    struct acpi_cpu_softc *sc;
272    struct acpi_softc	  *acpi_sc;
273    ACPI_STATUS		   status;
274    u_int		   features;
275    int			   cpu_id, drv_count, i;
276    driver_t 		  **drivers;
277    uint32_t		   cap_set[3];
278
279    /* UUID needed by _OSC evaluation */
280    static uint8_t cpu_oscuuid[16] = { 0x16, 0xA6, 0x77, 0x40, 0x0C, 0x29,
281				       0xBE, 0x47, 0x9E, 0xBD, 0xD8, 0x70,
282				       0x58, 0x71, 0x39, 0x53 };
283
284    ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
285
286    sc = device_get_softc(dev);
287    sc->cpu_dev = dev;
288    sc->cpu_handle = acpi_get_handle(dev);
289    cpu_id = (int)(intptr_t)acpi_get_private(dev);
290    cpu_softc[cpu_id] = sc;
291    pcpu_data = pcpu_find(cpu_id);
292    pcpu_data->pc_device = dev;
293    sc->cpu_pcpu = pcpu_data;
294    cpu_smi_cmd = AcpiGbl_FADT.SmiCommand;
295    cpu_cst_cnt = AcpiGbl_FADT.CstControl;
296
297    buf.Pointer = NULL;
298    buf.Length = ACPI_ALLOCATE_BUFFER;
299    status = AcpiEvaluateObject(sc->cpu_handle, NULL, NULL, &buf);
300    if (ACPI_FAILURE(status)) {
301	device_printf(dev, "attach failed to get Processor obj - %s\n",
302		      AcpiFormatException(status));
303	return (ENXIO);
304    }
305    obj = (ACPI_OBJECT *)buf.Pointer;
306    sc->cpu_p_blk = obj->Processor.PblkAddress;
307    sc->cpu_p_blk_len = obj->Processor.PblkLength;
308    sc->cpu_acpi_id = obj->Processor.ProcId;
309    AcpiOsFree(obj);
310    ACPI_DEBUG_PRINT((ACPI_DB_INFO, "acpi_cpu%d: P_BLK at %#x/%d\n",
311		     device_get_unit(dev), sc->cpu_p_blk, sc->cpu_p_blk_len));
312
313    /*
314     * If this is the first cpu we attach, create and initialize the generic
315     * resources that will be used by all acpi cpu devices.
316     */
317    if (device_get_unit(dev) == 0) {
318	/* Assume we won't be using generic Cx mode by default */
319	cpu_cx_generic = FALSE;
320
321	/* Install hw.acpi.cpu sysctl tree */
322	acpi_sc = acpi_device_get_parent_softc(dev);
323	sysctl_ctx_init(&cpu_sysctl_ctx);
324	cpu_sysctl_tree = SYSCTL_ADD_NODE(&cpu_sysctl_ctx,
325	    SYSCTL_CHILDREN(acpi_sc->acpi_sysctl_tree), OID_AUTO, "cpu",
326	    CTLFLAG_RD, 0, "node for CPU children");
327
328	/* Queue post cpu-probing task handler */
329	AcpiOsExecute(OSL_NOTIFY_HANDLER, acpi_cpu_startup, NULL);
330    }
331
332    /*
333     * Before calling any CPU methods, collect child driver feature hints
334     * and notify ACPI of them.  We support unified SMP power control
335     * so advertise this ourselves.  Note this is not the same as independent
336     * SMP control where each CPU can have different settings.
337     */
338    sc->cpu_features = ACPI_CAP_SMP_SAME | ACPI_CAP_SMP_SAME_C3;
339    if (devclass_get_drivers(acpi_cpu_devclass, &drivers, &drv_count) == 0) {
340	for (i = 0; i < drv_count; i++) {
341	    if (ACPI_GET_FEATURES(drivers[i], &features) == 0)
342		sc->cpu_features |= features;
343	}
344	free(drivers, M_TEMP);
345    }
346
347    /*
348     * CPU capabilities are specified in
349     * Intel Processor Vendor-Specific ACPI Interface Specification.
350     */
351    if (sc->cpu_features) {
352	arglist.Pointer = arg;
353	arglist.Count = 4;
354	arg[0].Type = ACPI_TYPE_BUFFER;
355	arg[0].Buffer.Length = sizeof(cpu_oscuuid);
356	arg[0].Buffer.Pointer = cpu_oscuuid;	/* UUID */
357	arg[1].Type = ACPI_TYPE_INTEGER;
358	arg[1].Integer.Value = 1;		/* revision */
359	arg[2].Type = ACPI_TYPE_INTEGER;
360	arg[2].Integer.Value = 1;		/* count */
361	arg[3].Type = ACPI_TYPE_BUFFER;
362	arg[3].Buffer.Length = sizeof(cap_set);	/* Capabilities buffer */
363	arg[3].Buffer.Pointer = (uint8_t *)cap_set;
364	cap_set[0] = 0;				/* status */
365	cap_set[1] = sc->cpu_features;
366	status = AcpiEvaluateObject(sc->cpu_handle, "_OSC", &arglist, NULL);
367	if (ACPI_SUCCESS(status)) {
368	    if (cap_set[0] != 0)
369		device_printf(dev, "_OSC returned status %#x\n", cap_set[0]);
370	}
371	else {
372	    arglist.Pointer = arg;
373	    arglist.Count = 1;
374	    arg[0].Type = ACPI_TYPE_BUFFER;
375	    arg[0].Buffer.Length = sizeof(cap_set);
376	    arg[0].Buffer.Pointer = (uint8_t *)cap_set;
377	    cap_set[0] = 1; /* revision */
378	    cap_set[1] = 1; /* number of capabilities integers */
379	    cap_set[2] = sc->cpu_features;
380	    AcpiEvaluateObject(sc->cpu_handle, "_PDC", &arglist, NULL);
381	}
382    }
383
384    /* Probe for Cx state support. */
385    acpi_cpu_cx_probe(sc);
386
387    return (0);
388}
389
390static void
391acpi_cpu_postattach(void *unused __unused)
392{
393    device_t *devices;
394    int err;
395    int i, n;
396
397    err = devclass_get_devices(acpi_cpu_devclass, &devices, &n);
398    if (err != 0) {
399	printf("devclass_get_devices(acpi_cpu_devclass) failed\n");
400	return;
401    }
402    for (i = 0; i < n; i++)
403	bus_generic_probe(devices[i]);
404    for (i = 0; i < n; i++)
405	bus_generic_attach(devices[i]);
406    free(devices, M_TEMP);
407}
408
409SYSINIT(acpi_cpu, SI_SUB_CONFIGURE, SI_ORDER_MIDDLE,
410    acpi_cpu_postattach, NULL);
411
412/*
413 * Disable any entry to the idle function during suspend and re-enable it
414 * during resume.
415 */
416static int
417acpi_cpu_suspend(device_t dev)
418{
419    int error;
420
421    error = bus_generic_suspend(dev);
422    if (error)
423	return (error);
424    cpu_disable_idle = TRUE;
425    return (0);
426}
427
428static int
429acpi_cpu_resume(device_t dev)
430{
431
432    cpu_disable_idle = FALSE;
433    return (bus_generic_resume(dev));
434}
435
436/*
437 * Find the nth present CPU and return its pc_cpuid as well as set the
438 * pc_acpi_id from the most reliable source.
439 */
440static int
441acpi_pcpu_get_id(uint32_t idx, uint32_t *acpi_id, uint32_t *cpu_id)
442{
443    struct pcpu	*pcpu_data;
444    uint32_t	 i;
445
446    KASSERT(acpi_id != NULL, ("Null acpi_id"));
447    KASSERT(cpu_id != NULL, ("Null cpu_id"));
448    CPU_FOREACH(i) {
449	pcpu_data = pcpu_find(i);
450	KASSERT(pcpu_data != NULL, ("no pcpu data for %d", i));
451	if (idx-- == 0) {
452	    /*
453	     * If pc_acpi_id was not initialized (e.g., a non-APIC UP box)
454	     * override it with the value from the ASL.  Otherwise, if the
455	     * two don't match, prefer the MADT-derived value.  Finally,
456	     * return the pc_cpuid to reference this processor.
457	     */
458	    if (pcpu_data->pc_acpi_id == 0xffffffff)
459		pcpu_data->pc_acpi_id = *acpi_id;
460	    else if (pcpu_data->pc_acpi_id != *acpi_id)
461		*acpi_id = pcpu_data->pc_acpi_id;
462	    *cpu_id = pcpu_data->pc_cpuid;
463	    return (0);
464	}
465    }
466
467    return (ESRCH);
468}
469
470static struct resource_list *
471acpi_cpu_get_rlist(device_t dev, device_t child)
472{
473    struct acpi_cpu_device *ad;
474
475    ad = device_get_ivars(child);
476    if (ad == NULL)
477	return (NULL);
478    return (&ad->ad_rl);
479}
480
481static device_t
482acpi_cpu_add_child(device_t dev, int order, const char *name, int unit)
483{
484    struct acpi_cpu_device *ad;
485    device_t child;
486
487    if ((ad = malloc(sizeof(*ad), M_TEMP, M_NOWAIT | M_ZERO)) == NULL)
488	return (NULL);
489
490    resource_list_init(&ad->ad_rl);
491
492    child = device_add_child_ordered(dev, order, name, unit);
493    if (child != NULL)
494	device_set_ivars(child, ad);
495    else
496	free(ad, M_TEMP);
497    return (child);
498}
499
500static int
501acpi_cpu_read_ivar(device_t dev, device_t child, int index, uintptr_t *result)
502{
503    struct acpi_cpu_softc *sc;
504
505    sc = device_get_softc(dev);
506    switch (index) {
507    case ACPI_IVAR_HANDLE:
508	*result = (uintptr_t)sc->cpu_handle;
509	break;
510    case CPU_IVAR_PCPU:
511	*result = (uintptr_t)sc->cpu_pcpu;
512	break;
513    default:
514	return (ENOENT);
515    }
516    return (0);
517}
518
519static int
520acpi_cpu_shutdown(device_t dev)
521{
522    ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
523
524    /* Allow children to shutdown first. */
525    bus_generic_shutdown(dev);
526
527    /*
528     * Disable any entry to the idle function.  There is a small race where
529     * an idle thread have passed this check but not gone to sleep.  This
530     * is ok since device_shutdown() does not free the softc, otherwise
531     * we'd have to be sure all threads were evicted before returning.
532     */
533    cpu_disable_idle = TRUE;
534
535    return_VALUE (0);
536}
537
538static void
539acpi_cpu_cx_probe(struct acpi_cpu_softc *sc)
540{
541    ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
542
543    /* Use initial sleep value of 1 sec. to start with lowest idle state. */
544    sc->cpu_prev_sleep = 1000000;
545    sc->cpu_cx_lowest = 0;
546
547    /*
548     * Check for the ACPI 2.0 _CST sleep states object. If we can't find
549     * any, we'll revert to generic FADT/P_BLK Cx control method which will
550     * be handled by acpi_cpu_startup. We need to defer to after having
551     * probed all the cpus in the system before probing for generic Cx
552     * states as we may already have found cpus with valid _CST packages
553     */
554    if (!cpu_cx_generic && acpi_cpu_cx_cst(sc) != 0) {
555	/*
556	 * We were unable to find a _CST package for this cpu or there
557	 * was an error parsing it. Switch back to generic mode.
558	 */
559	cpu_cx_generic = TRUE;
560	if (bootverbose)
561	    device_printf(sc->cpu_dev, "switching to generic Cx mode\n");
562    }
563
564    /*
565     * TODO: _CSD Package should be checked here.
566     */
567}
568
569static void
570acpi_cpu_generic_cx_probe(struct acpi_cpu_softc *sc)
571{
572    ACPI_GENERIC_ADDRESS	 gas;
573    struct acpi_cx		*cx_ptr;
574
575    sc->cpu_cx_count = 0;
576    cx_ptr = sc->cpu_cx_states;
577
578    /* Use initial sleep value of 1 sec. to start with lowest idle state. */
579    sc->cpu_prev_sleep = 1000000;
580
581    /* C1 has been required since just after ACPI 1.0 */
582    cx_ptr->type = ACPI_STATE_C1;
583    cx_ptr->trans_lat = 0;
584    cx_ptr++;
585    sc->cpu_cx_count++;
586
587    /*
588     * The spec says P_BLK must be 6 bytes long.  However, some systems
589     * use it to indicate a fractional set of features present so we
590     * take 5 as C2.  Some may also have a value of 7 to indicate
591     * another C3 but most use _CST for this (as required) and having
592     * "only" C1-C3 is not a hardship.
593     */
594    if (sc->cpu_p_blk_len < 5)
595	return;
596
597    /* Validate and allocate resources for C2 (P_LVL2). */
598    gas.SpaceId = ACPI_ADR_SPACE_SYSTEM_IO;
599    gas.BitWidth = 8;
600    if (AcpiGbl_FADT.C2Latency <= 100) {
601	gas.Address = sc->cpu_p_blk + 4;
602	acpi_bus_alloc_gas(sc->cpu_dev, &cx_ptr->res_type, &sc->cpu_rid,
603	    &gas, &cx_ptr->p_lvlx, RF_SHAREABLE);
604	if (cx_ptr->p_lvlx != NULL) {
605	    sc->cpu_rid++;
606	    cx_ptr->type = ACPI_STATE_C2;
607	    cx_ptr->trans_lat = AcpiGbl_FADT.C2Latency;
608	    cx_ptr++;
609	    sc->cpu_cx_count++;
610	}
611    }
612    if (sc->cpu_p_blk_len < 6)
613	return;
614
615    /* Validate and allocate resources for C3 (P_LVL3). */
616    if (AcpiGbl_FADT.C3Latency <= 1000 && !(cpu_quirks & CPU_QUIRK_NO_C3)) {
617	gas.Address = sc->cpu_p_blk + 5;
618	acpi_bus_alloc_gas(sc->cpu_dev, &cx_ptr->res_type, &sc->cpu_rid, &gas,
619	    &cx_ptr->p_lvlx, RF_SHAREABLE);
620	if (cx_ptr->p_lvlx != NULL) {
621	    sc->cpu_rid++;
622	    cx_ptr->type = ACPI_STATE_C3;
623	    cx_ptr->trans_lat = AcpiGbl_FADT.C3Latency;
624	    cx_ptr++;
625	    sc->cpu_cx_count++;
626	}
627    }
628}
629
630/*
631 * Parse a _CST package and set up its Cx states.  Since the _CST object
632 * can change dynamically, our notify handler may call this function
633 * to clean up and probe the new _CST package.
634 */
635static int
636acpi_cpu_cx_cst(struct acpi_cpu_softc *sc)
637{
638    struct	 acpi_cx *cx_ptr;
639    ACPI_STATUS	 status;
640    ACPI_BUFFER	 buf;
641    ACPI_OBJECT	*top;
642    ACPI_OBJECT	*pkg;
643    uint32_t	 count;
644    int		 i;
645
646    ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
647
648    buf.Pointer = NULL;
649    buf.Length = ACPI_ALLOCATE_BUFFER;
650    status = AcpiEvaluateObject(sc->cpu_handle, "_CST", NULL, &buf);
651    if (ACPI_FAILURE(status))
652	return (ENXIO);
653
654    /* _CST is a package with a count and at least one Cx package. */
655    top = (ACPI_OBJECT *)buf.Pointer;
656    if (!ACPI_PKG_VALID(top, 2) || acpi_PkgInt32(top, 0, &count) != 0) {
657	device_printf(sc->cpu_dev, "invalid _CST package\n");
658	AcpiOsFree(buf.Pointer);
659	return (ENXIO);
660    }
661    if (count != top->Package.Count - 1) {
662	device_printf(sc->cpu_dev, "invalid _CST state count (%d != %d)\n",
663	       count, top->Package.Count - 1);
664	count = top->Package.Count - 1;
665    }
666    if (count > MAX_CX_STATES) {
667	device_printf(sc->cpu_dev, "_CST has too many states (%d)\n", count);
668	count = MAX_CX_STATES;
669    }
670
671    /* Set up all valid states. */
672    sc->cpu_cx_count = 0;
673    cx_ptr = sc->cpu_cx_states;
674    for (i = 0; i < count; i++) {
675	pkg = &top->Package.Elements[i + 1];
676	if (!ACPI_PKG_VALID(pkg, 4) ||
677	    acpi_PkgInt32(pkg, 1, &cx_ptr->type) != 0 ||
678	    acpi_PkgInt32(pkg, 2, &cx_ptr->trans_lat) != 0 ||
679	    acpi_PkgInt32(pkg, 3, &cx_ptr->power) != 0) {
680
681	    device_printf(sc->cpu_dev, "skipping invalid Cx state package\n");
682	    continue;
683	}
684
685	/* Validate the state to see if we should use it. */
686	switch (cx_ptr->type) {
687	case ACPI_STATE_C1:
688	    sc->cpu_non_c3 = i;
689	    cx_ptr++;
690	    sc->cpu_cx_count++;
691	    continue;
692	case ACPI_STATE_C2:
693	    if (cx_ptr->trans_lat > 100) {
694		ACPI_DEBUG_PRINT((ACPI_DB_INFO,
695				 "acpi_cpu%d: C2[%d] not available.\n",
696				 device_get_unit(sc->cpu_dev), i));
697		continue;
698	    }
699	    sc->cpu_non_c3 = i;
700	    break;
701	case ACPI_STATE_C3:
702	default:
703	    if (cx_ptr->trans_lat > 1000 ||
704		(cpu_quirks & CPU_QUIRK_NO_C3) != 0) {
705
706		ACPI_DEBUG_PRINT((ACPI_DB_INFO,
707				 "acpi_cpu%d: C3[%d] not available.\n",
708				 device_get_unit(sc->cpu_dev), i));
709		continue;
710	    }
711	    break;
712	}
713
714#ifdef notyet
715	/* Free up any previous register. */
716	if (cx_ptr->p_lvlx != NULL) {
717	    bus_release_resource(sc->cpu_dev, 0, 0, cx_ptr->p_lvlx);
718	    cx_ptr->p_lvlx = NULL;
719	}
720#endif
721
722	/* Allocate the control register for C2 or C3. */
723	acpi_PkgGas(sc->cpu_dev, pkg, 0, &cx_ptr->res_type, &sc->cpu_rid,
724	    &cx_ptr->p_lvlx, RF_SHAREABLE);
725	if (cx_ptr->p_lvlx) {
726	    sc->cpu_rid++;
727	    ACPI_DEBUG_PRINT((ACPI_DB_INFO,
728			     "acpi_cpu%d: Got C%d - %d latency\n",
729			     device_get_unit(sc->cpu_dev), cx_ptr->type,
730			     cx_ptr->trans_lat));
731	    cx_ptr++;
732	    sc->cpu_cx_count++;
733	}
734    }
735    AcpiOsFree(buf.Pointer);
736
737    return (0);
738}
739
740/*
741 * Call this *after* all CPUs have been attached.
742 */
743static void
744acpi_cpu_startup(void *arg)
745{
746    struct acpi_cpu_softc *sc;
747    int i;
748
749    /* Get set of CPU devices */
750    devclass_get_devices(acpi_cpu_devclass, &cpu_devices, &cpu_ndevices);
751
752    /*
753     * Setup any quirks that might necessary now that we have probed
754     * all the CPUs
755     */
756    acpi_cpu_quirks();
757
758    cpu_cx_count = 0;
759    if (cpu_cx_generic) {
760	/*
761	 * We are using generic Cx mode, probe for available Cx states
762	 * for all processors.
763	 */
764	for (i = 0; i < cpu_ndevices; i++) {
765	    sc = device_get_softc(cpu_devices[i]);
766	    acpi_cpu_generic_cx_probe(sc);
767	    if (sc->cpu_cx_count > cpu_cx_count)
768		    cpu_cx_count = sc->cpu_cx_count;
769	}
770
771	/*
772	 * Find the highest Cx state common to all CPUs
773	 * in the system, taking quirks into account.
774	 */
775	for (i = 0; i < cpu_ndevices; i++) {
776	    sc = device_get_softc(cpu_devices[i]);
777	    if (sc->cpu_cx_count < cpu_cx_count)
778		cpu_cx_count = sc->cpu_cx_count;
779	}
780    } else {
781	/*
782	 * We are using _CST mode, remove C3 state if necessary.
783	 * Update the largest Cx state supported in the global cpu_cx_count.
784	 * It will be used in the global Cx sysctl handler.
785	 * As we now know for sure that we will be using _CST mode
786	 * install our notify handler.
787	 */
788	for (i = 0; i < cpu_ndevices; i++) {
789	    sc = device_get_softc(cpu_devices[i]);
790	    if (cpu_quirks & CPU_QUIRK_NO_C3) {
791		sc->cpu_cx_count = sc->cpu_non_c3 + 1;
792	    }
793	    if (sc->cpu_cx_count > cpu_cx_count)
794		cpu_cx_count = sc->cpu_cx_count;
795	    AcpiInstallNotifyHandler(sc->cpu_handle, ACPI_DEVICE_NOTIFY,
796		acpi_cpu_notify, sc);
797	}
798    }
799
800    /* Perform Cx final initialization. */
801    for (i = 0; i < cpu_ndevices; i++) {
802	sc = device_get_softc(cpu_devices[i]);
803	acpi_cpu_startup_cx(sc);
804    }
805
806    /* Add a sysctl handler to handle global Cx lowest setting */
807    SYSCTL_ADD_PROC(&cpu_sysctl_ctx, SYSCTL_CHILDREN(cpu_sysctl_tree),
808	OID_AUTO, "cx_lowest", CTLTYPE_STRING | CTLFLAG_RW,
809	NULL, 0, acpi_cpu_global_cx_lowest_sysctl, "A",
810	"Global lowest Cx sleep state to use");
811
812    /* Take over idling from cpu_idle_default(). */
813    cpu_cx_lowest = 0;
814    cpu_disable_idle = FALSE;
815    cpu_idle_hook = acpi_cpu_idle;
816}
817
818static void
819acpi_cpu_cx_list(struct acpi_cpu_softc *sc)
820{
821    struct sbuf sb;
822    int i;
823
824    /*
825     * Set up the list of Cx states
826     */
827    sc->cpu_non_c3 = 0;
828    sbuf_new(&sb, sc->cpu_cx_supported, sizeof(sc->cpu_cx_supported),
829	SBUF_FIXEDLEN);
830    for (i = 0; i < sc->cpu_cx_count; i++) {
831	sbuf_printf(&sb, "C%d/%d ", i + 1, sc->cpu_cx_states[i].trans_lat);
832	if (sc->cpu_cx_states[i].type < ACPI_STATE_C3)
833	    sc->cpu_non_c3 = i;
834    }
835    sbuf_trim(&sb);
836    sbuf_finish(&sb);
837}
838
839static void
840acpi_cpu_startup_cx(struct acpi_cpu_softc *sc)
841{
842    acpi_cpu_cx_list(sc);
843
844    SYSCTL_ADD_STRING(&sc->cpu_sysctl_ctx,
845		      SYSCTL_CHILDREN(device_get_sysctl_tree(sc->cpu_dev)),
846		      OID_AUTO, "cx_supported", CTLFLAG_RD,
847		      sc->cpu_cx_supported, 0,
848		      "Cx/microsecond values for supported Cx states");
849    SYSCTL_ADD_PROC(&sc->cpu_sysctl_ctx,
850		    SYSCTL_CHILDREN(device_get_sysctl_tree(sc->cpu_dev)),
851		    OID_AUTO, "cx_lowest", CTLTYPE_STRING | CTLFLAG_RW,
852		    (void *)sc, 0, acpi_cpu_cx_lowest_sysctl, "A",
853		    "lowest Cx sleep state to use");
854    SYSCTL_ADD_PROC(&sc->cpu_sysctl_ctx,
855		    SYSCTL_CHILDREN(device_get_sysctl_tree(sc->cpu_dev)),
856		    OID_AUTO, "cx_usage", CTLTYPE_STRING | CTLFLAG_RD,
857		    (void *)sc, 0, acpi_cpu_usage_sysctl, "A",
858		    "percent usage for each Cx state");
859
860#ifdef notyet
861    /* Signal platform that we can handle _CST notification. */
862    if (!cpu_cx_generic && cpu_cst_cnt != 0) {
863	ACPI_LOCK(acpi);
864	AcpiOsWritePort(cpu_smi_cmd, cpu_cst_cnt, 8);
865	ACPI_UNLOCK(acpi);
866    }
867#endif
868}
869
870/*
871 * Idle the CPU in the lowest state possible.  This function is called with
872 * interrupts disabled.  Note that once it re-enables interrupts, a task
873 * switch can occur so do not access shared data (i.e. the softc) after
874 * interrupts are re-enabled.
875 */
876static void
877acpi_cpu_idle()
878{
879    struct	acpi_cpu_softc *sc;
880    struct	acpi_cx *cx_next;
881    uint32_t	start_time, end_time;
882    int		bm_active, cx_next_idx, i;
883
884    /* If disabled, return immediately. */
885    if (cpu_disable_idle) {
886	ACPI_ENABLE_IRQS();
887	return;
888    }
889
890    /*
891     * Look up our CPU id to get our softc.  If it's NULL, we'll use C1
892     * since there is no ACPI processor object for this CPU.  This occurs
893     * for logical CPUs in the HTT case.
894     */
895    sc = cpu_softc[PCPU_GET(cpuid)];
896    if (sc == NULL) {
897	acpi_cpu_c1();
898	return;
899    }
900
901    /* Find the lowest state that has small enough latency. */
902    cx_next_idx = 0;
903    for (i = sc->cpu_cx_lowest; i >= 0; i--) {
904	if (sc->cpu_cx_states[i].trans_lat * 3 <= sc->cpu_prev_sleep) {
905	    cx_next_idx = i;
906	    break;
907	}
908    }
909
910    /*
911     * Check for bus master activity.  If there was activity, clear
912     * the bit and use the lowest non-C3 state.  Note that the USB
913     * driver polling for new devices keeps this bit set all the
914     * time if USB is loaded.
915     */
916    if ((cpu_quirks & CPU_QUIRK_NO_BM_CTRL) == 0) {
917	AcpiReadBitRegister(ACPI_BITREG_BUS_MASTER_STATUS, &bm_active);
918	if (bm_active != 0) {
919	    AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_STATUS, 1);
920	    cx_next_idx = min(cx_next_idx, sc->cpu_non_c3);
921	}
922    }
923
924    /* Select the next state and update statistics. */
925    cx_next = &sc->cpu_cx_states[cx_next_idx];
926    sc->cpu_cx_stats[cx_next_idx]++;
927    KASSERT(cx_next->type != ACPI_STATE_C0, ("acpi_cpu_idle: C0 sleep"));
928
929    /*
930     * Execute HLT (or equivalent) and wait for an interrupt.  We can't
931     * calculate the time spent in C1 since the place we wake up is an
932     * ISR.  Assume we slept half of quantum and return.
933     */
934    if (cx_next->type == ACPI_STATE_C1) {
935	sc->cpu_prev_sleep = (sc->cpu_prev_sleep * 3 + 500000 / hz) / 4;
936	acpi_cpu_c1();
937	return;
938    }
939
940    /*
941     * For C3, disable bus master arbitration and enable bus master wake
942     * if BM control is available, otherwise flush the CPU cache.
943     */
944    if (cx_next->type == ACPI_STATE_C3) {
945	if ((cpu_quirks & CPU_QUIRK_NO_BM_CTRL) == 0) {
946	    AcpiWriteBitRegister(ACPI_BITREG_ARB_DISABLE, 1);
947	    AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_RLD, 1);
948	} else
949	    ACPI_FLUSH_CPU_CACHE();
950    }
951
952    /*
953     * Read from P_LVLx to enter C2(+), checking time spent asleep.
954     * Use the ACPI timer for measuring sleep time.  Since we need to
955     * get the time very close to the CPU start/stop clock logic, this
956     * is the only reliable time source.
957     */
958    AcpiHwRead(&start_time, &AcpiGbl_FADT.XPmTimerBlock);
959    CPU_GET_REG(cx_next->p_lvlx, 1);
960
961    /*
962     * Read the end time twice.  Since it may take an arbitrary time
963     * to enter the idle state, the first read may be executed before
964     * the processor has stopped.  Doing it again provides enough
965     * margin that we are certain to have a correct value.
966     */
967    AcpiHwRead(&end_time, &AcpiGbl_FADT.XPmTimerBlock);
968    AcpiHwRead(&end_time, &AcpiGbl_FADT.XPmTimerBlock);
969
970    /* Enable bus master arbitration and disable bus master wakeup. */
971    if (cx_next->type == ACPI_STATE_C3 &&
972	(cpu_quirks & CPU_QUIRK_NO_BM_CTRL) == 0) {
973	AcpiWriteBitRegister(ACPI_BITREG_ARB_DISABLE, 0);
974	AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_RLD, 0);
975    }
976    ACPI_ENABLE_IRQS();
977
978    /* Find the actual time asleep in microseconds. */
979    end_time = acpi_TimerDelta(end_time, start_time);
980    sc->cpu_prev_sleep = (sc->cpu_prev_sleep * 3 + PM_USEC(end_time)) / 4;
981}
982
983/*
984 * Re-evaluate the _CST object when we are notified that it changed.
985 *
986 * XXX Re-evaluation disabled until locking is done.
987 */
988static void
989acpi_cpu_notify(ACPI_HANDLE h, UINT32 notify, void *context)
990{
991    struct acpi_cpu_softc *sc = (struct acpi_cpu_softc *)context;
992    struct acpi_cpu_softc *isc;
993    int i;
994
995    if (notify != ACPI_NOTIFY_CX_STATES)
996	return;
997
998    /* Update the list of Cx states. */
999    acpi_cpu_cx_cst(sc);
1000    acpi_cpu_cx_list(sc);
1001
1002    /* Update the new lowest useable Cx state for all CPUs. */
1003    ACPI_SERIAL_BEGIN(cpu);
1004    cpu_cx_count = 0;
1005    for (i = 0; i < cpu_ndevices; i++) {
1006	isc = device_get_softc(cpu_devices[i]);
1007	if (isc->cpu_cx_count > cpu_cx_count)
1008	    cpu_cx_count = isc->cpu_cx_count;
1009    }
1010    ACPI_SERIAL_END(cpu);
1011}
1012
1013static int
1014acpi_cpu_quirks(void)
1015{
1016    device_t acpi_dev;
1017    uint32_t val;
1018
1019    ACPI_FUNCTION_TRACE((char *)(uintptr_t)__func__);
1020
1021    /*
1022     * Bus mastering arbitration control is needed to keep caches coherent
1023     * while sleeping in C3.  If it's not present but a working flush cache
1024     * instruction is present, flush the caches before entering C3 instead.
1025     * Otherwise, just disable C3 completely.
1026     */
1027    if (AcpiGbl_FADT.Pm2ControlBlock == 0 ||
1028	AcpiGbl_FADT.Pm2ControlLength == 0) {
1029	if ((AcpiGbl_FADT.Flags & ACPI_FADT_WBINVD) &&
1030	    (AcpiGbl_FADT.Flags & ACPI_FADT_WBINVD_FLUSH) == 0) {
1031	    cpu_quirks |= CPU_QUIRK_NO_BM_CTRL;
1032	    ACPI_DEBUG_PRINT((ACPI_DB_INFO,
1033		"acpi_cpu: no BM control, using flush cache method\n"));
1034	} else {
1035	    cpu_quirks |= CPU_QUIRK_NO_C3;
1036	    ACPI_DEBUG_PRINT((ACPI_DB_INFO,
1037		"acpi_cpu: no BM control, C3 not available\n"));
1038	}
1039    }
1040
1041    /*
1042     * If we are using generic Cx mode, C3 on multiple CPUs requires using
1043     * the expensive flush cache instruction.
1044     */
1045    if (cpu_cx_generic && mp_ncpus > 1) {
1046	cpu_quirks |= CPU_QUIRK_NO_BM_CTRL;
1047	ACPI_DEBUG_PRINT((ACPI_DB_INFO,
1048	    "acpi_cpu: SMP, using flush cache mode for C3\n"));
1049    }
1050
1051    /* Look for various quirks of the PIIX4 part. */
1052    acpi_dev = pci_find_device(PCI_VENDOR_INTEL, PCI_DEVICE_82371AB_3);
1053    if (acpi_dev != NULL) {
1054	switch (pci_get_revid(acpi_dev)) {
1055	/*
1056	 * Disable C3 support for all PIIX4 chipsets.  Some of these parts
1057	 * do not report the BMIDE status to the BM status register and
1058	 * others have a livelock bug if Type-F DMA is enabled.  Linux
1059	 * works around the BMIDE bug by reading the BM status directly
1060	 * but we take the simpler approach of disabling C3 for these
1061	 * parts.
1062	 *
1063	 * See erratum #18 ("C3 Power State/BMIDE and Type-F DMA
1064	 * Livelock") from the January 2002 PIIX4 specification update.
1065	 * Applies to all PIIX4 models.
1066	 *
1067	 * Also, make sure that all interrupts cause a "Stop Break"
1068	 * event to exit from C2 state.
1069	 * Also, BRLD_EN_BM (ACPI_BITREG_BUS_MASTER_RLD in ACPI-speak)
1070	 * should be set to zero, otherwise it causes C2 to short-sleep.
1071	 * PIIX4 doesn't properly support C3 and bus master activity
1072	 * need not break out of C2.
1073	 */
1074	case PCI_REVISION_A_STEP:
1075	case PCI_REVISION_B_STEP:
1076	case PCI_REVISION_4E:
1077	case PCI_REVISION_4M:
1078	    cpu_quirks |= CPU_QUIRK_NO_C3;
1079	    ACPI_DEBUG_PRINT((ACPI_DB_INFO,
1080		"acpi_cpu: working around PIIX4 bug, disabling C3\n"));
1081
1082	    val = pci_read_config(acpi_dev, PIIX4_DEVACTB_REG, 4);
1083	    if ((val & PIIX4_STOP_BREAK_MASK) != PIIX4_STOP_BREAK_MASK) {
1084		ACPI_DEBUG_PRINT((ACPI_DB_INFO,
1085		    "acpi_cpu: PIIX4: enabling IRQs to generate Stop Break\n"));
1086	    	val |= PIIX4_STOP_BREAK_MASK;
1087		pci_write_config(acpi_dev, PIIX4_DEVACTB_REG, val, 4);
1088	    }
1089	    AcpiReadBitRegister(ACPI_BITREG_BUS_MASTER_RLD, &val);
1090	    if (val) {
1091		ACPI_DEBUG_PRINT((ACPI_DB_INFO,
1092		    "acpi_cpu: PIIX4: reset BRLD_EN_BM\n"));
1093		AcpiWriteBitRegister(ACPI_BITREG_BUS_MASTER_RLD, 0);
1094	    }
1095	    break;
1096	default:
1097	    break;
1098	}
1099    }
1100
1101    return (0);
1102}
1103
1104static int
1105acpi_cpu_usage_sysctl(SYSCTL_HANDLER_ARGS)
1106{
1107    struct acpi_cpu_softc *sc;
1108    struct sbuf	 sb;
1109    char	 buf[128];
1110    int		 i;
1111    uintmax_t	 fract, sum, whole;
1112
1113    sc = (struct acpi_cpu_softc *) arg1;
1114    sum = 0;
1115    for (i = 0; i < sc->cpu_cx_count; i++)
1116	sum += sc->cpu_cx_stats[i];
1117    sbuf_new(&sb, buf, sizeof(buf), SBUF_FIXEDLEN);
1118    for (i = 0; i < sc->cpu_cx_count; i++) {
1119	if (sum > 0) {
1120	    whole = (uintmax_t)sc->cpu_cx_stats[i] * 100;
1121	    fract = (whole % sum) * 100;
1122	    sbuf_printf(&sb, "%u.%02u%% ", (u_int)(whole / sum),
1123		(u_int)(fract / sum));
1124	} else
1125	    sbuf_printf(&sb, "0.00%% ");
1126    }
1127    sbuf_printf(&sb, "last %dus", sc->cpu_prev_sleep);
1128    sbuf_trim(&sb);
1129    sbuf_finish(&sb);
1130    sysctl_handle_string(oidp, sbuf_data(&sb), sbuf_len(&sb), req);
1131    sbuf_delete(&sb);
1132
1133    return (0);
1134}
1135
1136static int
1137acpi_cpu_set_cx_lowest(struct acpi_cpu_softc *sc, int val)
1138{
1139    int i;
1140
1141    ACPI_SERIAL_ASSERT(cpu);
1142    sc->cpu_cx_lowest = val;
1143
1144    /* If not disabling, cache the new lowest non-C3 state. */
1145    sc->cpu_non_c3 = 0;
1146    for (i = sc->cpu_cx_lowest; i >= 0; i--) {
1147	if (sc->cpu_cx_states[i].type < ACPI_STATE_C3) {
1148	    sc->cpu_non_c3 = i;
1149	    break;
1150	}
1151    }
1152
1153    /* Reset the statistics counters. */
1154    bzero(sc->cpu_cx_stats, sizeof(sc->cpu_cx_stats));
1155    return (0);
1156}
1157
1158static int
1159acpi_cpu_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS)
1160{
1161    struct	 acpi_cpu_softc *sc;
1162    char	 state[8];
1163    int		 val, error;
1164
1165    sc = (struct acpi_cpu_softc *) arg1;
1166    snprintf(state, sizeof(state), "C%d", sc->cpu_cx_lowest + 1);
1167    error = sysctl_handle_string(oidp, state, sizeof(state), req);
1168    if (error != 0 || req->newptr == NULL)
1169	return (error);
1170    if (strlen(state) < 2 || toupper(state[0]) != 'C')
1171	return (EINVAL);
1172    val = (int) strtol(state + 1, NULL, 10) - 1;
1173    if (val < 0 || val > sc->cpu_cx_count - 1)
1174	return (EINVAL);
1175
1176    ACPI_SERIAL_BEGIN(cpu);
1177    acpi_cpu_set_cx_lowest(sc, val);
1178    ACPI_SERIAL_END(cpu);
1179
1180    return (0);
1181}
1182
1183static int
1184acpi_cpu_global_cx_lowest_sysctl(SYSCTL_HANDLER_ARGS)
1185{
1186    struct	acpi_cpu_softc *sc;
1187    char	state[8];
1188    int		val, error, i;
1189
1190    snprintf(state, sizeof(state), "C%d", cpu_cx_lowest + 1);
1191    error = sysctl_handle_string(oidp, state, sizeof(state), req);
1192    if (error != 0 || req->newptr == NULL)
1193	return (error);
1194    if (strlen(state) < 2 || toupper(state[0]) != 'C')
1195	return (EINVAL);
1196    val = (int) strtol(state + 1, NULL, 10) - 1;
1197    if (val < 0 || val > cpu_cx_count - 1)
1198	return (EINVAL);
1199    cpu_cx_lowest = val;
1200
1201    /* Update the new lowest useable Cx state for all CPUs. */
1202    ACPI_SERIAL_BEGIN(cpu);
1203    for (i = 0; i < cpu_ndevices; i++) {
1204	sc = device_get_softc(cpu_devices[i]);
1205	acpi_cpu_set_cx_lowest(sc, val);
1206    }
1207    ACPI_SERIAL_END(cpu);
1208
1209    return (0);
1210}
1211