libzfs_status.c revision 252766
1/*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21
22/*
23 * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved.
24 * Copyright (c) 2012 by Delphix. All rights reserved.
25 * Copyright (c) 2013 Steven Hartland. All rights reserved.
26 */
27
28/*
29 * This file contains the functions which analyze the status of a pool.  This
30 * include both the status of an active pool, as well as the status exported
31 * pools.  Returns one of the ZPOOL_STATUS_* defines describing the status of
32 * the pool.  This status is independent (to a certain degree) from the state of
33 * the pool.  A pool's state describes only whether or not it is capable of
34 * providing the necessary fault tolerance for data.  The status describes the
35 * overall status of devices.  A pool that is online can still have a device
36 * that is experiencing errors.
37 *
38 * Only a subset of the possible faults can be detected using 'zpool status',
39 * and not all possible errors correspond to a FMA message ID.  The explanation
40 * is left up to the caller, depending on whether it is a live pool or an
41 * import.
42 */
43
44#include <libzfs.h>
45#include <string.h>
46#include <unistd.h>
47#include "libzfs_impl.h"
48#include "zfeature_common.h"
49
50/*
51 * Message ID table.  This must be kept in sync with the ZPOOL_STATUS_* defines
52 * in libzfs.h.  Note that there are some status results which go past the end
53 * of this table, and hence have no associated message ID.
54 */
55static char *zfs_msgid_table[] = {
56	"ZFS-8000-14",
57	"ZFS-8000-2Q",
58	"ZFS-8000-3C",
59	"ZFS-8000-4J",
60	"ZFS-8000-5E",
61	"ZFS-8000-6X",
62	"ZFS-8000-72",
63	"ZFS-8000-8A",
64	"ZFS-8000-9P",
65	"ZFS-8000-A5",
66	"ZFS-8000-EY",
67	"ZFS-8000-HC",
68	"ZFS-8000-JQ",
69	"ZFS-8000-K4",
70};
71
72#define	NMSGID	(sizeof (zfs_msgid_table) / sizeof (zfs_msgid_table[0]))
73
74/* ARGSUSED */
75static int
76vdev_missing(uint64_t state, uint64_t aux, uint64_t errs)
77{
78	return (state == VDEV_STATE_CANT_OPEN &&
79	    aux == VDEV_AUX_OPEN_FAILED);
80}
81
82/* ARGSUSED */
83static int
84vdev_faulted(uint64_t state, uint64_t aux, uint64_t errs)
85{
86	return (state == VDEV_STATE_FAULTED);
87}
88
89/* ARGSUSED */
90static int
91vdev_errors(uint64_t state, uint64_t aux, uint64_t errs)
92{
93	return (state == VDEV_STATE_DEGRADED || errs != 0);
94}
95
96/* ARGSUSED */
97static int
98vdev_broken(uint64_t state, uint64_t aux, uint64_t errs)
99{
100	return (state == VDEV_STATE_CANT_OPEN);
101}
102
103/* ARGSUSED */
104static int
105vdev_offlined(uint64_t state, uint64_t aux, uint64_t errs)
106{
107	return (state == VDEV_STATE_OFFLINE);
108}
109
110/* ARGSUSED */
111static int
112vdev_removed(uint64_t state, uint64_t aux, uint64_t errs)
113{
114	return (state == VDEV_STATE_REMOVED);
115}
116
117/*
118 * Detect if any leaf devices that have seen errors or could not be opened.
119 */
120static boolean_t
121find_vdev_problem(nvlist_t *vdev, int (*func)(uint64_t, uint64_t, uint64_t))
122{
123	nvlist_t **child;
124	vdev_stat_t *vs;
125	uint_t c, children;
126	char *type;
127
128	/*
129	 * Ignore problems within a 'replacing' vdev, since we're presumably in
130	 * the process of repairing any such errors, and don't want to call them
131	 * out again.  We'll pick up the fact that a resilver is happening
132	 * later.
133	 */
134	verify(nvlist_lookup_string(vdev, ZPOOL_CONFIG_TYPE, &type) == 0);
135	if (strcmp(type, VDEV_TYPE_REPLACING) == 0)
136		return (B_FALSE);
137
138	if (nvlist_lookup_nvlist_array(vdev, ZPOOL_CONFIG_CHILDREN, &child,
139	    &children) == 0) {
140		for (c = 0; c < children; c++)
141			if (find_vdev_problem(child[c], func))
142				return (B_TRUE);
143	} else {
144		verify(nvlist_lookup_uint64_array(vdev, ZPOOL_CONFIG_VDEV_STATS,
145		    (uint64_t **)&vs, &c) == 0);
146
147		if (func(vs->vs_state, vs->vs_aux,
148		    vs->vs_read_errors +
149		    vs->vs_write_errors +
150		    vs->vs_checksum_errors))
151			return (B_TRUE);
152	}
153
154	/*
155	 * Check any L2 cache devs
156	 */
157	if (nvlist_lookup_nvlist_array(vdev, ZPOOL_CONFIG_L2CACHE, &child,
158	    &children) == 0) {
159		for (c = 0; c < children; c++)
160			if (find_vdev_problem(child[c], func))
161				return (B_TRUE);
162	}
163
164	return (B_FALSE);
165}
166
167/*
168 * Active pool health status.
169 *
170 * To determine the status for a pool, we make several passes over the config,
171 * picking the most egregious error we find.  In order of importance, we do the
172 * following:
173 *
174 *	- Check for a complete and valid configuration
175 *	- Look for any faulted or missing devices in a non-replicated config
176 *	- Check for any data errors
177 *	- Check for any faulted or missing devices in a replicated config
178 *	- Look for any devices showing errors
179 *	- Check for any resilvering devices
180 *
181 * There can obviously be multiple errors within a single pool, so this routine
182 * only picks the most damaging of all the current errors to report.
183 */
184static zpool_status_t
185check_status(nvlist_t *config, boolean_t isimport)
186{
187	nvlist_t *nvroot;
188	vdev_stat_t *vs;
189	pool_scan_stat_t *ps = NULL;
190	uint_t vsc, psc;
191	uint64_t nerr;
192	uint64_t version;
193	uint64_t stateval;
194	uint64_t suspended;
195	uint64_t hostid = 0;
196
197	verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_VERSION,
198	    &version) == 0);
199	verify(nvlist_lookup_nvlist(config, ZPOOL_CONFIG_VDEV_TREE,
200	    &nvroot) == 0);
201	verify(nvlist_lookup_uint64_array(nvroot, ZPOOL_CONFIG_VDEV_STATS,
202	    (uint64_t **)&vs, &vsc) == 0);
203	verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_STATE,
204	    &stateval) == 0);
205
206	/*
207	 * Currently resilvering a vdev
208	 */
209	(void) nvlist_lookup_uint64_array(nvroot, ZPOOL_CONFIG_SCAN_STATS,
210	    (uint64_t **)&ps, &psc);
211	if (ps && ps->pss_func == POOL_SCAN_RESILVER &&
212	    ps->pss_state == DSS_SCANNING)
213		return (ZPOOL_STATUS_RESILVERING);
214
215	/*
216	 * Pool last accessed by another system.
217	 */
218	(void) nvlist_lookup_uint64(config, ZPOOL_CONFIG_HOSTID, &hostid);
219	if (hostid != 0 && (unsigned long)hostid != gethostid() &&
220	    stateval == POOL_STATE_ACTIVE)
221		return (ZPOOL_STATUS_HOSTID_MISMATCH);
222
223	/*
224	 * Newer on-disk version.
225	 */
226	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
227	    vs->vs_aux == VDEV_AUX_VERSION_NEWER)
228		return (ZPOOL_STATUS_VERSION_NEWER);
229
230	/*
231	 * Unsupported feature(s).
232	 */
233	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
234	    vs->vs_aux == VDEV_AUX_UNSUP_FEAT) {
235		nvlist_t *nvinfo;
236
237		verify(nvlist_lookup_nvlist(config, ZPOOL_CONFIG_LOAD_INFO,
238		    &nvinfo) == 0);
239		if (nvlist_exists(nvinfo, ZPOOL_CONFIG_CAN_RDONLY))
240			return (ZPOOL_STATUS_UNSUP_FEAT_WRITE);
241		return (ZPOOL_STATUS_UNSUP_FEAT_READ);
242	}
243
244	/*
245	 * Check that the config is complete.
246	 */
247	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
248	    vs->vs_aux == VDEV_AUX_BAD_GUID_SUM)
249		return (ZPOOL_STATUS_BAD_GUID_SUM);
250
251	/*
252	 * Check whether the pool has suspended due to failed I/O.
253	 */
254	if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_SUSPENDED,
255	    &suspended) == 0) {
256		if (suspended == ZIO_FAILURE_MODE_CONTINUE)
257			return (ZPOOL_STATUS_IO_FAILURE_CONTINUE);
258		return (ZPOOL_STATUS_IO_FAILURE_WAIT);
259	}
260
261	/*
262	 * Could not read a log.
263	 */
264	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
265	    vs->vs_aux == VDEV_AUX_BAD_LOG) {
266		return (ZPOOL_STATUS_BAD_LOG);
267	}
268
269	/*
270	 * Bad devices in non-replicated config.
271	 */
272	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
273	    find_vdev_problem(nvroot, vdev_faulted))
274		return (ZPOOL_STATUS_FAULTED_DEV_NR);
275
276	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
277	    find_vdev_problem(nvroot, vdev_missing))
278		return (ZPOOL_STATUS_MISSING_DEV_NR);
279
280	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
281	    find_vdev_problem(nvroot, vdev_broken))
282		return (ZPOOL_STATUS_CORRUPT_LABEL_NR);
283
284	/*
285	 * Corrupted pool metadata
286	 */
287	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
288	    vs->vs_aux == VDEV_AUX_CORRUPT_DATA)
289		return (ZPOOL_STATUS_CORRUPT_POOL);
290
291	/*
292	 * Persistent data errors.
293	 */
294	if (!isimport) {
295		if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_ERRCOUNT,
296		    &nerr) == 0 && nerr != 0)
297			return (ZPOOL_STATUS_CORRUPT_DATA);
298	}
299
300	/*
301	 * Missing devices in a replicated config.
302	 */
303	if (find_vdev_problem(nvroot, vdev_faulted))
304		return (ZPOOL_STATUS_FAULTED_DEV_R);
305	if (find_vdev_problem(nvroot, vdev_missing))
306		return (ZPOOL_STATUS_MISSING_DEV_R);
307	if (find_vdev_problem(nvroot, vdev_broken))
308		return (ZPOOL_STATUS_CORRUPT_LABEL_R);
309
310	/*
311	 * Devices with errors
312	 */
313	if (!isimport && find_vdev_problem(nvroot, vdev_errors))
314		return (ZPOOL_STATUS_FAILING_DEV);
315
316	/*
317	 * Offlined devices
318	 */
319	if (find_vdev_problem(nvroot, vdev_offlined))
320		return (ZPOOL_STATUS_OFFLINE_DEV);
321
322	/*
323	 * Removed device
324	 */
325	if (find_vdev_problem(nvroot, vdev_removed))
326		return (ZPOOL_STATUS_REMOVED_DEV);
327
328	/*
329	 * Outdated, but usable, version
330	 */
331	if (SPA_VERSION_IS_SUPPORTED(version) && version != SPA_VERSION)
332		return (ZPOOL_STATUS_VERSION_OLDER);
333
334	/*
335	 * Usable pool with disabled features
336	 */
337	if (version >= SPA_VERSION_FEATURES) {
338		int i;
339		nvlist_t *feat;
340
341		if (isimport) {
342			feat = fnvlist_lookup_nvlist(config,
343			    ZPOOL_CONFIG_LOAD_INFO);
344			feat = fnvlist_lookup_nvlist(feat,
345			    ZPOOL_CONFIG_ENABLED_FEAT);
346		} else {
347			feat = fnvlist_lookup_nvlist(config,
348			    ZPOOL_CONFIG_FEATURE_STATS);
349		}
350
351		for (i = 0; i < SPA_FEATURES; i++) {
352			zfeature_info_t *fi = &spa_feature_table[i];
353			if (!nvlist_exists(feat, fi->fi_guid))
354				return (ZPOOL_STATUS_FEAT_DISABLED);
355		}
356	}
357
358	return (ZPOOL_STATUS_OK);
359}
360
361zpool_status_t
362zpool_get_status(zpool_handle_t *zhp, char **msgid)
363{
364	zpool_status_t ret = check_status(zhp->zpool_config, B_FALSE);
365
366	if (ret >= NMSGID)
367		*msgid = NULL;
368	else
369		*msgid = zfs_msgid_table[ret];
370
371	return (ret);
372}
373
374zpool_status_t
375zpool_import_status(nvlist_t *config, char **msgid)
376{
377	zpool_status_t ret = check_status(config, B_TRUE);
378
379	if (ret >= NMSGID)
380		*msgid = NULL;
381	else
382		*msgid = zfs_msgid_table[ret];
383
384	return (ret);
385}
386
387static void
388dump_ddt_stat(const ddt_stat_t *dds, int h)
389{
390	char refcnt[6];
391	char blocks[6], lsize[6], psize[6], dsize[6];
392	char ref_blocks[6], ref_lsize[6], ref_psize[6], ref_dsize[6];
393
394	if (dds == NULL || dds->dds_blocks == 0)
395		return;
396
397	if (h == -1)
398		(void) strcpy(refcnt, "Total");
399	else
400		zfs_nicenum(1ULL << h, refcnt, sizeof (refcnt));
401
402	zfs_nicenum(dds->dds_blocks, blocks, sizeof (blocks));
403	zfs_nicenum(dds->dds_lsize, lsize, sizeof (lsize));
404	zfs_nicenum(dds->dds_psize, psize, sizeof (psize));
405	zfs_nicenum(dds->dds_dsize, dsize, sizeof (dsize));
406	zfs_nicenum(dds->dds_ref_blocks, ref_blocks, sizeof (ref_blocks));
407	zfs_nicenum(dds->dds_ref_lsize, ref_lsize, sizeof (ref_lsize));
408	zfs_nicenum(dds->dds_ref_psize, ref_psize, sizeof (ref_psize));
409	zfs_nicenum(dds->dds_ref_dsize, ref_dsize, sizeof (ref_dsize));
410
411	(void) printf("%6s   %6s   %5s   %5s   %5s   %6s   %5s   %5s   %5s\n",
412	    refcnt,
413	    blocks, lsize, psize, dsize,
414	    ref_blocks, ref_lsize, ref_psize, ref_dsize);
415}
416
417/*
418 * Print the DDT histogram and the column totals.
419 */
420void
421zpool_dump_ddt(const ddt_stat_t *dds_total, const ddt_histogram_t *ddh)
422{
423	int h;
424
425	(void) printf("\n");
426
427	(void) printf("bucket   "
428	    "           allocated             "
429	    "          referenced          \n");
430	(void) printf("______   "
431	    "______________________________   "
432	    "______________________________\n");
433
434	(void) printf("%6s   %6s   %5s   %5s   %5s   %6s   %5s   %5s   %5s\n",
435	    "refcnt",
436	    "blocks", "LSIZE", "PSIZE", "DSIZE",
437	    "blocks", "LSIZE", "PSIZE", "DSIZE");
438
439	(void) printf("%6s   %6s   %5s   %5s   %5s   %6s   %5s   %5s   %5s\n",
440	    "------",
441	    "------", "-----", "-----", "-----",
442	    "------", "-----", "-----", "-----");
443
444	for (h = 0; h < 64; h++)
445		dump_ddt_stat(&ddh->ddh_stat[h], h);
446
447	dump_ddt_stat(dds_total, -1);
448
449	(void) printf("\n");
450}
451