libzfs_status.c revision 243674
1230557Sjimharris/* 2230557Sjimharris * CDDL HEADER START 3230557Sjimharris * 4230557Sjimharris * The contents of this file are subject to the terms of the 5230557Sjimharris * Common Development and Distribution License (the "License"). 6230557Sjimharris * You may not use this file except in compliance with the License. 7230557Sjimharris * 8230557Sjimharris * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9230557Sjimharris * or http://www.opensolaris.org/os/licensing. 10230557Sjimharris * See the License for the specific language governing permissions 11230557Sjimharris * and limitations under the License. 12230557Sjimharris * 13230557Sjimharris * When distributing Covered Code, include this CDDL HEADER in each 14230557Sjimharris * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15230557Sjimharris * If applicable, add the following below this CDDL HEADER, with the 16230557Sjimharris * fields enclosed by brackets "[]" replaced with your own identifying 17230557Sjimharris * information: Portions Copyright [yyyy] [name of copyright owner] 18230557Sjimharris * 19230557Sjimharris * CDDL HEADER END 20230557Sjimharris */ 21230557Sjimharris 22230557Sjimharris/* 23230557Sjimharris * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved. 24230557Sjimharris * Copyright (c) 2012 by Delphix. All rights reserved. 25230557Sjimharris */ 26230557Sjimharris 27230557Sjimharris/* 28230557Sjimharris * This file contains the functions which analyze the status of a pool. This 29230557Sjimharris * include both the status of an active pool, as well as the status exported 30230557Sjimharris * pools. Returns one of the ZPOOL_STATUS_* defines describing the status of 31230557Sjimharris * the pool. This status is independent (to a certain degree) from the state of 32230557Sjimharris * the pool. A pool's state describes only whether or not it is capable of 33230557Sjimharris * providing the necessary fault tolerance for data. The status describes the 34230557Sjimharris * overall status of devices. A pool that is online can still have a device 35230557Sjimharris * that is experiencing errors. 36230557Sjimharris * 37230557Sjimharris * Only a subset of the possible faults can be detected using 'zpool status', 38230557Sjimharris * and not all possible errors correspond to a FMA message ID. The explanation 39230557Sjimharris * is left up to the caller, depending on whether it is a live pool or an 40230557Sjimharris * import. 41230557Sjimharris */ 42230557Sjimharris 43230557Sjimharris#include <libzfs.h> 44230557Sjimharris#include <string.h> 45230557Sjimharris#include <unistd.h> 46230557Sjimharris#include "libzfs_impl.h" 47230557Sjimharris#include "zfeature_common.h" 48230557Sjimharris 49230557Sjimharris/* 50230557Sjimharris * Message ID table. This must be kept in sync with the ZPOOL_STATUS_* defines 51230557Sjimharris * in libzfs.h. Note that there are some status results which go past the end 52230557Sjimharris * of this table, and hence have no associated message ID. 53230557Sjimharris */ 54230557Sjimharrisstatic char *zfs_msgid_table[] = { 55230557Sjimharris "ZFS-8000-14", 56230557Sjimharris "ZFS-8000-2Q", 57230557Sjimharris "ZFS-8000-3C", 58230557Sjimharris "ZFS-8000-4J", 59230557Sjimharris "ZFS-8000-5E", 60230557Sjimharris "ZFS-8000-6X", 61230557Sjimharris "ZFS-8000-72", 62230557Sjimharris "ZFS-8000-8A", 63230557Sjimharris "ZFS-8000-9P", 64230557Sjimharris "ZFS-8000-A5", 65230557Sjimharris "ZFS-8000-EY", 66230557Sjimharris "ZFS-8000-HC", 67230557Sjimharris "ZFS-8000-JQ", 68230557Sjimharris "ZFS-8000-K4", 69230557Sjimharris}; 70230557Sjimharris 71230557Sjimharris#define NMSGID (sizeof (zfs_msgid_table) / sizeof (zfs_msgid_table[0])) 72230557Sjimharris 73230557Sjimharris/* ARGSUSED */ 74230557Sjimharrisstatic int 75230557Sjimharrisvdev_missing(uint64_t state, uint64_t aux, uint64_t errs) 76230557Sjimharris{ 77230557Sjimharris return (state == VDEV_STATE_CANT_OPEN && 78230557Sjimharris aux == VDEV_AUX_OPEN_FAILED); 79230557Sjimharris} 80230557Sjimharris 81230557Sjimharris/* ARGSUSED */ 82230557Sjimharrisstatic int 83230557Sjimharrisvdev_faulted(uint64_t state, uint64_t aux, uint64_t errs) 84230557Sjimharris{ 85230557Sjimharris return (state == VDEV_STATE_FAULTED); 86230557Sjimharris} 87230557Sjimharris 88230557Sjimharris/* ARGSUSED */ 89230557Sjimharrisstatic int 90230557Sjimharrisvdev_errors(uint64_t state, uint64_t aux, uint64_t errs) 91230557Sjimharris{ 92230557Sjimharris return (state == VDEV_STATE_DEGRADED || errs != 0); 93230557Sjimharris} 94230557Sjimharris 95/* ARGSUSED */ 96static int 97vdev_broken(uint64_t state, uint64_t aux, uint64_t errs) 98{ 99 return (state == VDEV_STATE_CANT_OPEN); 100} 101 102/* ARGSUSED */ 103static int 104vdev_offlined(uint64_t state, uint64_t aux, uint64_t errs) 105{ 106 return (state == VDEV_STATE_OFFLINE); 107} 108 109/* ARGSUSED */ 110static int 111vdev_removed(uint64_t state, uint64_t aux, uint64_t errs) 112{ 113 return (state == VDEV_STATE_REMOVED); 114} 115 116/* 117 * Detect if any leaf devices that have seen errors or could not be opened. 118 */ 119static boolean_t 120find_vdev_problem(nvlist_t *vdev, int (*func)(uint64_t, uint64_t, uint64_t)) 121{ 122 nvlist_t **child; 123 vdev_stat_t *vs; 124 uint_t c, children; 125 char *type; 126 127 /* 128 * Ignore problems within a 'replacing' vdev, since we're presumably in 129 * the process of repairing any such errors, and don't want to call them 130 * out again. We'll pick up the fact that a resilver is happening 131 * later. 132 */ 133 verify(nvlist_lookup_string(vdev, ZPOOL_CONFIG_TYPE, &type) == 0); 134 if (strcmp(type, VDEV_TYPE_REPLACING) == 0) 135 return (B_FALSE); 136 137 if (nvlist_lookup_nvlist_array(vdev, ZPOOL_CONFIG_CHILDREN, &child, 138 &children) == 0) { 139 for (c = 0; c < children; c++) 140 if (find_vdev_problem(child[c], func)) 141 return (B_TRUE); 142 } else { 143 verify(nvlist_lookup_uint64_array(vdev, ZPOOL_CONFIG_VDEV_STATS, 144 (uint64_t **)&vs, &c) == 0); 145 146 if (func(vs->vs_state, vs->vs_aux, 147 vs->vs_read_errors + 148 vs->vs_write_errors + 149 vs->vs_checksum_errors)) 150 return (B_TRUE); 151 } 152 153 return (B_FALSE); 154} 155 156/* 157 * Active pool health status. 158 * 159 * To determine the status for a pool, we make several passes over the config, 160 * picking the most egregious error we find. In order of importance, we do the 161 * following: 162 * 163 * - Check for a complete and valid configuration 164 * - Look for any faulted or missing devices in a non-replicated config 165 * - Check for any data errors 166 * - Check for any faulted or missing devices in a replicated config 167 * - Look for any devices showing errors 168 * - Check for any resilvering devices 169 * 170 * There can obviously be multiple errors within a single pool, so this routine 171 * only picks the most damaging of all the current errors to report. 172 */ 173static zpool_status_t 174check_status(nvlist_t *config, boolean_t isimport) 175{ 176 nvlist_t *nvroot; 177 vdev_stat_t *vs; 178 pool_scan_stat_t *ps = NULL; 179 uint_t vsc, psc; 180 uint64_t nerr; 181 uint64_t version; 182 uint64_t stateval; 183 uint64_t suspended; 184 uint64_t hostid = 0; 185 186 verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_VERSION, 187 &version) == 0); 188 verify(nvlist_lookup_nvlist(config, ZPOOL_CONFIG_VDEV_TREE, 189 &nvroot) == 0); 190 verify(nvlist_lookup_uint64_array(nvroot, ZPOOL_CONFIG_VDEV_STATS, 191 (uint64_t **)&vs, &vsc) == 0); 192 verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_STATE, 193 &stateval) == 0); 194 195 /* 196 * Currently resilvering a vdev 197 */ 198 (void) nvlist_lookup_uint64_array(nvroot, ZPOOL_CONFIG_SCAN_STATS, 199 (uint64_t **)&ps, &psc); 200 if (ps && ps->pss_func == POOL_SCAN_RESILVER && 201 ps->pss_state == DSS_SCANNING) 202 return (ZPOOL_STATUS_RESILVERING); 203 204 /* 205 * Pool last accessed by another system. 206 */ 207 (void) nvlist_lookup_uint64(config, ZPOOL_CONFIG_HOSTID, &hostid); 208 if (hostid != 0 && (unsigned long)hostid != gethostid() && 209 stateval == POOL_STATE_ACTIVE) 210 return (ZPOOL_STATUS_HOSTID_MISMATCH); 211 212 /* 213 * Newer on-disk version. 214 */ 215 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 216 vs->vs_aux == VDEV_AUX_VERSION_NEWER) 217 return (ZPOOL_STATUS_VERSION_NEWER); 218 219 /* 220 * Unsupported feature(s). 221 */ 222 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 223 vs->vs_aux == VDEV_AUX_UNSUP_FEAT) { 224 nvlist_t *nvinfo; 225 226 verify(nvlist_lookup_nvlist(config, ZPOOL_CONFIG_LOAD_INFO, 227 &nvinfo) == 0); 228 if (nvlist_exists(nvinfo, ZPOOL_CONFIG_CAN_RDONLY)) 229 return (ZPOOL_STATUS_UNSUP_FEAT_WRITE); 230 return (ZPOOL_STATUS_UNSUP_FEAT_READ); 231 } 232 233 /* 234 * Check that the config is complete. 235 */ 236 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 237 vs->vs_aux == VDEV_AUX_BAD_GUID_SUM) 238 return (ZPOOL_STATUS_BAD_GUID_SUM); 239 240 /* 241 * Check whether the pool has suspended due to failed I/O. 242 */ 243 if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_SUSPENDED, 244 &suspended) == 0) { 245 if (suspended == ZIO_FAILURE_MODE_CONTINUE) 246 return (ZPOOL_STATUS_IO_FAILURE_CONTINUE); 247 return (ZPOOL_STATUS_IO_FAILURE_WAIT); 248 } 249 250 /* 251 * Could not read a log. 252 */ 253 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 254 vs->vs_aux == VDEV_AUX_BAD_LOG) { 255 return (ZPOOL_STATUS_BAD_LOG); 256 } 257 258 /* 259 * Bad devices in non-replicated config. 260 */ 261 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 262 find_vdev_problem(nvroot, vdev_faulted)) 263 return (ZPOOL_STATUS_FAULTED_DEV_NR); 264 265 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 266 find_vdev_problem(nvroot, vdev_missing)) 267 return (ZPOOL_STATUS_MISSING_DEV_NR); 268 269 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 270 find_vdev_problem(nvroot, vdev_broken)) 271 return (ZPOOL_STATUS_CORRUPT_LABEL_NR); 272 273 /* 274 * Corrupted pool metadata 275 */ 276 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 277 vs->vs_aux == VDEV_AUX_CORRUPT_DATA) 278 return (ZPOOL_STATUS_CORRUPT_POOL); 279 280 /* 281 * Persistent data errors. 282 */ 283 if (!isimport) { 284 if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_ERRCOUNT, 285 &nerr) == 0 && nerr != 0) 286 return (ZPOOL_STATUS_CORRUPT_DATA); 287 } 288 289 /* 290 * Missing devices in a replicated config. 291 */ 292 if (find_vdev_problem(nvroot, vdev_faulted)) 293 return (ZPOOL_STATUS_FAULTED_DEV_R); 294 if (find_vdev_problem(nvroot, vdev_missing)) 295 return (ZPOOL_STATUS_MISSING_DEV_R); 296 if (find_vdev_problem(nvroot, vdev_broken)) 297 return (ZPOOL_STATUS_CORRUPT_LABEL_R); 298 299 /* 300 * Devices with errors 301 */ 302 if (!isimport && find_vdev_problem(nvroot, vdev_errors)) 303 return (ZPOOL_STATUS_FAILING_DEV); 304 305 /* 306 * Offlined devices 307 */ 308 if (find_vdev_problem(nvroot, vdev_offlined)) 309 return (ZPOOL_STATUS_OFFLINE_DEV); 310 311 /* 312 * Removed device 313 */ 314 if (find_vdev_problem(nvroot, vdev_removed)) 315 return (ZPOOL_STATUS_REMOVED_DEV); 316 317 /* 318 * Outdated, but usable, version 319 */ 320 if (SPA_VERSION_IS_SUPPORTED(version) && version != SPA_VERSION) 321 return (ZPOOL_STATUS_VERSION_OLDER); 322 323 /* 324 * Usable pool with disabled features 325 */ 326 if (version >= SPA_VERSION_FEATURES) { 327 int i; 328 nvlist_t *feat; 329 330 if (isimport) { 331 feat = fnvlist_lookup_nvlist(config, 332 ZPOOL_CONFIG_LOAD_INFO); 333 feat = fnvlist_lookup_nvlist(feat, 334 ZPOOL_CONFIG_ENABLED_FEAT); 335 } else { 336 feat = fnvlist_lookup_nvlist(config, 337 ZPOOL_CONFIG_FEATURE_STATS); 338 } 339 340 for (i = 0; i < SPA_FEATURES; i++) { 341 zfeature_info_t *fi = &spa_feature_table[i]; 342 if (!nvlist_exists(feat, fi->fi_guid)) 343 return (ZPOOL_STATUS_FEAT_DISABLED); 344 } 345 } 346 347 return (ZPOOL_STATUS_OK); 348} 349 350zpool_status_t 351zpool_get_status(zpool_handle_t *zhp, char **msgid) 352{ 353 zpool_status_t ret = check_status(zhp->zpool_config, B_FALSE); 354 355 if (ret >= NMSGID) 356 *msgid = NULL; 357 else 358 *msgid = zfs_msgid_table[ret]; 359 360 return (ret); 361} 362 363zpool_status_t 364zpool_import_status(nvlist_t *config, char **msgid) 365{ 366 zpool_status_t ret = check_status(config, B_TRUE); 367 368 if (ret >= NMSGID) 369 *msgid = NULL; 370 else 371 *msgid = zfs_msgid_table[ret]; 372 373 return (ret); 374} 375 376static void 377dump_ddt_stat(const ddt_stat_t *dds, int h) 378{ 379 char refcnt[6]; 380 char blocks[6], lsize[6], psize[6], dsize[6]; 381 char ref_blocks[6], ref_lsize[6], ref_psize[6], ref_dsize[6]; 382 383 if (dds == NULL || dds->dds_blocks == 0) 384 return; 385 386 if (h == -1) 387 (void) strcpy(refcnt, "Total"); 388 else 389 zfs_nicenum(1ULL << h, refcnt, sizeof (refcnt)); 390 391 zfs_nicenum(dds->dds_blocks, blocks, sizeof (blocks)); 392 zfs_nicenum(dds->dds_lsize, lsize, sizeof (lsize)); 393 zfs_nicenum(dds->dds_psize, psize, sizeof (psize)); 394 zfs_nicenum(dds->dds_dsize, dsize, sizeof (dsize)); 395 zfs_nicenum(dds->dds_ref_blocks, ref_blocks, sizeof (ref_blocks)); 396 zfs_nicenum(dds->dds_ref_lsize, ref_lsize, sizeof (ref_lsize)); 397 zfs_nicenum(dds->dds_ref_psize, ref_psize, sizeof (ref_psize)); 398 zfs_nicenum(dds->dds_ref_dsize, ref_dsize, sizeof (ref_dsize)); 399 400 (void) printf("%6s %6s %5s %5s %5s %6s %5s %5s %5s\n", 401 refcnt, 402 blocks, lsize, psize, dsize, 403 ref_blocks, ref_lsize, ref_psize, ref_dsize); 404} 405 406/* 407 * Print the DDT histogram and the column totals. 408 */ 409void 410zpool_dump_ddt(const ddt_stat_t *dds_total, const ddt_histogram_t *ddh) 411{ 412 int h; 413 414 (void) printf("\n"); 415 416 (void) printf("bucket " 417 " allocated " 418 " referenced \n"); 419 (void) printf("______ " 420 "______________________________ " 421 "______________________________\n"); 422 423 (void) printf("%6s %6s %5s %5s %5s %6s %5s %5s %5s\n", 424 "refcnt", 425 "blocks", "LSIZE", "PSIZE", "DSIZE", 426 "blocks", "LSIZE", "PSIZE", "DSIZE"); 427 428 (void) printf("%6s %6s %5s %5s %5s %6s %5s %5s %5s\n", 429 "------", 430 "------", "-----", "-----", "-----", 431 "------", "-----", "-----", "-----"); 432 433 for (h = 0; h < 64; h++) 434 dump_ddt_stat(&ddh->ddh_stat[h], h); 435 436 dump_ddt_stat(dds_total, -1); 437 438 (void) printf("\n"); 439} 440