1/* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22/* 23 * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved. 24 * Copyright (c) 2012 by Delphix. All rights reserved. 25 * Copyright (c) 2013 Steven Hartland. All rights reserved. 26 */ 27 28/* 29 * This file contains the functions which analyze the status of a pool. This 30 * include both the status of an active pool, as well as the status exported 31 * pools. Returns one of the ZPOOL_STATUS_* defines describing the status of 32 * the pool. This status is independent (to a certain degree) from the state of 33 * the pool. A pool's state describes only whether or not it is capable of 34 * providing the necessary fault tolerance for data. The status describes the 35 * overall status of devices. A pool that is online can still have a device 36 * that is experiencing errors. 37 * 38 * Only a subset of the possible faults can be detected using 'zpool status', 39 * and not all possible errors correspond to a FMA message ID. The explanation 40 * is left up to the caller, depending on whether it is a live pool or an 41 * import. 42 */ 43 44#include <libzfs.h> 45#include <string.h> 46#include <unistd.h> 47#include "libzfs_impl.h" 48#include "zfeature_common.h" 49 50/* 51 * Message ID table. This must be kept in sync with the ZPOOL_STATUS_* defines 52 * in libzfs.h. Note that there are some status results which go past the end 53 * of this table, and hence have no associated message ID. 54 */ 55static char *zfs_msgid_table[] = { 56 "ZFS-8000-14", /* ZPOOL_STATUS_CORRUPT_CACHE */ 57 "ZFS-8000-2Q", /* ZPOOL_STATUS_MISSING_DEV_R */ 58 "ZFS-8000-3C", /* ZPOOL_STATUS_MISSING_DEV_NR */ 59 "ZFS-8000-4J", /* ZPOOL_STATUS_CORRUPT_LABEL_R */ 60 "ZFS-8000-5E", /* ZPOOL_STATUS_CORRUPT_LABEL_NR */ 61 "ZFS-8000-6X", /* ZPOOL_STATUS_BAD_GUID_SUM */ 62 "ZFS-8000-72", /* ZPOOL_STATUS_CORRUPT_POOL */ 63 "ZFS-8000-8A", /* ZPOOL_STATUS_CORRUPT_DATA */ 64 "ZFS-8000-9P", /* ZPOOL_STATUS_FAILING_DEV */ 65 "ZFS-8000-A5", /* ZPOOL_STATUS_VERSION_NEWER */ 66 "ZFS-8000-EY", /* ZPOOL_STATUS_HOSTID_MISMATCH */ 67 "ZFS-8000-EY", /* ZPOOL_STATUS_HOSTID_ACTIVE */ 68 "ZFS-8000-EY", /* ZPOOL_STATUS_HOSTID_REQUIRED */ 69 "ZFS-8000-HC", /* ZPOOL_STATUS_IO_FAILURE_WAIT */ 70 "ZFS-8000-JQ", /* ZPOOL_STATUS_IO_FAILURE_CONTINUE */ 71 "ZFS-8000-MM", /* ZPOOL_STATUS_IO_FAILURE_MMP */ 72 "ZFS-8000-K4", /* ZPOOL_STATUS_BAD_LOG */ 73 /* 74 * The following results have no message ID. 75 * ZPOOL_STATUS_UNSUP_FEAT_READ 76 * ZPOOL_STATUS_UNSUP_FEAT_WRITE 77 * ZPOOL_STATUS_FAULTED_DEV_R 78 * ZPOOL_STATUS_FAULTED_DEV_NR 79 * ZPOOL_STATUS_VERSION_OLDER 80 * ZPOOL_STATUS_FEAT_DISABLED 81 * ZPOOL_STATUS_RESILVERING 82 * ZPOOL_STATUS_OFFLINE_DEV 83 * ZPOOL_STATUS_REMOVED_DEV 84 * ZPOOL_STATUS_OK 85 */ 86}; 87 88#define NMSGID (sizeof (zfs_msgid_table) / sizeof (zfs_msgid_table[0])) 89 90/* ARGSUSED */ 91static int 92vdev_missing(vdev_stat_t *vs, uint_t vsc) 93{ 94 return (vs->vs_state == VDEV_STATE_CANT_OPEN && 95 vs->vs_aux == VDEV_AUX_OPEN_FAILED); 96} 97 98/* ARGSUSED */ 99static int 100vdev_faulted(vdev_stat_t *vs, uint_t vsc) 101{ 102 return (vs->vs_state == VDEV_STATE_FAULTED); 103} 104 105/* ARGSUSED */ 106static int 107vdev_errors(vdev_stat_t *vs, uint_t vsc) 108{ 109 return (vs->vs_state == VDEV_STATE_DEGRADED || 110 vs->vs_read_errors != 0 || vs->vs_write_errors != 0 || 111 vs->vs_checksum_errors != 0); 112} 113 114/* ARGSUSED */ 115static int 116vdev_broken(vdev_stat_t *vs, uint_t vsc) 117{ 118 return (vs->vs_state == VDEV_STATE_CANT_OPEN); 119} 120 121/* ARGSUSED */ 122static int 123vdev_offlined(vdev_stat_t *vs, uint_t vsc) 124{ 125 return (vs->vs_state == VDEV_STATE_OFFLINE); 126} 127 128/* ARGSUSED */ 129static int 130vdev_removed(vdev_stat_t *vs, uint_t vsc) 131{ 132 return (vs->vs_state == VDEV_STATE_REMOVED); 133} 134 135static int 136vdev_non_native_ashift(vdev_stat_t *vs, uint_t vsc) 137{ 138 return (VDEV_STAT_VALID(vs_physical_ashift, vsc) && 139 vs->vs_configured_ashift < vs->vs_physical_ashift); 140} 141 142/* 143 * Detect if any leaf devices that have seen errors or could not be opened. 144 */ 145static boolean_t 146find_vdev_problem(nvlist_t *vdev, int (*func)(vdev_stat_t *, uint_t), 147 boolean_t ignore_replacing) 148{ 149 nvlist_t **child; 150 vdev_stat_t *vs; 151 uint_t c, vsc, children; 152 153 /* 154 * Ignore problems within a 'replacing' vdev, since we're presumably in 155 * the process of repairing any such errors, and don't want to call them 156 * out again. We'll pick up the fact that a resilver is happening 157 * later. 158 */ 159 if (ignore_replacing == B_TRUE) { 160 char *type; 161 162 verify(nvlist_lookup_string(vdev, ZPOOL_CONFIG_TYPE, 163 &type) == 0); 164 if (strcmp(type, VDEV_TYPE_REPLACING) == 0) 165 return (B_FALSE); 166 } 167 168 if (nvlist_lookup_nvlist_array(vdev, ZPOOL_CONFIG_CHILDREN, &child, 169 &children) == 0) { 170 for (c = 0; c < children; c++) 171 if (find_vdev_problem(child[c], func, ignore_replacing)) 172 return (B_TRUE); 173 } else { 174 verify(nvlist_lookup_uint64_array(vdev, ZPOOL_CONFIG_VDEV_STATS, 175 (uint64_t **)&vs, &vsc) == 0); 176 177 if (func(vs, vsc) != 0) 178 return (B_TRUE); 179 } 180 181 /* 182 * Check any L2 cache devs 183 */ 184 if (nvlist_lookup_nvlist_array(vdev, ZPOOL_CONFIG_L2CACHE, &child, 185 &children) == 0) { 186 for (c = 0; c < children; c++) 187 if (find_vdev_problem(child[c], func, ignore_replacing)) 188 return (B_TRUE); 189 } 190 191 return (B_FALSE); 192} 193 194/* 195 * Active pool health status. 196 * 197 * To determine the status for a pool, we make several passes over the config, 198 * picking the most egregious error we find. In order of importance, we do the 199 * following: 200 * 201 * - Check for a complete and valid configuration 202 * - Look for any faulted or missing devices in a non-replicated config 203 * - Check for any data errors 204 * - Check for any faulted or missing devices in a replicated config 205 * - Look for any devices showing errors 206 * - Check for any resilvering devices 207 * 208 * There can obviously be multiple errors within a single pool, so this routine 209 * only picks the most damaging of all the current errors to report. 210 */ 211static zpool_status_t 212check_status(nvlist_t *config, boolean_t isimport) 213{ 214 nvlist_t *nvroot; 215 vdev_stat_t *vs; 216 pool_scan_stat_t *ps = NULL; 217 uint_t vsc, psc; 218 uint64_t nerr; 219 uint64_t version; 220 uint64_t stateval; 221 uint64_t suspended; 222 uint64_t hostid = 0; 223 unsigned long system_hostid = get_system_hostid(); 224 225 verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_VERSION, 226 &version) == 0); 227 verify(nvlist_lookup_nvlist(config, ZPOOL_CONFIG_VDEV_TREE, 228 &nvroot) == 0); 229 verify(nvlist_lookup_uint64_array(nvroot, ZPOOL_CONFIG_VDEV_STATS, 230 (uint64_t **)&vs, &vsc) == 0); 231 verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_STATE, 232 &stateval) == 0); 233 234 /* 235 * Currently resilvering a vdev 236 */ 237 (void) nvlist_lookup_uint64_array(nvroot, ZPOOL_CONFIG_SCAN_STATS, 238 (uint64_t **)&ps, &psc); 239 if (ps != NULL && ps->pss_func == POOL_SCAN_RESILVER && 240 ps->pss_state == DSS_SCANNING) 241 return (ZPOOL_STATUS_RESILVERING); 242 243 /* 244 * The multihost property is set and the pool may be active. 245 */ 246 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 247 vs->vs_aux == VDEV_AUX_ACTIVE) { 248 mmp_state_t mmp_state; 249 nvlist_t *nvinfo; 250 251 nvinfo = fnvlist_lookup_nvlist(config, ZPOOL_CONFIG_LOAD_INFO); 252 mmp_state = fnvlist_lookup_uint64(nvinfo, 253 ZPOOL_CONFIG_MMP_STATE); 254 255 if (mmp_state == MMP_STATE_ACTIVE) 256 return (ZPOOL_STATUS_HOSTID_ACTIVE); 257 else if (mmp_state == MMP_STATE_NO_HOSTID) 258 return (ZPOOL_STATUS_HOSTID_REQUIRED); 259 else 260 return (ZPOOL_STATUS_HOSTID_MISMATCH); 261 } 262 263 /* 264 * Pool last accessed by another system. 265 */ 266 (void) nvlist_lookup_uint64(config, ZPOOL_CONFIG_HOSTID, &hostid); 267 if (hostid != 0 && (unsigned long)hostid != system_hostid && 268 stateval == POOL_STATE_ACTIVE) 269 return (ZPOOL_STATUS_HOSTID_MISMATCH); 270 271 /* 272 * Newer on-disk version. 273 */ 274 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 275 vs->vs_aux == VDEV_AUX_VERSION_NEWER) 276 return (ZPOOL_STATUS_VERSION_NEWER); 277 278 /* 279 * Unsupported feature(s). 280 */ 281 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 282 vs->vs_aux == VDEV_AUX_UNSUP_FEAT) { 283 nvlist_t *nvinfo; 284 285 verify(nvlist_lookup_nvlist(config, ZPOOL_CONFIG_LOAD_INFO, 286 &nvinfo) == 0); 287 if (nvlist_exists(nvinfo, ZPOOL_CONFIG_CAN_RDONLY)) 288 return (ZPOOL_STATUS_UNSUP_FEAT_WRITE); 289 return (ZPOOL_STATUS_UNSUP_FEAT_READ); 290 } 291 292 /* 293 * Check that the config is complete. 294 */ 295 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 296 vs->vs_aux == VDEV_AUX_BAD_GUID_SUM) 297 return (ZPOOL_STATUS_BAD_GUID_SUM); 298 299 /* 300 * Check whether the pool has suspended. 301 */ 302 if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_SUSPENDED, 303 &suspended) == 0) { 304 uint64_t reason; 305 306 if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_SUSPENDED_REASON, 307 &reason) == 0 && reason == ZIO_SUSPEND_MMP) 308 return (ZPOOL_STATUS_IO_FAILURE_MMP); 309 310 if (suspended == ZIO_FAILURE_MODE_CONTINUE) 311 return (ZPOOL_STATUS_IO_FAILURE_CONTINUE); 312 return (ZPOOL_STATUS_IO_FAILURE_WAIT); 313 } 314 315 /* 316 * Could not read a log. 317 */ 318 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 319 vs->vs_aux == VDEV_AUX_BAD_LOG) { 320 return (ZPOOL_STATUS_BAD_LOG); 321 } 322 323 /* 324 * Bad devices in non-replicated config. 325 */ 326 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 327 find_vdev_problem(nvroot, vdev_faulted, B_TRUE)) 328 return (ZPOOL_STATUS_FAULTED_DEV_NR); 329 330 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 331 find_vdev_problem(nvroot, vdev_missing, B_TRUE)) 332 return (ZPOOL_STATUS_MISSING_DEV_NR); 333 334 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 335 find_vdev_problem(nvroot, vdev_broken, B_TRUE)) 336 return (ZPOOL_STATUS_CORRUPT_LABEL_NR); 337 338 /* 339 * Corrupted pool metadata 340 */ 341 if (vs->vs_state == VDEV_STATE_CANT_OPEN && 342 vs->vs_aux == VDEV_AUX_CORRUPT_DATA) 343 return (ZPOOL_STATUS_CORRUPT_POOL); 344 345 /* 346 * Persistent data errors. 347 */ 348 if (!isimport) { 349 if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_ERRCOUNT, 350 &nerr) == 0 && nerr != 0) 351 return (ZPOOL_STATUS_CORRUPT_DATA); 352 } 353 354 /* 355 * Missing devices in a replicated config. 356 */ 357 if (find_vdev_problem(nvroot, vdev_faulted, B_TRUE)) 358 return (ZPOOL_STATUS_FAULTED_DEV_R); 359 if (find_vdev_problem(nvroot, vdev_missing, B_TRUE)) 360 return (ZPOOL_STATUS_MISSING_DEV_R); 361 if (find_vdev_problem(nvroot, vdev_broken, B_TRUE)) 362 return (ZPOOL_STATUS_CORRUPT_LABEL_R); 363 364 /* 365 * Devices with errors 366 */ 367 if (!isimport && find_vdev_problem(nvroot, vdev_errors, B_TRUE)) 368 return (ZPOOL_STATUS_FAILING_DEV); 369 370 /* 371 * Offlined devices 372 */ 373 if (find_vdev_problem(nvroot, vdev_offlined, B_TRUE)) 374 return (ZPOOL_STATUS_OFFLINE_DEV); 375 376 /* 377 * Removed device 378 */ 379 if (find_vdev_problem(nvroot, vdev_removed, B_TRUE)) 380 return (ZPOOL_STATUS_REMOVED_DEV); 381 382 /* 383 * Suboptimal, but usable, ashift configuration. 384 */ 385 if (find_vdev_problem(nvroot, vdev_non_native_ashift, B_FALSE)) 386 return (ZPOOL_STATUS_NON_NATIVE_ASHIFT); 387 388 /* 389 * Outdated, but usable, version 390 */ 391 if (SPA_VERSION_IS_SUPPORTED(version) && version != SPA_VERSION) 392 return (ZPOOL_STATUS_VERSION_OLDER); 393 394 /* 395 * Usable pool with disabled features 396 */ 397 if (version >= SPA_VERSION_FEATURES) { 398 int i; 399 nvlist_t *feat; 400 401 if (isimport) { 402 feat = fnvlist_lookup_nvlist(config, 403 ZPOOL_CONFIG_LOAD_INFO); 404 if (nvlist_exists(feat, ZPOOL_CONFIG_ENABLED_FEAT)) 405 feat = fnvlist_lookup_nvlist(feat, 406 ZPOOL_CONFIG_ENABLED_FEAT); 407 } else { 408 feat = fnvlist_lookup_nvlist(config, 409 ZPOOL_CONFIG_FEATURE_STATS); 410 } 411 412 for (i = 0; i < SPA_FEATURES; i++) { 413 zfeature_info_t *fi = &spa_feature_table[i]; 414 if (!nvlist_exists(feat, fi->fi_guid)) 415 return (ZPOOL_STATUS_FEAT_DISABLED); 416 } 417 } 418 419 return (ZPOOL_STATUS_OK); 420} 421 422zpool_status_t 423zpool_get_status(zpool_handle_t *zhp, char **msgid) 424{ 425 zpool_status_t ret = check_status(zhp->zpool_config, B_FALSE); 426 427 if (ret >= NMSGID) 428 *msgid = NULL; 429 else 430 *msgid = zfs_msgid_table[ret]; 431 432 return (ret); 433} 434 435zpool_status_t 436zpool_import_status(nvlist_t *config, char **msgid) 437{ 438 zpool_status_t ret = check_status(config, B_TRUE); 439 440 if (ret >= NMSGID) 441 *msgid = NULL; 442 else 443 *msgid = zfs_msgid_table[ret]; 444 445 return (ret); 446} 447 448static void 449dump_ddt_stat(const ddt_stat_t *dds, int h) 450{ 451 char refcnt[6]; 452 char blocks[6], lsize[6], psize[6], dsize[6]; 453 char ref_blocks[6], ref_lsize[6], ref_psize[6], ref_dsize[6]; 454 455 if (dds == NULL || dds->dds_blocks == 0) 456 return; 457 458 if (h == -1) 459 (void) strcpy(refcnt, "Total"); 460 else 461 zfs_nicenum(1ULL << h, refcnt, sizeof (refcnt)); 462 463 zfs_nicenum(dds->dds_blocks, blocks, sizeof (blocks)); 464 zfs_nicenum(dds->dds_lsize, lsize, sizeof (lsize)); 465 zfs_nicenum(dds->dds_psize, psize, sizeof (psize)); 466 zfs_nicenum(dds->dds_dsize, dsize, sizeof (dsize)); 467 zfs_nicenum(dds->dds_ref_blocks, ref_blocks, sizeof (ref_blocks)); 468 zfs_nicenum(dds->dds_ref_lsize, ref_lsize, sizeof (ref_lsize)); 469 zfs_nicenum(dds->dds_ref_psize, ref_psize, sizeof (ref_psize)); 470 zfs_nicenum(dds->dds_ref_dsize, ref_dsize, sizeof (ref_dsize)); 471 472 (void) printf("%6s %6s %5s %5s %5s %6s %5s %5s %5s\n", 473 refcnt, 474 blocks, lsize, psize, dsize, 475 ref_blocks, ref_lsize, ref_psize, ref_dsize); 476} 477 478/* 479 * Print the DDT histogram and the column totals. 480 */ 481void 482zpool_dump_ddt(const ddt_stat_t *dds_total, const ddt_histogram_t *ddh) 483{ 484 int h; 485 486 (void) printf("\n"); 487 488 (void) printf("bucket " 489 " allocated " 490 " referenced \n"); 491 (void) printf("______ " 492 "______________________________ " 493 "______________________________\n"); 494 495 (void) printf("%6s %6s %5s %5s %5s %6s %5s %5s %5s\n", 496 "refcnt", 497 "blocks", "LSIZE", "PSIZE", "DSIZE", 498 "blocks", "LSIZE", "PSIZE", "DSIZE"); 499 500 (void) printf("%6s %6s %5s %5s %5s %6s %5s %5s %5s\n", 501 "------", 502 "------", "-----", "-----", "-----", 503 "------", "-----", "-----", "-----"); 504 505 for (h = 0; h < 64; h++) 506 dump_ddt_stat(&ddh->ddh_stat[h], h); 507 508 dump_ddt_stat(dds_total, -1); 509 510 (void) printf("\n"); 511} 512