md_nvidia.c revision 220209
1219974Smav/*- 2219974Smav * Copyright (c) 2011 Alexander Motin <mav@FreeBSD.org> 3219974Smav * All rights reserved. 4219974Smav * 5219974Smav * Redistribution and use in source and binary forms, with or without 6219974Smav * modification, are permitted provided that the following conditions 7219974Smav * are met: 8219974Smav * 1. Redistributions of source code must retain the above copyright 9219974Smav * notice, this list of conditions and the following disclaimer. 10219974Smav * 2. Redistributions in binary form must reproduce the above copyright 11219974Smav * notice, this list of conditions and the following disclaimer in the 12219974Smav * documentation and/or other materials provided with the distribution. 13219974Smav * 14219974Smav * THIS SOFTWARE IS PROVIDED BY THE AUTHORS AND CONTRIBUTORS ``AS IS'' AND 15219974Smav * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16219974Smav * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17219974Smav * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHORS OR CONTRIBUTORS BE LIABLE 18219974Smav * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19219974Smav * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20219974Smav * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21219974Smav * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22219974Smav * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23219974Smav * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24219974Smav * SUCH DAMAGE. 25219974Smav */ 26219974Smav 27219974Smav#include <sys/cdefs.h> 28219974Smav__FBSDID("$FreeBSD: head/sys/geom/raid/md_nvidia.c 220209 2011-03-31 16:14:35Z mav $"); 29219974Smav 30219974Smav#include <sys/param.h> 31219974Smav#include <sys/bio.h> 32219974Smav#include <sys/endian.h> 33219974Smav#include <sys/kernel.h> 34219974Smav#include <sys/kobj.h> 35219974Smav#include <sys/limits.h> 36219974Smav#include <sys/lock.h> 37219974Smav#include <sys/malloc.h> 38219974Smav#include <sys/mutex.h> 39219974Smav#include <sys/systm.h> 40219974Smav#include <sys/taskqueue.h> 41219974Smav#include <geom/geom.h> 42219974Smav#include "geom/raid/g_raid.h" 43219974Smav#include "g_raid_md_if.h" 44219974Smav 45219974Smavstatic MALLOC_DEFINE(M_MD_NVIDIA, "md_nvidia_data", "GEOM_RAID NVIDIA metadata"); 46219974Smav 47219974Smavstruct nvidia_raid_conf { 48219974Smav uint8_t nvidia_id[8]; 49219974Smav#define NVIDIA_MAGIC "NVIDIA " 50219974Smav 51219974Smav uint32_t config_size; 52219974Smav uint32_t checksum; 53219974Smav uint16_t version; 54219974Smav uint8_t disk_number; 55219974Smav uint8_t dummy_0; 56219974Smav uint32_t total_sectors; 57219974Smav uint32_t sector_size; 58219974Smav uint8_t name[16]; 59219974Smav uint8_t revision[4]; 60219974Smav uint32_t disk_status; 61219974Smav 62219974Smav uint32_t magic_0; 63219974Smav#define NVIDIA_MAGIC0 0x00640044 64219974Smav 65219974Smav uint64_t volume_id[2]; 66219974Smav uint8_t state; 67219974Smav#define NVIDIA_S_IDLE 0 68219974Smav#define NVIDIA_S_INIT 2 69219974Smav#define NVIDIA_S_REBUILD 3 70219974Smav#define NVIDIA_S_UPGRADE 4 71219974Smav#define NVIDIA_S_SYNC 5 72219974Smav uint8_t array_width; 73219974Smav uint8_t total_disks; 74219974Smav uint8_t orig_array_width; 75219974Smav uint16_t type; 76219974Smav#define NVIDIA_T_RAID0 0x0080 77219974Smav#define NVIDIA_T_RAID1 0x0081 78219974Smav#define NVIDIA_T_RAID3 0x0083 79219974Smav#define NVIDIA_T_RAID5 0x0085 /* RLQ = 00/02? */ 80219974Smav#define NVIDIA_T_RAID5_SYM 0x0095 /* RLQ = 03 */ 81219974Smav#define NVIDIA_T_RAID10 0x008a 82219974Smav#define NVIDIA_T_RAID01 0x8180 83219974Smav#define NVIDIA_T_CONCAT 0x00ff 84219974Smav 85219974Smav uint16_t dummy_3; 86219974Smav uint32_t strip_sectors; 87219974Smav uint32_t strip_bytes; 88219974Smav uint32_t strip_shift; 89219974Smav uint32_t strip_mask; 90219974Smav uint32_t stripe_sectors; 91219974Smav uint32_t stripe_bytes; 92219974Smav uint32_t rebuild_lba; 93219974Smav uint32_t orig_type; 94219974Smav uint32_t orig_total_sectors; 95219974Smav uint32_t status; 96219974Smav#define NVIDIA_S_BOOTABLE 0x00000001 97219974Smav#define NVIDIA_S_DEGRADED 0x00000002 98219974Smav 99219974Smav uint32_t filler[98]; 100219974Smav} __packed; 101219974Smav 102219974Smavstruct g_raid_md_nvidia_perdisk { 103219974Smav struct nvidia_raid_conf *pd_meta; 104219974Smav int pd_disk_pos; 105219974Smav off_t pd_disk_size; 106219974Smav}; 107219974Smav 108219974Smavstruct g_raid_md_nvidia_object { 109219974Smav struct g_raid_md_object mdio_base; 110219974Smav uint64_t mdio_volume_id[2]; 111219974Smav struct nvidia_raid_conf *mdio_meta; 112219974Smav struct callout mdio_start_co; /* STARTING state timer. */ 113219974Smav int mdio_total_disks; 114219974Smav int mdio_disks_present; 115219974Smav int mdio_started; 116219974Smav int mdio_incomplete; 117219974Smav struct root_hold_token *mdio_rootmount; /* Root mount delay token. */ 118219974Smav}; 119219974Smav 120219974Smavstatic g_raid_md_create_t g_raid_md_create_nvidia; 121219974Smavstatic g_raid_md_taste_t g_raid_md_taste_nvidia; 122219974Smavstatic g_raid_md_event_t g_raid_md_event_nvidia; 123219974Smavstatic g_raid_md_ctl_t g_raid_md_ctl_nvidia; 124219974Smavstatic g_raid_md_write_t g_raid_md_write_nvidia; 125219974Smavstatic g_raid_md_fail_disk_t g_raid_md_fail_disk_nvidia; 126219974Smavstatic g_raid_md_free_disk_t g_raid_md_free_disk_nvidia; 127219974Smavstatic g_raid_md_free_t g_raid_md_free_nvidia; 128219974Smav 129219974Smavstatic kobj_method_t g_raid_md_nvidia_methods[] = { 130219974Smav KOBJMETHOD(g_raid_md_create, g_raid_md_create_nvidia), 131219974Smav KOBJMETHOD(g_raid_md_taste, g_raid_md_taste_nvidia), 132219974Smav KOBJMETHOD(g_raid_md_event, g_raid_md_event_nvidia), 133219974Smav KOBJMETHOD(g_raid_md_ctl, g_raid_md_ctl_nvidia), 134219974Smav KOBJMETHOD(g_raid_md_write, g_raid_md_write_nvidia), 135219974Smav KOBJMETHOD(g_raid_md_fail_disk, g_raid_md_fail_disk_nvidia), 136219974Smav KOBJMETHOD(g_raid_md_free_disk, g_raid_md_free_disk_nvidia), 137219974Smav KOBJMETHOD(g_raid_md_free, g_raid_md_free_nvidia), 138219974Smav { 0, 0 } 139219974Smav}; 140219974Smav 141219974Smavstatic struct g_raid_md_class g_raid_md_nvidia_class = { 142219974Smav "NVIDIA", 143219974Smav g_raid_md_nvidia_methods, 144219974Smav sizeof(struct g_raid_md_nvidia_object), 145219974Smav .mdc_priority = 100 146219974Smav}; 147219974Smav 148219974Smavstatic int NVIDIANodeID = 1; 149219974Smav 150219974Smavstatic void 151219974Smavg_raid_md_nvidia_print(struct nvidia_raid_conf *meta) 152219974Smav{ 153219974Smav 154219974Smav if (g_raid_debug < 1) 155219974Smav return; 156219974Smav 157219974Smav printf("********* ATA NVIDIA RAID Metadata *********\n"); 158219974Smav printf("nvidia_id <%.8s>\n", meta->nvidia_id); 159219974Smav printf("config_size %u\n", meta->config_size); 160219974Smav printf("checksum 0x%08x\n", meta->checksum); 161219974Smav printf("version 0x%04x\n", meta->version); 162219974Smav printf("disk_number %d\n", meta->disk_number); 163219974Smav printf("dummy_0 0x%02x\n", meta->dummy_0); 164219974Smav printf("total_sectors %u\n", meta->total_sectors); 165219974Smav printf("sector_size %u\n", meta->sector_size); 166219974Smav printf("name <%.16s>\n", meta->name); 167219974Smav printf("revision 0x%02x%02x%02x%02x\n", 168219974Smav meta->revision[0], meta->revision[1], 169219974Smav meta->revision[2], meta->revision[3]); 170219974Smav printf("disk_status 0x%08x\n", meta->disk_status); 171219974Smav printf("magic_0 0x%08x\n", meta->magic_0); 172219974Smav printf("volume_id 0x%016jx%016jx\n", 173219974Smav meta->volume_id[1], meta->volume_id[0]); 174219974Smav printf("state 0x%02x\n", meta->state); 175219974Smav printf("array_width %u\n", meta->array_width); 176219974Smav printf("total_disks %u\n", meta->total_disks); 177219974Smav printf("orig_array_width %u\n", meta->orig_array_width); 178219974Smav printf("type 0x%04x\n", meta->type); 179219974Smav printf("dummy_3 0x%04x\n", meta->dummy_3); 180219974Smav printf("strip_sectors %u\n", meta->strip_sectors); 181219974Smav printf("strip_bytes %u\n", meta->strip_bytes); 182219974Smav printf("strip_shift %u\n", meta->strip_shift); 183219974Smav printf("strip_mask 0x%08x\n", meta->strip_mask); 184219974Smav printf("stripe_sectors %u\n", meta->stripe_sectors); 185219974Smav printf("stripe_bytes %u\n", meta->stripe_bytes); 186219974Smav printf("rebuild_lba %u\n", meta->rebuild_lba); 187219974Smav printf("orig_type 0x%04x\n", meta->orig_type); 188219974Smav printf("orig_total_sectors %u\n", meta->orig_total_sectors); 189219974Smav printf("status 0x%08x\n", meta->status); 190219974Smav printf("=================================================\n"); 191219974Smav} 192219974Smav 193219974Smavstatic struct nvidia_raid_conf * 194219974Smavnvidia_meta_copy(struct nvidia_raid_conf *meta) 195219974Smav{ 196219974Smav struct nvidia_raid_conf *nmeta; 197219974Smav 198219974Smav nmeta = malloc(sizeof(*meta), M_MD_NVIDIA, M_WAITOK); 199219974Smav memcpy(nmeta, meta, sizeof(*meta)); 200219974Smav return (nmeta); 201219974Smav} 202219974Smav 203219974Smavstatic int 204219974Smavnvidia_meta_translate_disk(struct nvidia_raid_conf *meta, int md_disk_pos) 205219974Smav{ 206219974Smav int disk_pos; 207219974Smav 208219974Smav if (md_disk_pos >= 0 && meta->type == NVIDIA_T_RAID01) { 209219974Smav disk_pos = (md_disk_pos / meta->array_width) + 210219974Smav (md_disk_pos % meta->array_width) * meta->array_width; 211219974Smav } else 212219974Smav disk_pos = md_disk_pos; 213219974Smav return (disk_pos); 214219974Smav} 215219974Smav 216219974Smavstatic void 217219974Smavnvidia_meta_get_name(struct nvidia_raid_conf *meta, char *buf) 218219974Smav{ 219219974Smav int i; 220219974Smav 221219974Smav strncpy(buf, meta->name, 16); 222219974Smav buf[16] = 0; 223219974Smav for (i = 15; i >= 0; i--) { 224219974Smav if (buf[i] > 0x20) 225219974Smav break; 226219974Smav buf[i] = 0; 227219974Smav } 228219974Smav} 229219974Smav 230219974Smavstatic void 231219974Smavnvidia_meta_put_name(struct nvidia_raid_conf *meta, char *buf) 232219974Smav{ 233219974Smav 234219974Smav memset(meta->name, 0x20, 16); 235219974Smav memcpy(meta->name, buf, MIN(strlen(buf), 16)); 236219974Smav} 237219974Smav 238219974Smavstatic struct nvidia_raid_conf * 239219974Smavnvidia_meta_read(struct g_consumer *cp) 240219974Smav{ 241219974Smav struct g_provider *pp; 242219974Smav struct nvidia_raid_conf *meta; 243219974Smav char *buf; 244219974Smav int error, i; 245219974Smav uint32_t checksum, *ptr; 246219974Smav 247219974Smav pp = cp->provider; 248219974Smav 249219974Smav /* Read the anchor sector. */ 250219974Smav buf = g_read_data(cp, 251219974Smav pp->mediasize - 2 * pp->sectorsize, pp->sectorsize, &error); 252219974Smav if (buf == NULL) { 253219974Smav G_RAID_DEBUG(1, "Cannot read metadata from %s (error=%d).", 254219974Smav pp->name, error); 255219974Smav return (NULL); 256219974Smav } 257219974Smav meta = malloc(sizeof(*meta), M_MD_NVIDIA, M_WAITOK); 258219974Smav memcpy(meta, buf, min(sizeof(*meta), pp->sectorsize)); 259219974Smav g_free(buf); 260219974Smav 261219974Smav /* Check if this is an NVIDIA RAID struct */ 262219974Smav if (strncmp(meta->nvidia_id, NVIDIA_MAGIC, strlen(NVIDIA_MAGIC))) { 263219974Smav G_RAID_DEBUG(1, "NVIDIA signature check failed on %s", pp->name); 264219974Smav free(meta, M_MD_NVIDIA); 265219974Smav return (NULL); 266219974Smav } 267219974Smav if (meta->config_size > 128 || 268219974Smav meta->config_size < 30) { 269219974Smav G_RAID_DEBUG(1, "NVIDIA metadata size looks wrong: %d", 270219974Smav meta->config_size); 271219974Smav free(meta, M_MD_NVIDIA); 272219974Smav return (NULL); 273219974Smav } 274219974Smav 275219974Smav /* Check metadata checksum. */ 276219974Smav for (checksum = 0, ptr = (uint32_t *)meta, 277219974Smav i = 0; i < meta->config_size; i++) 278219974Smav checksum += *ptr++; 279219974Smav if (checksum != 0) { 280219974Smav G_RAID_DEBUG(1, "NVIDIA checksum check failed on %s", pp->name); 281219974Smav free(meta, M_MD_NVIDIA); 282219974Smav return (NULL); 283219974Smav } 284219974Smav 285219974Smav /* Check volume state. */ 286219974Smav if (meta->state != NVIDIA_S_IDLE && meta->state != NVIDIA_S_INIT && 287219974Smav meta->state != NVIDIA_S_REBUILD && meta->state != NVIDIA_S_SYNC) { 288219974Smav G_RAID_DEBUG(1, "NVIDIA unknown state on %s (0x%02x)", 289219974Smav pp->name, meta->state); 290219974Smav free(meta, M_MD_NVIDIA); 291219974Smav return (NULL); 292219974Smav } 293219974Smav 294219974Smav /* Check raid type. */ 295219974Smav if (meta->type != NVIDIA_T_RAID0 && meta->type != NVIDIA_T_RAID1 && 296219974Smav meta->type != NVIDIA_T_RAID3 && meta->type != NVIDIA_T_RAID5 && 297219974Smav meta->type != NVIDIA_T_RAID5_SYM && 298219974Smav meta->type != NVIDIA_T_RAID01 && meta->type != NVIDIA_T_CONCAT) { 299219974Smav G_RAID_DEBUG(1, "NVIDIA unknown RAID level on %s (0x%02x)", 300219974Smav pp->name, meta->type); 301219974Smav free(meta, M_MD_NVIDIA); 302219974Smav return (NULL); 303219974Smav } 304219974Smav 305219974Smav return (meta); 306219974Smav} 307219974Smav 308219974Smavstatic int 309219974Smavnvidia_meta_write(struct g_consumer *cp, struct nvidia_raid_conf *meta) 310219974Smav{ 311219974Smav struct g_provider *pp; 312219974Smav char *buf; 313219974Smav int error, i; 314219974Smav uint32_t checksum, *ptr; 315219974Smav 316219974Smav pp = cp->provider; 317219974Smav 318219974Smav /* Recalculate checksum for case if metadata were changed. */ 319219974Smav meta->checksum = 0; 320219974Smav for (checksum = 0, ptr = (uint32_t *)meta, 321219974Smav i = 0; i < meta->config_size; i++) 322219974Smav checksum += *ptr++; 323219974Smav meta->checksum -= checksum; 324219974Smav 325219974Smav /* Create and fill buffer. */ 326219974Smav buf = malloc(pp->sectorsize, M_MD_NVIDIA, M_WAITOK | M_ZERO); 327219974Smav memcpy(buf, meta, sizeof(*meta)); 328219974Smav 329219974Smav /* Write metadata. */ 330219974Smav error = g_write_data(cp, 331219974Smav pp->mediasize - 2 * pp->sectorsize, buf, pp->sectorsize); 332219974Smav if (error != 0) { 333219974Smav G_RAID_DEBUG(1, "Cannot write metadata to %s (error=%d).", 334219974Smav pp->name, error); 335219974Smav } 336219974Smav 337219974Smav free(buf, M_MD_NVIDIA); 338219974Smav return (error); 339219974Smav} 340219974Smav 341219974Smavstatic int 342219974Smavnvidia_meta_erase(struct g_consumer *cp) 343219974Smav{ 344219974Smav struct g_provider *pp; 345219974Smav char *buf; 346219974Smav int error; 347219974Smav 348219974Smav pp = cp->provider; 349219974Smav buf = malloc(pp->sectorsize, M_MD_NVIDIA, M_WAITOK | M_ZERO); 350219974Smav error = g_write_data(cp, 351219974Smav pp->mediasize - 2 * pp->sectorsize, buf, pp->sectorsize); 352219974Smav if (error != 0) { 353219974Smav G_RAID_DEBUG(1, "Cannot erase metadata on %s (error=%d).", 354219974Smav pp->name, error); 355219974Smav } 356219974Smav free(buf, M_MD_NVIDIA); 357219974Smav return (error); 358219974Smav} 359219974Smav 360219974Smavstatic struct g_raid_disk * 361219974Smavg_raid_md_nvidia_get_disk(struct g_raid_softc *sc, int id) 362219974Smav{ 363219974Smav struct g_raid_disk *disk; 364219974Smav struct g_raid_md_nvidia_perdisk *pd; 365219974Smav 366219974Smav TAILQ_FOREACH(disk, &sc->sc_disks, d_next) { 367219974Smav pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data; 368219974Smav if (pd->pd_disk_pos == id) 369219974Smav break; 370219974Smav } 371219974Smav return (disk); 372219974Smav} 373219974Smav 374219974Smavstatic int 375219974Smavg_raid_md_nvidia_supported(int level, int qual, int disks, int force) 376219974Smav{ 377219974Smav 378219974Smav switch (level) { 379219974Smav case G_RAID_VOLUME_RL_RAID0: 380219974Smav if (disks < 1) 381219974Smav return (0); 382219974Smav if (!force && (disks < 2 || disks > 6)) 383219974Smav return (0); 384219974Smav break; 385219974Smav case G_RAID_VOLUME_RL_RAID1: 386219974Smav if (disks < 1) 387219974Smav return (0); 388219974Smav if (!force && (disks != 2)) 389219974Smav return (0); 390219974Smav break; 391219974Smav case G_RAID_VOLUME_RL_RAID1E: 392219974Smav if (disks < 2) 393219974Smav return (0); 394219974Smav if (disks % 2 != 0) 395219974Smav return (0); 396219974Smav if (!force && (disks < 4)) 397219974Smav return (0); 398219974Smav break; 399219974Smav case G_RAID_VOLUME_RL_SINGLE: 400219974Smav if (disks != 1) 401219974Smav return (0); 402219974Smav break; 403219974Smav case G_RAID_VOLUME_RL_CONCAT: 404219974Smav if (disks < 2) 405219974Smav return (0); 406219974Smav break; 407219974Smav case G_RAID_VOLUME_RL_RAID5: 408219974Smav if (disks < 3) 409219974Smav return (0); 410219974Smav break; 411219974Smav default: 412219974Smav return (0); 413219974Smav } 414219974Smav if (qual != G_RAID_VOLUME_RLQ_NONE) 415219974Smav return (0); 416219974Smav return (1); 417219974Smav} 418219974Smav 419219974Smavstatic int 420219974Smavg_raid_md_nvidia_start_disk(struct g_raid_disk *disk) 421219974Smav{ 422219974Smav struct g_raid_softc *sc; 423219974Smav struct g_raid_subdisk *sd, *tmpsd; 424219974Smav struct g_raid_disk *olddisk, *tmpdisk; 425219974Smav struct g_raid_md_object *md; 426219974Smav struct g_raid_md_nvidia_object *mdi; 427219974Smav struct g_raid_md_nvidia_perdisk *pd, *oldpd; 428219974Smav struct nvidia_raid_conf *meta; 429219974Smav int disk_pos, resurrection = 0; 430219974Smav 431219974Smav sc = disk->d_softc; 432219974Smav md = sc->sc_md; 433219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 434219974Smav meta = mdi->mdio_meta; 435219974Smav pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data; 436219974Smav olddisk = NULL; 437219974Smav 438219974Smav /* Find disk position in metadata by it's serial. */ 439219974Smav if (pd->pd_meta != NULL) { 440219974Smav disk_pos = pd->pd_meta->disk_number; 441219974Smav if (disk_pos >= meta->total_disks || mdi->mdio_started) 442219974Smav disk_pos = -3; 443219974Smav } else 444219974Smav disk_pos = -3; 445219974Smav /* For RAID0+1 we need to translate order. */ 446219974Smav disk_pos = nvidia_meta_translate_disk(meta, disk_pos); 447219974Smav if (disk_pos < 0) { 448219974Smav G_RAID_DEBUG1(1, sc, "Unknown, probably new or stale disk"); 449219974Smav /* If we are in the start process, that's all for now. */ 450219974Smav if (!mdi->mdio_started) 451219974Smav goto nofit; 452219974Smav /* 453219974Smav * If we have already started - try to get use of the disk. 454219974Smav * Try to replace OFFLINE disks first, then FAILED. 455219974Smav */ 456219974Smav TAILQ_FOREACH(tmpdisk, &sc->sc_disks, d_next) { 457219974Smav if (tmpdisk->d_state != G_RAID_DISK_S_OFFLINE && 458219974Smav tmpdisk->d_state != G_RAID_DISK_S_FAILED) 459219974Smav continue; 460219974Smav /* Make sure this disk is big enough. */ 461219974Smav TAILQ_FOREACH(sd, &tmpdisk->d_subdisks, sd_next) { 462219974Smav if (sd->sd_offset + sd->sd_size + 2 * 512 > 463219974Smav pd->pd_disk_size) { 464219974Smav G_RAID_DEBUG1(1, sc, 465219974Smav "Disk too small (%ju < %ju)", 466219974Smav pd->pd_disk_size, 467219974Smav sd->sd_offset + sd->sd_size + 512); 468219974Smav break; 469219974Smav } 470219974Smav } 471219974Smav if (sd != NULL) 472219974Smav continue; 473219974Smav if (tmpdisk->d_state == G_RAID_DISK_S_OFFLINE) { 474219974Smav olddisk = tmpdisk; 475219974Smav break; 476219974Smav } else if (olddisk == NULL) 477219974Smav olddisk = tmpdisk; 478219974Smav } 479219974Smav if (olddisk == NULL) { 480219974Smavnofit: 481219974Smav g_raid_change_disk_state(disk, G_RAID_DISK_S_SPARE); 482219974Smav return (1); 483219974Smav } 484219974Smav oldpd = (struct g_raid_md_nvidia_perdisk *)olddisk->d_md_data; 485219974Smav disk_pos = oldpd->pd_disk_pos; 486219974Smav resurrection = 1; 487219974Smav } 488219974Smav 489219974Smav if (olddisk == NULL) { 490219974Smav /* Find placeholder by position. */ 491219974Smav olddisk = g_raid_md_nvidia_get_disk(sc, disk_pos); 492219974Smav if (olddisk == NULL) 493219974Smav panic("No disk at position %d!", disk_pos); 494219974Smav if (olddisk->d_state != G_RAID_DISK_S_OFFLINE) { 495219974Smav G_RAID_DEBUG1(1, sc, "More then one disk for pos %d", 496219974Smav disk_pos); 497219974Smav g_raid_change_disk_state(disk, G_RAID_DISK_S_STALE); 498219974Smav return (0); 499219974Smav } 500219974Smav oldpd = (struct g_raid_md_nvidia_perdisk *)olddisk->d_md_data; 501219974Smav } 502219974Smav 503219974Smav /* Replace failed disk or placeholder with new disk. */ 504219974Smav TAILQ_FOREACH_SAFE(sd, &olddisk->d_subdisks, sd_next, tmpsd) { 505219974Smav TAILQ_REMOVE(&olddisk->d_subdisks, sd, sd_next); 506219974Smav TAILQ_INSERT_TAIL(&disk->d_subdisks, sd, sd_next); 507219974Smav sd->sd_disk = disk; 508219974Smav } 509219974Smav oldpd->pd_disk_pos = -2; 510219974Smav pd->pd_disk_pos = disk_pos; 511219974Smav 512219974Smav /* If it was placeholder -- destroy it. */ 513219974Smav if (olddisk->d_state == G_RAID_DISK_S_OFFLINE) { 514219974Smav g_raid_destroy_disk(olddisk); 515219974Smav } else { 516219974Smav /* Otherwise, make it STALE_FAILED. */ 517219974Smav g_raid_change_disk_state(olddisk, G_RAID_DISK_S_STALE_FAILED); 518219974Smav } 519219974Smav 520219974Smav /* Welcome the new disk. */ 521219974Smav if (resurrection) 522219974Smav g_raid_change_disk_state(disk, G_RAID_DISK_S_ACTIVE); 523219974Smav else// if (pd->pd_meta->disk_status == NVIDIA_S_CURRENT || 524219974Smav //pd->pd_meta->disk_status == NVIDIA_S_REBUILD) 525219974Smav g_raid_change_disk_state(disk, G_RAID_DISK_S_ACTIVE); 526219974Smav// else 527219974Smav// g_raid_change_disk_state(disk, G_RAID_DISK_S_FAILED); 528219974Smav TAILQ_FOREACH(sd, &disk->d_subdisks, sd_next) { 529219974Smav 530219974Smav /* 531219974Smav * Different disks may have different sizes, 532219974Smav * in concat mode. Update from real disk size. 533219974Smav */ 534219974Smav if (meta->type == NVIDIA_T_CONCAT) 535219974Smav sd->sd_size = pd->pd_disk_size - 0x800 * 512; 536219974Smav 537219974Smav if (resurrection) { 538219974Smav /* New or ex-spare disk. */ 539219974Smav g_raid_change_subdisk_state(sd, 540219974Smav G_RAID_SUBDISK_S_NEW); 541219974Smav } else if (meta->state == NVIDIA_S_REBUILD && 542219974Smav (pd->pd_meta->disk_status & 0x100)) { 543219974Smav /* Rebuilding disk. */ 544219974Smav g_raid_change_subdisk_state(sd, 545219974Smav G_RAID_SUBDISK_S_REBUILD); 546219974Smav sd->sd_rebuild_pos = (off_t)pd->pd_meta->rebuild_lba / 547219974Smav meta->array_width * pd->pd_meta->sector_size; 548219974Smav } else if (meta->state == NVIDIA_S_SYNC) { 549219974Smav /* Resyncing/dirty disk. */ 550219974Smav g_raid_change_subdisk_state(sd, 551219974Smav G_RAID_SUBDISK_S_RESYNC); 552219974Smav sd->sd_rebuild_pos = (off_t)pd->pd_meta->rebuild_lba / 553219974Smav meta->array_width * pd->pd_meta->sector_size; 554219974Smav } else { 555219974Smav /* Up to date disk. */ 556219974Smav g_raid_change_subdisk_state(sd, 557219974Smav G_RAID_SUBDISK_S_ACTIVE); 558219974Smav } 559219974Smav g_raid_event_send(sd, G_RAID_SUBDISK_E_NEW, 560219974Smav G_RAID_EVENT_SUBDISK); 561219974Smav } 562219974Smav 563219974Smav /* Update status of our need for spare. */ 564219974Smav if (mdi->mdio_started) { 565219974Smav mdi->mdio_incomplete = 566219974Smav (g_raid_ndisks(sc, G_RAID_DISK_S_ACTIVE) < 567219974Smav mdi->mdio_total_disks); 568219974Smav } 569219974Smav 570219974Smav return (resurrection); 571219974Smav} 572219974Smav 573219974Smavstatic void 574219974Smavg_disk_md_nvidia_retaste(void *arg, int pending) 575219974Smav{ 576219974Smav 577219974Smav G_RAID_DEBUG(1, "Array is not complete, trying to retaste."); 578219974Smav g_retaste(&g_raid_class); 579219974Smav free(arg, M_MD_NVIDIA); 580219974Smav} 581219974Smav 582219974Smavstatic void 583219974Smavg_raid_md_nvidia_refill(struct g_raid_softc *sc) 584219974Smav{ 585219974Smav struct g_raid_md_object *md; 586219974Smav struct g_raid_md_nvidia_object *mdi; 587219974Smav struct nvidia_raid_conf *meta; 588219974Smav struct g_raid_disk *disk; 589219974Smav struct task *task; 590219974Smav int update, na; 591219974Smav 592219974Smav md = sc->sc_md; 593219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 594219974Smav meta = mdi->mdio_meta; 595219974Smav update = 0; 596219974Smav do { 597219974Smav /* Make sure we miss anything. */ 598219974Smav na = g_raid_ndisks(sc, G_RAID_DISK_S_ACTIVE); 599219974Smav if (na == mdi->mdio_total_disks) 600219974Smav break; 601219974Smav 602219974Smav G_RAID_DEBUG1(1, md->mdo_softc, 603219974Smav "Array is not complete (%d of %d), " 604219974Smav "trying to refill.", na, mdi->mdio_total_disks); 605219974Smav 606219974Smav /* Try to get use some of STALE disks. */ 607219974Smav TAILQ_FOREACH(disk, &sc->sc_disks, d_next) { 608219974Smav if (disk->d_state == G_RAID_DISK_S_STALE) { 609219974Smav update += g_raid_md_nvidia_start_disk(disk); 610219974Smav if (disk->d_state == G_RAID_DISK_S_ACTIVE) 611219974Smav break; 612219974Smav } 613219974Smav } 614219974Smav if (disk != NULL) 615219974Smav continue; 616219974Smav 617219974Smav /* Try to get use some of SPARE disks. */ 618219974Smav TAILQ_FOREACH(disk, &sc->sc_disks, d_next) { 619219974Smav if (disk->d_state == G_RAID_DISK_S_SPARE) { 620219974Smav update += g_raid_md_nvidia_start_disk(disk); 621219974Smav if (disk->d_state == G_RAID_DISK_S_ACTIVE) 622219974Smav break; 623219974Smav } 624219974Smav } 625219974Smav } while (disk != NULL); 626219974Smav 627219974Smav /* Write new metadata if we changed something. */ 628219974Smav if (update) { 629219974Smav g_raid_md_write_nvidia(md, NULL, NULL, NULL); 630219974Smav meta = mdi->mdio_meta; 631219974Smav } 632219974Smav 633219974Smav /* Update status of our need for spare. */ 634219974Smav mdi->mdio_incomplete = (g_raid_ndisks(sc, G_RAID_DISK_S_ACTIVE) < 635219974Smav mdi->mdio_total_disks); 636219974Smav 637219974Smav /* Request retaste hoping to find spare. */ 638219974Smav if (mdi->mdio_incomplete) { 639219974Smav task = malloc(sizeof(struct task), 640219974Smav M_MD_NVIDIA, M_WAITOK | M_ZERO); 641219974Smav TASK_INIT(task, 0, g_disk_md_nvidia_retaste, task); 642219974Smav taskqueue_enqueue(taskqueue_swi, task); 643219974Smav } 644219974Smav} 645219974Smav 646219974Smavstatic void 647219974Smavg_raid_md_nvidia_start(struct g_raid_softc *sc) 648219974Smav{ 649219974Smav struct g_raid_md_object *md; 650219974Smav struct g_raid_md_nvidia_object *mdi; 651219974Smav struct g_raid_md_nvidia_perdisk *pd; 652219974Smav struct nvidia_raid_conf *meta; 653219974Smav struct g_raid_volume *vol; 654219974Smav struct g_raid_subdisk *sd; 655219974Smav struct g_raid_disk *disk; 656219974Smav off_t size; 657219974Smav int j, disk_pos; 658219974Smav char buf[17]; 659219974Smav 660219974Smav md = sc->sc_md; 661219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 662219974Smav meta = mdi->mdio_meta; 663219974Smav 664219974Smav /* Create volumes and subdisks. */ 665219974Smav nvidia_meta_get_name(meta, buf); 666219974Smav vol = g_raid_create_volume(sc, buf, -1); 667219974Smav vol->v_mediasize = (off_t)meta->total_sectors * 512; 668219974Smav vol->v_raid_level_qualifier = G_RAID_VOLUME_RLQ_NONE; 669219974Smav if (meta->type == NVIDIA_T_RAID0) { 670219974Smav vol->v_raid_level = G_RAID_VOLUME_RL_RAID0; 671219974Smav size = vol->v_mediasize / mdi->mdio_total_disks; 672219974Smav } else if (meta->type == NVIDIA_T_RAID1) { 673219974Smav vol->v_raid_level = G_RAID_VOLUME_RL_RAID1; 674219974Smav size = vol->v_mediasize; 675219974Smav } else if (meta->type == NVIDIA_T_RAID01) { 676219974Smav vol->v_raid_level = G_RAID_VOLUME_RL_RAID1E; 677219974Smav size = vol->v_mediasize / (mdi->mdio_total_disks / 2); 678219974Smav } else if (meta->type == NVIDIA_T_CONCAT) { 679219974Smav if (mdi->mdio_total_disks == 1) 680219974Smav vol->v_raid_level = G_RAID_VOLUME_RL_SINGLE; 681219974Smav else 682219974Smav vol->v_raid_level = G_RAID_VOLUME_RL_CONCAT; 683219974Smav size = 0; 684219974Smav } else if (meta->type == NVIDIA_T_RAID5) { 685219974Smav vol->v_raid_level = G_RAID_VOLUME_RL_RAID5; 686219974Smav size = vol->v_mediasize / (mdi->mdio_total_disks - 1); 687219974Smav } else if (meta->type == NVIDIA_T_RAID5_SYM) { 688219974Smav vol->v_raid_level = G_RAID_VOLUME_RL_RAID5; 689219974Smav// vol->v_raid_level_qualifier = 0x03; 690219974Smav size = vol->v_mediasize / (mdi->mdio_total_disks - 1); 691219974Smav } else { 692219974Smav vol->v_raid_level = G_RAID_VOLUME_RL_UNKNOWN; 693219974Smav size = 0; 694219974Smav } 695219974Smav vol->v_strip_size = meta->strip_sectors * 512; //ZZZ 696219974Smav vol->v_disks_count = mdi->mdio_total_disks; 697219974Smav vol->v_sectorsize = 512; //ZZZ 698219974Smav for (j = 0; j < vol->v_disks_count; j++) { 699219974Smav sd = &vol->v_subdisks[j]; 700219974Smav sd->sd_offset = 0; 701219974Smav sd->sd_size = size; 702219974Smav } 703219974Smav g_raid_start_volume(vol); 704219974Smav 705219974Smav /* Create disk placeholders to store data for later writing. */ 706219974Smav for (disk_pos = 0; disk_pos < mdi->mdio_total_disks; disk_pos++) { 707219974Smav pd = malloc(sizeof(*pd), M_MD_NVIDIA, M_WAITOK | M_ZERO); 708219974Smav pd->pd_disk_pos = disk_pos; 709219974Smav disk = g_raid_create_disk(sc); 710219974Smav disk->d_md_data = (void *)pd; 711219974Smav disk->d_state = G_RAID_DISK_S_OFFLINE; 712219974Smav sd = &vol->v_subdisks[disk_pos]; 713219974Smav sd->sd_disk = disk; 714219974Smav TAILQ_INSERT_TAIL(&disk->d_subdisks, sd, sd_next); 715219974Smav } 716219974Smav 717219974Smav /* Make all disks found till the moment take their places. */ 718219974Smav do { 719219974Smav TAILQ_FOREACH(disk, &sc->sc_disks, d_next) { 720219974Smav if (disk->d_state == G_RAID_DISK_S_NONE) { 721219974Smav g_raid_md_nvidia_start_disk(disk); 722219974Smav break; 723219974Smav } 724219974Smav } 725219974Smav } while (disk != NULL); 726219974Smav 727219974Smav mdi->mdio_started = 1; 728219974Smav G_RAID_DEBUG1(0, sc, "Array started."); 729219974Smav g_raid_md_write_nvidia(md, NULL, NULL, NULL); 730219974Smav 731219974Smav /* Pickup any STALE/SPARE disks to refill array if needed. */ 732219974Smav g_raid_md_nvidia_refill(sc); 733219974Smav 734219974Smav g_raid_event_send(vol, G_RAID_VOLUME_E_START, G_RAID_EVENT_VOLUME); 735219974Smav 736219974Smav callout_stop(&mdi->mdio_start_co); 737219974Smav G_RAID_DEBUG1(1, sc, "root_mount_rel %p", mdi->mdio_rootmount); 738219974Smav root_mount_rel(mdi->mdio_rootmount); 739219974Smav mdi->mdio_rootmount = NULL; 740219974Smav} 741219974Smav 742219974Smavstatic void 743219974Smavg_raid_md_nvidia_new_disk(struct g_raid_disk *disk) 744219974Smav{ 745219974Smav struct g_raid_softc *sc; 746219974Smav struct g_raid_md_object *md; 747219974Smav struct g_raid_md_nvidia_object *mdi; 748219974Smav struct nvidia_raid_conf *pdmeta; 749219974Smav struct g_raid_md_nvidia_perdisk *pd; 750219974Smav 751219974Smav sc = disk->d_softc; 752219974Smav md = sc->sc_md; 753219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 754219974Smav pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data; 755219974Smav pdmeta = pd->pd_meta; 756219974Smav 757219974Smav if (mdi->mdio_started) { 758219974Smav if (g_raid_md_nvidia_start_disk(disk)) 759219974Smav g_raid_md_write_nvidia(md, NULL, NULL, NULL); 760219974Smav } else { 761219974Smav if (mdi->mdio_meta == NULL || 762219974Smav mdi->mdio_meta->disk_number >= mdi->mdio_meta->total_disks) { 763219974Smav G_RAID_DEBUG1(1, sc, "Newer disk"); 764219974Smav if (mdi->mdio_meta != NULL) 765219974Smav free(mdi->mdio_meta, M_MD_NVIDIA); 766219974Smav mdi->mdio_meta = nvidia_meta_copy(pdmeta); 767219974Smav mdi->mdio_total_disks = pdmeta->total_disks; 768219974Smav mdi->mdio_disks_present = 1; 769219974Smav } else if (pdmeta->disk_number < mdi->mdio_meta->total_disks) { 770219974Smav mdi->mdio_disks_present++; 771219974Smav G_RAID_DEBUG1(1, sc, "Matching disk (%d of %d up)", 772219974Smav mdi->mdio_disks_present, 773219974Smav mdi->mdio_total_disks); 774219974Smav } else 775219974Smav G_RAID_DEBUG1(1, sc, "Spare disk"); 776219974Smav 777219974Smav /* If we collected all needed disks - start array. */ 778219974Smav if (mdi->mdio_disks_present == mdi->mdio_total_disks) 779219974Smav g_raid_md_nvidia_start(sc); 780219974Smav } 781219974Smav} 782219974Smav 783219974Smavstatic void 784219974Smavg_raid_nvidia_go(void *arg) 785219974Smav{ 786219974Smav struct g_raid_softc *sc; 787219974Smav struct g_raid_md_object *md; 788219974Smav struct g_raid_md_nvidia_object *mdi; 789219974Smav 790219974Smav sc = arg; 791219974Smav md = sc->sc_md; 792219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 793219974Smav if (!mdi->mdio_started) { 794219974Smav G_RAID_DEBUG1(0, sc, "Force array start due to timeout."); 795219974Smav g_raid_event_send(sc, G_RAID_NODE_E_START, 0); 796219974Smav } 797219974Smav} 798219974Smav 799219974Smavstatic int 800219974Smavg_raid_md_create_nvidia(struct g_raid_md_object *md, struct g_class *mp, 801219974Smav struct g_geom **gp) 802219974Smav{ 803219974Smav struct g_raid_softc *sc; 804219974Smav struct g_raid_md_nvidia_object *mdi; 805219974Smav char name[32]; 806219974Smav 807219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 808219974Smav arc4rand(&mdi->mdio_volume_id, 16, 0); 809219974Smav snprintf(name, sizeof(name), "NVIDIA-%d", 810219974Smav atomic_fetchadd_int(&NVIDIANodeID, 1)); 811219974Smav sc = g_raid_create_node(mp, name, md); 812219974Smav if (sc == NULL) 813219974Smav return (G_RAID_MD_TASTE_FAIL); 814219974Smav md->mdo_softc = sc; 815219974Smav *gp = sc->sc_geom; 816219974Smav return (G_RAID_MD_TASTE_NEW); 817219974Smav} 818219974Smav 819219974Smavstatic int 820219974Smavg_raid_md_taste_nvidia(struct g_raid_md_object *md, struct g_class *mp, 821219974Smav struct g_consumer *cp, struct g_geom **gp) 822219974Smav{ 823219974Smav struct g_consumer *rcp; 824219974Smav struct g_provider *pp; 825219974Smav struct g_raid_md_nvidia_object *mdi, *mdi1; 826219974Smav struct g_raid_softc *sc; 827219974Smav struct g_raid_disk *disk; 828219974Smav struct nvidia_raid_conf *meta; 829219974Smav struct g_raid_md_nvidia_perdisk *pd; 830219974Smav struct g_geom *geom; 831219974Smav int error, disk_pos, result, spare, len; 832219974Smav char name[32]; 833219974Smav uint16_t vendor; 834219974Smav 835219974Smav G_RAID_DEBUG(1, "Tasting NVIDIA on %s", cp->provider->name); 836219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 837219974Smav pp = cp->provider; 838219974Smav 839219974Smav /* Read metadata from device. */ 840219974Smav meta = NULL; 841219974Smav spare = 0; 842219974Smav vendor = 0xffff; 843219974Smav disk_pos = 0; 844219974Smav if (g_access(cp, 1, 0, 0) != 0) 845219974Smav return (G_RAID_MD_TASTE_FAIL); 846219974Smav g_topology_unlock(); 847219974Smav len = 2; 848219974Smav if (pp->geom->rank == 1) 849219974Smav g_io_getattr("GEOM::hba_vendor", cp, &len, &vendor); 850219974Smav meta = nvidia_meta_read(cp); 851219974Smav g_topology_lock(); 852219974Smav g_access(cp, -1, 0, 0); 853219974Smav if (meta == NULL) { 854219974Smav if (g_raid_aggressive_spare) { 855219974Smav if (vendor == 0x10de) { 856219974Smav G_RAID_DEBUG(1, 857219974Smav "No NVIDIA metadata, forcing spare."); 858219974Smav spare = 2; 859219974Smav goto search; 860219974Smav } else { 861219974Smav G_RAID_DEBUG(1, 862219974Smav "NVIDIA vendor mismatch 0x%04x != 0x10de", 863219974Smav vendor); 864219974Smav } 865219974Smav } 866219974Smav return (G_RAID_MD_TASTE_FAIL); 867219974Smav } 868219974Smav 869219974Smav /* Check this disk position in obtained metadata. */ 870219974Smav disk_pos = meta->disk_number; 871219974Smav if (disk_pos == -1) { 872219974Smav G_RAID_DEBUG(1, "NVIDIA disk position not found"); 873219974Smav goto fail1; 874219974Smav } 875219974Smav 876219974Smav /* Metadata valid. Print it. */ 877219974Smav g_raid_md_nvidia_print(meta); 878219974Smav G_RAID_DEBUG(1, "NVIDIA disk position %d", disk_pos); 879219974Smav spare = 0;//(meta->type == NVIDIA_T_SPARE) ? 1 : 0; 880219974Smav 881219974Smavsearch: 882219974Smav /* Search for matching node. */ 883219974Smav sc = NULL; 884219974Smav mdi1 = NULL; 885219974Smav LIST_FOREACH(geom, &mp->geom, geom) { 886219974Smav sc = geom->softc; 887219974Smav if (sc == NULL) 888219974Smav continue; 889219974Smav if (sc->sc_stopping != 0) 890219974Smav continue; 891219974Smav if (sc->sc_md->mdo_class != md->mdo_class) 892219974Smav continue; 893219974Smav mdi1 = (struct g_raid_md_nvidia_object *)sc->sc_md; 894219974Smav if (spare) { 895219974Smav if (mdi1->mdio_incomplete) 896219974Smav break; 897219974Smav } else { 898219974Smav if (memcmp(&mdi1->mdio_volume_id, 899219974Smav &meta->volume_id, 16) == 0) 900219974Smav break; 901219974Smav } 902219974Smav } 903219974Smav 904219974Smav /* Found matching node. */ 905219974Smav if (geom != NULL) { 906219974Smav G_RAID_DEBUG(1, "Found matching array %s", sc->sc_name); 907219974Smav result = G_RAID_MD_TASTE_EXISTING; 908219974Smav 909219974Smav } else if (spare) { /* Not found needy node -- left for later. */ 910219974Smav G_RAID_DEBUG(1, "Spare is not needed at this time"); 911219974Smav goto fail1; 912219974Smav 913219974Smav } else { /* Not found matching node -- create one. */ 914219974Smav result = G_RAID_MD_TASTE_NEW; 915219974Smav memcpy(&mdi->mdio_volume_id, &meta->volume_id, 16); 916219974Smav snprintf(name, sizeof(name), "NVIDIA-%d", 917219974Smav atomic_fetchadd_int(&NVIDIANodeID, 1)); 918219974Smav sc = g_raid_create_node(mp, name, md); 919219974Smav md->mdo_softc = sc; 920219974Smav geom = sc->sc_geom; 921219974Smav callout_init(&mdi->mdio_start_co, 1); 922219974Smav callout_reset(&mdi->mdio_start_co, g_raid_start_timeout * hz, 923219974Smav g_raid_nvidia_go, sc); 924219974Smav mdi->mdio_rootmount = root_mount_hold("GRAID-NVIDIA"); 925219974Smav G_RAID_DEBUG1(1, sc, "root_mount_hold %p", mdi->mdio_rootmount); 926219974Smav } 927219974Smav 928219974Smav rcp = g_new_consumer(geom); 929219974Smav g_attach(rcp, pp); 930219974Smav if (g_access(rcp, 1, 1, 1) != 0) 931219974Smav ; //goto fail1; 932219974Smav 933219974Smav g_topology_unlock(); 934219974Smav sx_xlock(&sc->sc_lock); 935219974Smav 936219974Smav pd = malloc(sizeof(*pd), M_MD_NVIDIA, M_WAITOK | M_ZERO); 937219974Smav pd->pd_meta = meta; 938219974Smav if (spare == 2) { 939219974Smav pd->pd_disk_pos = -3; 940219974Smav } else { 941219974Smav pd->pd_disk_pos = -1; 942219974Smav } 943219974Smav pd->pd_disk_size = pp->mediasize; 944219974Smav disk = g_raid_create_disk(sc); 945219974Smav disk->d_md_data = (void *)pd; 946219974Smav disk->d_consumer = rcp; 947219974Smav rcp->private = disk; 948219974Smav 949219974Smav /* Read kernel dumping information. */ 950219974Smav disk->d_kd.offset = 0; 951219974Smav disk->d_kd.length = OFF_MAX; 952219974Smav len = sizeof(disk->d_kd); 953219974Smav error = g_io_getattr("GEOM::kerneldump", rcp, &len, &disk->d_kd); 954219974Smav if (disk->d_kd.di.dumper == NULL) 955219974Smav G_RAID_DEBUG1(2, sc, "Dumping not supported by %s: %d.", 956219974Smav rcp->provider->name, error); 957219974Smav 958219974Smav g_raid_md_nvidia_new_disk(disk); 959219974Smav 960219974Smav sx_xunlock(&sc->sc_lock); 961219974Smav g_topology_lock(); 962219974Smav *gp = geom; 963219974Smav return (result); 964219974Smavfail1: 965219974Smav free(meta, M_MD_NVIDIA); 966219974Smav return (G_RAID_MD_TASTE_FAIL); 967219974Smav} 968219974Smav 969219974Smavstatic int 970219974Smavg_raid_md_event_nvidia(struct g_raid_md_object *md, 971219974Smav struct g_raid_disk *disk, u_int event) 972219974Smav{ 973219974Smav struct g_raid_softc *sc; 974219974Smav struct g_raid_subdisk *sd; 975219974Smav struct g_raid_md_nvidia_object *mdi; 976219974Smav struct g_raid_md_nvidia_perdisk *pd; 977219974Smav 978219974Smav sc = md->mdo_softc; 979219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 980219974Smav if (disk == NULL) { 981219974Smav switch (event) { 982219974Smav case G_RAID_NODE_E_START: 983219974Smav if (!mdi->mdio_started) { 984219974Smav /* Bump volume ID to drop missing disks. */ 985219974Smav arc4rand(&mdi->mdio_volume_id, 16, 0); 986219974Smav g_raid_md_nvidia_start(sc); 987219974Smav } 988219974Smav return (0); 989219974Smav } 990219974Smav return (-1); 991219974Smav } 992219974Smav pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data; 993219974Smav switch (event) { 994219974Smav case G_RAID_DISK_E_DISCONNECTED: 995219974Smav /* If disk was assigned, just update statuses. */ 996219974Smav if (pd->pd_disk_pos >= 0) { 997219974Smav g_raid_change_disk_state(disk, G_RAID_DISK_S_OFFLINE); 998219974Smav if (disk->d_consumer) { 999219974Smav g_raid_kill_consumer(sc, disk->d_consumer); 1000219974Smav disk->d_consumer = NULL; 1001219974Smav } 1002219974Smav TAILQ_FOREACH(sd, &disk->d_subdisks, sd_next) { 1003219974Smav g_raid_change_subdisk_state(sd, 1004219974Smav G_RAID_SUBDISK_S_NONE); 1005219974Smav g_raid_event_send(sd, G_RAID_SUBDISK_E_DISCONNECTED, 1006219974Smav G_RAID_EVENT_SUBDISK); 1007219974Smav } 1008219974Smav } else { 1009219974Smav /* Otherwise -- delete. */ 1010219974Smav g_raid_change_disk_state(disk, G_RAID_DISK_S_NONE); 1011219974Smav g_raid_destroy_disk(disk); 1012219974Smav } 1013219974Smav 1014219974Smav if (mdi->mdio_started) { 1015219974Smav /* Bump volume ID to prevent disk resurrection. */ 1016219974Smav if (pd->pd_disk_pos >= 0) 1017219974Smav arc4rand(&mdi->mdio_volume_id, 16, 0); 1018219974Smav 1019219974Smav /* Write updated metadata to all disks. */ 1020219974Smav g_raid_md_write_nvidia(md, NULL, NULL, NULL); 1021219974Smav } 1022219974Smav 1023219974Smav /* Check if anything left except placeholders. */ 1024219974Smav if (g_raid_ndisks(sc, -1) == 1025219974Smav g_raid_ndisks(sc, G_RAID_DISK_S_OFFLINE)) 1026219974Smav g_raid_destroy_node(sc, 0); 1027219974Smav else 1028219974Smav g_raid_md_nvidia_refill(sc); 1029219974Smav return (0); 1030219974Smav } 1031219974Smav return (-2); 1032219974Smav} 1033219974Smav 1034219974Smavstatic int 1035219974Smavg_raid_md_ctl_nvidia(struct g_raid_md_object *md, 1036219974Smav struct gctl_req *req) 1037219974Smav{ 1038219974Smav struct g_raid_softc *sc; 1039219974Smav struct g_raid_volume *vol; 1040219974Smav struct g_raid_subdisk *sd; 1041219974Smav struct g_raid_disk *disk; 1042219974Smav struct g_raid_md_nvidia_object *mdi; 1043219974Smav struct g_raid_md_nvidia_perdisk *pd; 1044219974Smav struct g_consumer *cp; 1045219974Smav struct g_provider *pp; 1046219974Smav char arg[16]; 1047219974Smav const char *verb, *volname, *levelname, *diskname; 1048219974Smav int *nargs, *force; 1049219974Smav off_t size, sectorsize, strip; 1050219974Smav intmax_t *sizearg, *striparg; 1051219974Smav int numdisks, i, len, level, qual, update; 1052219974Smav int error; 1053219974Smav 1054219974Smav sc = md->mdo_softc; 1055219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 1056219974Smav verb = gctl_get_param(req, "verb", NULL); 1057219974Smav nargs = gctl_get_paraml(req, "nargs", sizeof(*nargs)); 1058219974Smav error = 0; 1059219974Smav if (strcmp(verb, "label") == 0) { 1060219974Smav 1061219974Smav if (*nargs < 4) { 1062219974Smav gctl_error(req, "Invalid number of arguments."); 1063219974Smav return (-1); 1064219974Smav } 1065219974Smav volname = gctl_get_asciiparam(req, "arg1"); 1066219974Smav if (volname == NULL) { 1067219974Smav gctl_error(req, "No volume name."); 1068219974Smav return (-2); 1069219974Smav } 1070219974Smav levelname = gctl_get_asciiparam(req, "arg2"); 1071219974Smav if (levelname == NULL) { 1072219974Smav gctl_error(req, "No RAID level."); 1073219974Smav return (-3); 1074219974Smav } 1075219974Smav if (g_raid_volume_str2level(levelname, &level, &qual)) { 1076219974Smav gctl_error(req, "Unknown RAID level '%s'.", levelname); 1077219974Smav return (-4); 1078219974Smav } 1079219974Smav numdisks = *nargs - 3; 1080219974Smav force = gctl_get_paraml(req, "force", sizeof(*force)); 1081219974Smav if (!g_raid_md_nvidia_supported(level, qual, numdisks, 1082219974Smav force ? *force : 0)) { 1083219974Smav gctl_error(req, "Unsupported RAID level " 1084219974Smav "(0x%02x/0x%02x), or number of disks (%d).", 1085219974Smav level, qual, numdisks); 1086219974Smav return (-5); 1087219974Smav } 1088219974Smav 1089219974Smav /* Search for disks, connect them and probe. */ 1090219974Smav size = 0x7fffffffffffffffllu; 1091219974Smav sectorsize = 0; 1092219974Smav for (i = 0; i < numdisks; i++) { 1093219974Smav snprintf(arg, sizeof(arg), "arg%d", i + 3); 1094219974Smav diskname = gctl_get_asciiparam(req, arg); 1095219974Smav if (diskname == NULL) { 1096219974Smav gctl_error(req, "No disk name (%s).", arg); 1097219974Smav error = -6; 1098219974Smav break; 1099219974Smav } 1100219974Smav if (strcmp(diskname, "NONE") == 0) { 1101219974Smav cp = NULL; 1102219974Smav pp = NULL; 1103219974Smav } else { 1104219974Smav g_topology_lock(); 1105219974Smav cp = g_raid_open_consumer(sc, diskname); 1106219974Smav if (cp == NULL) { 1107219974Smav gctl_error(req, "Can't open '%s'.", 1108219974Smav diskname); 1109219974Smav g_topology_unlock(); 1110219974Smav error = -7; 1111219974Smav break; 1112219974Smav } 1113219974Smav pp = cp->provider; 1114219974Smav } 1115219974Smav pd = malloc(sizeof(*pd), M_MD_NVIDIA, M_WAITOK | M_ZERO); 1116219974Smav pd->pd_disk_pos = i; 1117219974Smav disk = g_raid_create_disk(sc); 1118219974Smav disk->d_md_data = (void *)pd; 1119219974Smav disk->d_consumer = cp; 1120219974Smav if (cp == NULL) 1121219974Smav continue; 1122219974Smav cp->private = disk; 1123219974Smav g_topology_unlock(); 1124219974Smav 1125219974Smav /* Read kernel dumping information. */ 1126219974Smav disk->d_kd.offset = 0; 1127219974Smav disk->d_kd.length = OFF_MAX; 1128219974Smav len = sizeof(disk->d_kd); 1129219974Smav g_io_getattr("GEOM::kerneldump", cp, &len, &disk->d_kd); 1130219974Smav if (disk->d_kd.di.dumper == NULL) 1131219974Smav G_RAID_DEBUG1(2, sc, 1132219974Smav "Dumping not supported by %s.", 1133219974Smav cp->provider->name); 1134219974Smav 1135219974Smav pd->pd_disk_size = pp->mediasize; 1136219974Smav if (size > pp->mediasize) 1137219974Smav size = pp->mediasize; 1138219974Smav if (sectorsize < pp->sectorsize) 1139219974Smav sectorsize = pp->sectorsize; 1140219974Smav } 1141219974Smav if (error != 0) 1142219974Smav return (error); 1143219974Smav 1144219974Smav /* Reserve space for metadata. */ 1145219974Smav size -= 2 * sectorsize; 1146219974Smav 1147219974Smav /* Handle size argument. */ 1148219974Smav len = sizeof(*sizearg); 1149219974Smav sizearg = gctl_get_param(req, "size", &len); 1150219974Smav if (sizearg != NULL && len == sizeof(*sizearg) && 1151219974Smav *sizearg > 0) { 1152219974Smav if (*sizearg > size) { 1153219974Smav gctl_error(req, "Size too big %lld > %lld.", 1154219974Smav (long long)*sizearg, (long long)size); 1155219974Smav return (-9); 1156219974Smav } 1157219974Smav size = *sizearg; 1158219974Smav } 1159219974Smav 1160219974Smav /* Handle strip argument. */ 1161219974Smav strip = 131072; 1162219974Smav len = sizeof(*striparg); 1163219974Smav striparg = gctl_get_param(req, "strip", &len); 1164219974Smav if (striparg != NULL && len == sizeof(*striparg) && 1165219974Smav *striparg > 0) { 1166219974Smav if (*striparg < sectorsize) { 1167219974Smav gctl_error(req, "Strip size too small."); 1168219974Smav return (-10); 1169219974Smav } 1170219974Smav if (*striparg % sectorsize != 0) { 1171219974Smav gctl_error(req, "Incorrect strip size."); 1172219974Smav return (-11); 1173219974Smav } 1174219974Smav if (strip > 65535 * sectorsize) { 1175219974Smav gctl_error(req, "Strip size too big."); 1176219974Smav return (-12); 1177219974Smav } 1178219974Smav strip = *striparg; 1179219974Smav } 1180219974Smav 1181219974Smav /* Round size down to strip or sector. */ 1182219974Smav if (level == G_RAID_VOLUME_RL_RAID1) 1183219974Smav size -= (size % sectorsize); 1184219974Smav else if (level == G_RAID_VOLUME_RL_RAID1E && 1185219974Smav (numdisks & 1) != 0) 1186219974Smav size -= (size % (2 * strip)); 1187219974Smav else 1188219974Smav size -= (size % strip); 1189219974Smav if (size <= 0) { 1190219974Smav gctl_error(req, "Size too small."); 1191219974Smav return (-13); 1192219974Smav } 1193219974Smav if (size > 0xffffffffffffllu * sectorsize) { 1194219974Smav gctl_error(req, "Size too big."); 1195219974Smav return (-14); 1196219974Smav } 1197219974Smav 1198219974Smav /* We have all we need, create things: volume, ... */ 1199219974Smav mdi->mdio_total_disks = numdisks; 1200219974Smav mdi->mdio_started = 1; 1201219974Smav vol = g_raid_create_volume(sc, volname, -1); 1202219974Smav vol->v_md_data = (void *)(intptr_t)0; 1203219974Smav vol->v_raid_level = level; 1204219974Smav vol->v_raid_level_qualifier = G_RAID_VOLUME_RLQ_NONE; 1205219974Smav vol->v_strip_size = strip; 1206219974Smav vol->v_disks_count = numdisks; 1207219974Smav if (level == G_RAID_VOLUME_RL_RAID0 || 1208219974Smav level == G_RAID_VOLUME_RL_CONCAT || 1209219974Smav level == G_RAID_VOLUME_RL_SINGLE) 1210219974Smav vol->v_mediasize = size * numdisks; 1211219974Smav else if (level == G_RAID_VOLUME_RL_RAID1) 1212219974Smav vol->v_mediasize = size; 1213219974Smav else if (level == G_RAID_VOLUME_RL_RAID5) 1214219974Smav vol->v_mediasize = size * (numdisks - 1); 1215219974Smav else { /* RAID1E */ 1216219974Smav vol->v_mediasize = ((size * numdisks) / strip / 2) * 1217219974Smav strip; 1218219974Smav } 1219219974Smav vol->v_sectorsize = sectorsize; 1220219974Smav g_raid_start_volume(vol); 1221219974Smav 1222219974Smav /* , and subdisks. */ 1223219974Smav TAILQ_FOREACH(disk, &sc->sc_disks, d_next) { 1224219974Smav pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data; 1225219974Smav sd = &vol->v_subdisks[pd->pd_disk_pos]; 1226219974Smav sd->sd_disk = disk; 1227219974Smav sd->sd_offset = 0; 1228219974Smav sd->sd_size = size; 1229219974Smav TAILQ_INSERT_TAIL(&disk->d_subdisks, sd, sd_next); 1230219974Smav if (sd->sd_disk->d_consumer != NULL) { 1231219974Smav g_raid_change_disk_state(disk, 1232219974Smav G_RAID_DISK_S_ACTIVE); 1233219974Smav g_raid_change_subdisk_state(sd, 1234219974Smav G_RAID_SUBDISK_S_ACTIVE); 1235219974Smav g_raid_event_send(sd, G_RAID_SUBDISK_E_NEW, 1236219974Smav G_RAID_EVENT_SUBDISK); 1237219974Smav } else { 1238219974Smav g_raid_change_disk_state(disk, G_RAID_DISK_S_OFFLINE); 1239219974Smav } 1240219974Smav } 1241219974Smav 1242219974Smav /* Write metadata based on created entities. */ 1243219974Smav G_RAID_DEBUG1(0, sc, "Array started."); 1244219974Smav g_raid_md_write_nvidia(md, NULL, NULL, NULL); 1245219974Smav 1246219974Smav /* Pickup any STALE/SPARE disks to refill array if needed. */ 1247219974Smav g_raid_md_nvidia_refill(sc); 1248219974Smav 1249219974Smav g_raid_event_send(vol, G_RAID_VOLUME_E_START, 1250219974Smav G_RAID_EVENT_VOLUME); 1251219974Smav return (0); 1252219974Smav } 1253219974Smav if (strcmp(verb, "delete") == 0) { 1254219974Smav 1255219974Smav /* Check if some volume is still open. */ 1256219974Smav force = gctl_get_paraml(req, "force", sizeof(*force)); 1257219974Smav if (force != NULL && *force == 0 && 1258219974Smav g_raid_nopens(sc) != 0) { 1259219974Smav gctl_error(req, "Some volume is still open."); 1260219974Smav return (-4); 1261219974Smav } 1262219974Smav 1263219974Smav TAILQ_FOREACH(disk, &sc->sc_disks, d_next) { 1264219974Smav if (disk->d_consumer) 1265219974Smav nvidia_meta_erase(disk->d_consumer); 1266219974Smav } 1267219974Smav g_raid_destroy_node(sc, 0); 1268219974Smav return (0); 1269219974Smav } 1270219974Smav if (strcmp(verb, "remove") == 0 || 1271219974Smav strcmp(verb, "fail") == 0) { 1272219974Smav if (*nargs < 2) { 1273219974Smav gctl_error(req, "Invalid number of arguments."); 1274219974Smav return (-1); 1275219974Smav } 1276219974Smav for (i = 1; i < *nargs; i++) { 1277219974Smav snprintf(arg, sizeof(arg), "arg%d", i); 1278219974Smav diskname = gctl_get_asciiparam(req, arg); 1279219974Smav if (diskname == NULL) { 1280219974Smav gctl_error(req, "No disk name (%s).", arg); 1281219974Smav error = -2; 1282219974Smav break; 1283219974Smav } 1284219974Smav if (strncmp(diskname, "/dev/", 5) == 0) 1285219974Smav diskname += 5; 1286219974Smav 1287219974Smav TAILQ_FOREACH(disk, &sc->sc_disks, d_next) { 1288219974Smav if (disk->d_consumer != NULL && 1289219974Smav disk->d_consumer->provider != NULL && 1290219974Smav strcmp(disk->d_consumer->provider->name, 1291219974Smav diskname) == 0) 1292219974Smav break; 1293219974Smav } 1294219974Smav if (disk == NULL) { 1295219974Smav gctl_error(req, "Disk '%s' not found.", 1296219974Smav diskname); 1297219974Smav error = -3; 1298219974Smav break; 1299219974Smav } 1300219974Smav 1301219974Smav if (strcmp(verb, "fail") == 0) { 1302219974Smav g_raid_md_fail_disk_nvidia(md, NULL, disk); 1303219974Smav continue; 1304219974Smav } 1305219974Smav 1306219974Smav pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data; 1307219974Smav 1308219974Smav /* Erase metadata on deleting disk. */ 1309219974Smav nvidia_meta_erase(disk->d_consumer); 1310219974Smav 1311219974Smav /* If disk was assigned, just update statuses. */ 1312219974Smav if (pd->pd_disk_pos >= 0) { 1313219974Smav g_raid_change_disk_state(disk, G_RAID_DISK_S_OFFLINE); 1314220209Smav g_raid_kill_consumer(sc, disk->d_consumer); 1315220209Smav disk->d_consumer = NULL; 1316219974Smav TAILQ_FOREACH(sd, &disk->d_subdisks, sd_next) { 1317219974Smav g_raid_change_subdisk_state(sd, 1318219974Smav G_RAID_SUBDISK_S_NONE); 1319219974Smav g_raid_event_send(sd, G_RAID_SUBDISK_E_DISCONNECTED, 1320219974Smav G_RAID_EVENT_SUBDISK); 1321219974Smav } 1322219974Smav } else { 1323219974Smav /* Otherwise -- delete. */ 1324219974Smav g_raid_change_disk_state(disk, G_RAID_DISK_S_NONE); 1325219974Smav g_raid_destroy_disk(disk); 1326219974Smav } 1327219974Smav } 1328219974Smav 1329219974Smav /* Write updated metadata to remaining disks. */ 1330219974Smav g_raid_md_write_nvidia(md, NULL, NULL, NULL); 1331219974Smav 1332219974Smav /* Check if anything left except placeholders. */ 1333219974Smav if (g_raid_ndisks(sc, -1) == 1334219974Smav g_raid_ndisks(sc, G_RAID_DISK_S_OFFLINE)) 1335219974Smav g_raid_destroy_node(sc, 0); 1336219974Smav else 1337219974Smav g_raid_md_nvidia_refill(sc); 1338219974Smav return (error); 1339219974Smav } 1340219974Smav if (strcmp(verb, "insert") == 0) { 1341219974Smav if (*nargs < 2) { 1342219974Smav gctl_error(req, "Invalid number of arguments."); 1343219974Smav return (-1); 1344219974Smav } 1345219974Smav update = 0; 1346219974Smav for (i = 1; i < *nargs; i++) { 1347219974Smav /* Get disk name. */ 1348219974Smav snprintf(arg, sizeof(arg), "arg%d", i); 1349219974Smav diskname = gctl_get_asciiparam(req, arg); 1350219974Smav if (diskname == NULL) { 1351219974Smav gctl_error(req, "No disk name (%s).", arg); 1352219974Smav error = -3; 1353219974Smav break; 1354219974Smav } 1355219974Smav 1356219974Smav /* Try to find provider with specified name. */ 1357219974Smav g_topology_lock(); 1358219974Smav cp = g_raid_open_consumer(sc, diskname); 1359219974Smav if (cp == NULL) { 1360219974Smav gctl_error(req, "Can't open disk '%s'.", 1361219974Smav diskname); 1362219974Smav g_topology_unlock(); 1363219974Smav error = -4; 1364219974Smav break; 1365219974Smav } 1366219974Smav pp = cp->provider; 1367219974Smav 1368219974Smav pd = malloc(sizeof(*pd), M_MD_NVIDIA, M_WAITOK | M_ZERO); 1369219974Smav pd->pd_disk_pos = -3; 1370219974Smav pd->pd_disk_size = pp->mediasize; 1371219974Smav 1372219974Smav disk = g_raid_create_disk(sc); 1373219974Smav disk->d_consumer = cp; 1374219974Smav disk->d_consumer->private = disk; 1375219974Smav disk->d_md_data = (void *)pd; 1376219974Smav cp->private = disk; 1377219974Smav g_topology_unlock(); 1378219974Smav 1379219974Smav /* Read kernel dumping information. */ 1380219974Smav disk->d_kd.offset = 0; 1381219974Smav disk->d_kd.length = OFF_MAX; 1382219974Smav len = sizeof(disk->d_kd); 1383219974Smav g_io_getattr("GEOM::kerneldump", cp, &len, &disk->d_kd); 1384219974Smav if (disk->d_kd.di.dumper == NULL) 1385219974Smav G_RAID_DEBUG1(2, sc, 1386219974Smav "Dumping not supported by %s.", 1387219974Smav cp->provider->name); 1388219974Smav 1389219974Smav /* Welcome the "new" disk. */ 1390219974Smav update += g_raid_md_nvidia_start_disk(disk); 1391219974Smav if (disk->d_state != G_RAID_DISK_S_SPARE && 1392219974Smav disk->d_state != G_RAID_DISK_S_ACTIVE) { 1393219974Smav gctl_error(req, "Disk '%s' doesn't fit.", 1394219974Smav diskname); 1395219974Smav g_raid_destroy_disk(disk); 1396219974Smav error = -8; 1397219974Smav break; 1398219974Smav } 1399219974Smav } 1400219974Smav 1401219974Smav /* Write new metadata if we changed something. */ 1402219974Smav if (update) 1403219974Smav g_raid_md_write_nvidia(md, NULL, NULL, NULL); 1404219974Smav return (error); 1405219974Smav } 1406219974Smav gctl_error(req, "Command '%s' is not supported.", verb); 1407219974Smav return (-100); 1408219974Smav} 1409219974Smav 1410219974Smavstatic int 1411219974Smavg_raid_md_write_nvidia(struct g_raid_md_object *md, struct g_raid_volume *tvol, 1412219974Smav struct g_raid_subdisk *tsd, struct g_raid_disk *tdisk) 1413219974Smav{ 1414219974Smav struct g_raid_softc *sc; 1415219974Smav struct g_raid_volume *vol; 1416219974Smav struct g_raid_subdisk *sd; 1417219974Smav struct g_raid_disk *disk; 1418219974Smav struct g_raid_md_nvidia_object *mdi; 1419219974Smav struct g_raid_md_nvidia_perdisk *pd; 1420219974Smav struct nvidia_raid_conf *meta; 1421219974Smav int i, spares; 1422219974Smav 1423219974Smav sc = md->mdo_softc; 1424219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 1425219974Smav 1426219974Smav if (sc->sc_stopping == G_RAID_DESTROY_HARD) 1427219974Smav return (0); 1428219974Smav 1429219974Smav /* There is only one volume. */ 1430219974Smav vol = TAILQ_FIRST(&sc->sc_volumes); 1431219974Smav 1432219974Smav /* Fill global fields. */ 1433219974Smav meta = malloc(sizeof(*meta), M_MD_NVIDIA, M_WAITOK | M_ZERO); 1434219974Smav if (mdi->mdio_meta) 1435219974Smav memcpy(meta, mdi->mdio_meta, sizeof(*meta)); 1436220209Smav memcpy(meta->nvidia_id, NVIDIA_MAGIC, sizeof(NVIDIA_MAGIC) - 1); 1437219974Smav meta->config_size = 30; 1438219974Smav meta->version = 0x0064; 1439219974Smav meta->total_sectors = vol->v_mediasize / vol->v_sectorsize; 1440219974Smav meta->sector_size = vol->v_sectorsize; 1441219974Smav nvidia_meta_put_name(meta, vol->v_name); 1442219974Smav meta->magic_0 = NVIDIA_MAGIC0; 1443219974Smav memcpy(&meta->volume_id, &mdi->mdio_volume_id, 16); 1444219974Smav meta->state = NVIDIA_S_IDLE; 1445219974Smav if (vol->v_raid_level == G_RAID_VOLUME_RL_RAID1) 1446219974Smav meta->array_width = 1; 1447219974Smav else if (vol->v_raid_level == G_RAID_VOLUME_RL_RAID1E) 1448219974Smav meta->array_width = vol->v_disks_count / 2; 1449219974Smav else if (vol->v_raid_level == G_RAID_VOLUME_RL_RAID5) 1450219974Smav meta->array_width = vol->v_disks_count - 1; 1451219974Smav else 1452219974Smav meta->array_width = vol->v_disks_count; 1453219974Smav meta->total_disks = vol->v_disks_count; 1454219974Smav meta->orig_array_width = meta->array_width; 1455219974Smav if (vol->v_raid_level == G_RAID_VOLUME_RL_RAID0) 1456219974Smav meta->type = NVIDIA_T_RAID0; 1457219974Smav else if (vol->v_raid_level == G_RAID_VOLUME_RL_RAID1) 1458219974Smav meta->type = NVIDIA_T_RAID1; 1459219974Smav else if (vol->v_raid_level == G_RAID_VOLUME_RL_RAID1E) 1460219974Smav meta->type = NVIDIA_T_RAID01; 1461219974Smav else if (vol->v_raid_level == G_RAID_VOLUME_RL_CONCAT || 1462219974Smav vol->v_raid_level == G_RAID_VOLUME_RL_SINGLE) 1463219974Smav meta->type = NVIDIA_T_CONCAT; 1464219974Smav// else if (vol->v_raid_level_qualifier == 0) 1465219974Smav// meta->type = NVIDIA_T_RAID5; 1466219974Smav else 1467219974Smav meta->type = NVIDIA_T_RAID5_SYM; 1468219974Smav meta->strip_sectors = vol->v_strip_size / vol->v_sectorsize; 1469219974Smav meta->strip_bytes = vol->v_strip_size; 1470219974Smav meta->strip_shift = ffs(meta->strip_sectors) - 1; 1471219974Smav meta->strip_mask = meta->strip_sectors - 1; 1472219974Smav meta->stripe_sectors = meta->strip_sectors * meta->orig_array_width; 1473219974Smav meta->stripe_bytes = meta->stripe_sectors * vol->v_sectorsize; 1474219974Smav meta->rebuild_lba = 0; 1475219974Smav meta->orig_type = meta->type; 1476219974Smav meta->orig_total_sectors = meta->total_sectors; 1477219974Smav meta->status = 0; 1478219974Smav 1479219974Smav for (i = 0; i < vol->v_disks_count; i++) { 1480219974Smav sd = &vol->v_subdisks[i]; 1481219974Smav if ((sd->sd_state == G_RAID_SUBDISK_S_STALE || 1482219974Smav sd->sd_state == G_RAID_SUBDISK_S_RESYNC || 1483219974Smav vol->v_dirty) && 1484219974Smav meta->state != NVIDIA_S_REBUILD) 1485219974Smav meta->state = NVIDIA_S_SYNC; 1486219974Smav else if (sd->sd_state == G_RAID_SUBDISK_S_NEW || 1487219974Smav sd->sd_state == G_RAID_SUBDISK_S_REBUILD) 1488219974Smav meta->state = NVIDIA_S_REBUILD; 1489219974Smav } 1490219974Smav 1491219974Smav /* We are done. Print meta data and store them to disks. */ 1492219974Smav if (mdi->mdio_meta != NULL) 1493219974Smav free(mdi->mdio_meta, M_MD_NVIDIA); 1494219974Smav mdi->mdio_meta = meta; 1495219974Smav spares = 0; 1496219974Smav TAILQ_FOREACH(disk, &sc->sc_disks, d_next) { 1497219974Smav pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data; 1498219974Smav if (disk->d_state != G_RAID_DISK_S_ACTIVE && 1499219974Smav disk->d_state != G_RAID_DISK_S_SPARE) 1500219974Smav continue; 1501219974Smav if (pd->pd_meta != NULL) { 1502219974Smav free(pd->pd_meta, M_MD_NVIDIA); 1503219974Smav pd->pd_meta = NULL; 1504219974Smav } 1505219974Smav pd->pd_meta = nvidia_meta_copy(meta); 1506219974Smav if ((sd = TAILQ_FIRST(&disk->d_subdisks)) != NULL) { 1507219974Smav /* For RAID0+1 we need to translate order. */ 1508219974Smav pd->pd_meta->disk_number = 1509219974Smav nvidia_meta_translate_disk(meta, sd->sd_pos); 1510219974Smav if (sd->sd_state != G_RAID_SUBDISK_S_ACTIVE) { 1511219974Smav pd->pd_meta->disk_status = 0x100; 1512219974Smav pd->pd_meta->rebuild_lba = 1513219974Smav sd->sd_rebuild_pos / vol->v_sectorsize * 1514219974Smav meta->array_width; 1515219974Smav } 1516219974Smav } else 1517219974Smav pd->pd_meta->disk_number = meta->total_disks + spares++; 1518219974Smav G_RAID_DEBUG(1, "Writing NVIDIA metadata to %s", 1519219974Smav g_raid_get_diskname(disk)); 1520219974Smav g_raid_md_nvidia_print(pd->pd_meta); 1521219974Smav nvidia_meta_write(disk->d_consumer, pd->pd_meta); 1522219974Smav } 1523219974Smav return (0); 1524219974Smav} 1525219974Smav 1526219974Smavstatic int 1527219974Smavg_raid_md_fail_disk_nvidia(struct g_raid_md_object *md, 1528219974Smav struct g_raid_subdisk *tsd, struct g_raid_disk *tdisk) 1529219974Smav{ 1530219974Smav struct g_raid_softc *sc; 1531219974Smav struct g_raid_md_nvidia_object *mdi; 1532219974Smav struct g_raid_md_nvidia_perdisk *pd; 1533219974Smav struct g_raid_subdisk *sd; 1534219974Smav 1535219974Smav sc = md->mdo_softc; 1536219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 1537219974Smav pd = (struct g_raid_md_nvidia_perdisk *)tdisk->d_md_data; 1538219974Smav 1539219974Smav /* We can't fail disk that is not a part of array now. */ 1540219974Smav if (pd->pd_disk_pos < 0) 1541219974Smav return (-1); 1542219974Smav 1543219974Smav /* Erase metadata to prevent disks's later resurrection. */ 1544219974Smav if (tdisk->d_consumer) 1545219974Smav nvidia_meta_erase(tdisk->d_consumer); 1546219974Smav 1547219974Smav /* Change states. */ 1548219974Smav g_raid_change_disk_state(tdisk, G_RAID_DISK_S_FAILED); 1549219974Smav TAILQ_FOREACH(sd, &tdisk->d_subdisks, sd_next) { 1550219974Smav g_raid_change_subdisk_state(sd, 1551219974Smav G_RAID_SUBDISK_S_FAILED); 1552219974Smav g_raid_event_send(sd, G_RAID_SUBDISK_E_FAILED, 1553219974Smav G_RAID_EVENT_SUBDISK); 1554219974Smav } 1555219974Smav 1556219974Smav /* Write updated metadata to remaining disks. */ 1557219974Smav g_raid_md_write_nvidia(md, NULL, NULL, tdisk); 1558219974Smav 1559219974Smav /* Check if anything left except placeholders. */ 1560219974Smav if (g_raid_ndisks(sc, -1) == 1561219974Smav g_raid_ndisks(sc, G_RAID_DISK_S_OFFLINE)) 1562219974Smav g_raid_destroy_node(sc, 0); 1563219974Smav else 1564219974Smav g_raid_md_nvidia_refill(sc); 1565219974Smav return (0); 1566219974Smav} 1567219974Smav 1568219974Smavstatic int 1569219974Smavg_raid_md_free_disk_nvidia(struct g_raid_md_object *md, 1570219974Smav struct g_raid_disk *disk) 1571219974Smav{ 1572219974Smav struct g_raid_md_nvidia_perdisk *pd; 1573219974Smav 1574219974Smav pd = (struct g_raid_md_nvidia_perdisk *)disk->d_md_data; 1575219974Smav if (pd->pd_meta != NULL) { 1576219974Smav free(pd->pd_meta, M_MD_NVIDIA); 1577219974Smav pd->pd_meta = NULL; 1578219974Smav } 1579219974Smav free(pd, M_MD_NVIDIA); 1580219974Smav disk->d_md_data = NULL; 1581219974Smav return (0); 1582219974Smav} 1583219974Smav 1584219974Smavstatic int 1585219974Smavg_raid_md_free_nvidia(struct g_raid_md_object *md) 1586219974Smav{ 1587219974Smav struct g_raid_md_nvidia_object *mdi; 1588219974Smav 1589219974Smav mdi = (struct g_raid_md_nvidia_object *)md; 1590219974Smav if (!mdi->mdio_started) { 1591219974Smav mdi->mdio_started = 0; 1592219974Smav callout_stop(&mdi->mdio_start_co); 1593219974Smav G_RAID_DEBUG1(1, md->mdo_softc, 1594219974Smav "root_mount_rel %p", mdi->mdio_rootmount); 1595219974Smav root_mount_rel(mdi->mdio_rootmount); 1596219974Smav mdi->mdio_rootmount = NULL; 1597219974Smav } 1598219974Smav if (mdi->mdio_meta != NULL) { 1599219974Smav free(mdi->mdio_meta, M_MD_NVIDIA); 1600219974Smav mdi->mdio_meta = NULL; 1601219974Smav } 1602219974Smav return (0); 1603219974Smav} 1604219974Smav 1605219974SmavG_RAID_MD_DECLARE(g_raid_md_nvidia); 1606