vmbus_chan.c revision 302819
1/*- 2 * Copyright (c) 2009-2012,2016 Microsoft Corp. 3 * Copyright (c) 2012 NetApp Inc. 4 * Copyright (c) 2012 Citrix Inc. 5 * All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice unmodified, this list of conditions, and the following 12 * disclaimer. 13 * 2. Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in the 15 * documentation and/or other materials provided with the distribution. 16 * 17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 18 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 19 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 20 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 21 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 22 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 23 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 24 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 25 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 27 */ 28 29#include <sys/cdefs.h> 30__FBSDID("$FreeBSD: head/sys/dev/hyperv/vmbus/hv_channel.c 302819 2016-07-14 08:15:13Z sephe $"); 31 32#include <sys/param.h> 33#include <sys/kernel.h> 34#include <sys/malloc.h> 35#include <sys/systm.h> 36#include <sys/mbuf.h> 37#include <sys/lock.h> 38#include <sys/mutex.h> 39#include <sys/sysctl.h> 40 41#include <machine/atomic.h> 42#include <machine/bus.h> 43 44#include <vm/vm.h> 45#include <vm/vm_param.h> 46#include <vm/pmap.h> 47 48#include <dev/hyperv/vmbus/hv_vmbus_priv.h> 49#include <dev/hyperv/vmbus/hyperv_var.h> 50#include <dev/hyperv/vmbus/vmbus_reg.h> 51#include <dev/hyperv/vmbus/vmbus_var.h> 52 53static void vmbus_chan_send_event(hv_vmbus_channel* channel); 54static void vmbus_chan_update_evtflagcnt(struct vmbus_softc *, 55 const struct hv_vmbus_channel *); 56static void vmbus_chan_task(void *, int); 57static void vmbus_chan_task_nobatch(void *, int); 58 59/** 60 * @brief Trigger an event notification on the specified channel 61 */ 62static void 63vmbus_chan_send_event(hv_vmbus_channel *channel) 64{ 65 struct vmbus_softc *sc = channel->vmbus_sc; 66 uint32_t chanid = channel->ch_id; 67 68 atomic_set_long(&sc->vmbus_tx_evtflags[chanid >> VMBUS_EVTFLAG_SHIFT], 69 1UL << (chanid & VMBUS_EVTFLAG_MASK)); 70 71 if (channel->ch_flags & VMBUS_CHAN_FLAG_HASMNF) { 72 atomic_set_int( 73 &sc->vmbus_mnf2->mnf_trigs[channel->ch_montrig_idx].mt_pending, 74 channel->ch_montrig_mask); 75 } else { 76 hypercall_signal_event(channel->ch_monprm_dma.hv_paddr); 77 } 78} 79 80static int 81vmbus_channel_sysctl_monalloc(SYSCTL_HANDLER_ARGS) 82{ 83 struct hv_vmbus_channel *chan = arg1; 84 int alloc = 0; 85 86 if (chan->ch_flags & VMBUS_CHAN_FLAG_HASMNF) 87 alloc = 1; 88 return sysctl_handle_int(oidp, &alloc, 0, req); 89} 90 91static void 92vmbus_channel_sysctl_create(hv_vmbus_channel* channel) 93{ 94 device_t dev; 95 struct sysctl_oid *devch_sysctl; 96 struct sysctl_oid *devch_id_sysctl, *devch_sub_sysctl; 97 struct sysctl_oid *devch_id_in_sysctl, *devch_id_out_sysctl; 98 struct sysctl_ctx_list *ctx; 99 uint32_t ch_id; 100 uint16_t sub_ch_id; 101 char name[16]; 102 103 hv_vmbus_channel* primary_ch = channel->ch_prichan; 104 105 if (primary_ch == NULL) { 106 dev = channel->ch_dev; 107 ch_id = channel->ch_id; 108 } else { 109 dev = primary_ch->ch_dev; 110 ch_id = primary_ch->ch_id; 111 sub_ch_id = channel->ch_subidx; 112 } 113 ctx = &channel->ch_sysctl_ctx; 114 sysctl_ctx_init(ctx); 115 /* This creates dev.DEVNAME.DEVUNIT.channel tree */ 116 devch_sysctl = SYSCTL_ADD_NODE(ctx, 117 SYSCTL_CHILDREN(device_get_sysctl_tree(dev)), 118 OID_AUTO, "channel", CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 119 /* This creates dev.DEVNAME.DEVUNIT.channel.CHANID tree */ 120 snprintf(name, sizeof(name), "%d", ch_id); 121 devch_id_sysctl = SYSCTL_ADD_NODE(ctx, 122 SYSCTL_CHILDREN(devch_sysctl), 123 OID_AUTO, name, CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 124 125 if (primary_ch != NULL) { 126 devch_sub_sysctl = SYSCTL_ADD_NODE(ctx, 127 SYSCTL_CHILDREN(devch_id_sysctl), 128 OID_AUTO, "sub", CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 129 snprintf(name, sizeof(name), "%d", sub_ch_id); 130 devch_id_sysctl = SYSCTL_ADD_NODE(ctx, 131 SYSCTL_CHILDREN(devch_sub_sysctl), 132 OID_AUTO, name, CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 133 134 SYSCTL_ADD_UINT(ctx, SYSCTL_CHILDREN(devch_id_sysctl), 135 OID_AUTO, "chanid", CTLFLAG_RD, 136 &channel->ch_id, 0, "channel id"); 137 } 138 SYSCTL_ADD_UINT(ctx, SYSCTL_CHILDREN(devch_id_sysctl), OID_AUTO, 139 "cpu", CTLFLAG_RD, &channel->target_cpu, 0, "owner CPU id"); 140 SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(devch_id_sysctl), OID_AUTO, 141 "monitor_allocated", CTLTYPE_INT | CTLFLAG_RD | CTLFLAG_MPSAFE, 142 channel, 0, vmbus_channel_sysctl_monalloc, "I", 143 "is monitor allocated to this channel"); 144 145 devch_id_in_sysctl = SYSCTL_ADD_NODE(ctx, 146 SYSCTL_CHILDREN(devch_id_sysctl), 147 OID_AUTO, 148 "in", 149 CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 150 devch_id_out_sysctl = SYSCTL_ADD_NODE(ctx, 151 SYSCTL_CHILDREN(devch_id_sysctl), 152 OID_AUTO, 153 "out", 154 CTLFLAG_RD | CTLFLAG_MPSAFE, 0, ""); 155 hv_ring_buffer_stat(ctx, 156 SYSCTL_CHILDREN(devch_id_in_sysctl), 157 &(channel->inbound), 158 "inbound ring buffer stats"); 159 hv_ring_buffer_stat(ctx, 160 SYSCTL_CHILDREN(devch_id_out_sysctl), 161 &(channel->outbound), 162 "outbound ring buffer stats"); 163} 164 165/** 166 * @brief Open the specified channel 167 */ 168int 169hv_vmbus_channel_open( 170 hv_vmbus_channel* new_channel, 171 uint32_t send_ring_buffer_size, 172 uint32_t recv_ring_buffer_size, 173 void* user_data, 174 uint32_t user_data_len, 175 hv_vmbus_pfn_channel_callback pfn_on_channel_callback, 176 void* context) 177{ 178 struct vmbus_softc *sc = new_channel->vmbus_sc; 179 const struct vmbus_chanmsg_chopen_resp *resp; 180 const struct vmbus_message *msg; 181 struct vmbus_chanmsg_chopen *req; 182 struct vmbus_msghc *mh; 183 uint32_t status; 184 int ret = 0; 185 void *in, *out; 186 187 if (user_data_len > VMBUS_CHANMSG_CHOPEN_UDATA_SIZE) { 188 device_printf(sc->vmbus_dev, 189 "invalid udata len %u for chan%u\n", 190 user_data_len, new_channel->ch_id); 191 return EINVAL; 192 } 193 194 if (atomic_testandset_int(&new_channel->ch_stflags, 195 VMBUS_CHAN_ST_OPENED_SHIFT)) 196 panic("double-open chan%u", new_channel->ch_id); 197 198 new_channel->on_channel_callback = pfn_on_channel_callback; 199 new_channel->channel_callback_context = context; 200 201 vmbus_chan_update_evtflagcnt(sc, new_channel); 202 203 new_channel->rxq = VMBUS_PCPU_GET(new_channel->vmbus_sc, event_tq, 204 new_channel->target_cpu); 205 if (new_channel->ch_flags & VMBUS_CHAN_FLAG_BATCHREAD) { 206 TASK_INIT(&new_channel->channel_task, 0, 207 vmbus_chan_task, new_channel); 208 } else { 209 TASK_INIT(&new_channel->channel_task, 0, 210 vmbus_chan_task_nobatch, new_channel); 211 } 212 213 /* Allocate the ring buffer */ 214 out = contigmalloc((send_ring_buffer_size + recv_ring_buffer_size), 215 M_DEVBUF, M_ZERO, 0UL, BUS_SPACE_MAXADDR, PAGE_SIZE, 0); 216 KASSERT(out != NULL, 217 ("Error VMBUS: contigmalloc failed to allocate Ring Buffer!")); 218 if (out == NULL) { 219 ret = ENOMEM; 220 goto failed; 221 } 222 223 in = ((uint8_t *) out + send_ring_buffer_size); 224 225 new_channel->ring_buffer_pages = out; 226 new_channel->ring_buffer_page_count = (send_ring_buffer_size + 227 recv_ring_buffer_size) >> PAGE_SHIFT; 228 new_channel->ring_buffer_size = send_ring_buffer_size + 229 recv_ring_buffer_size; 230 231 hv_vmbus_ring_buffer_init( 232 &new_channel->outbound, 233 out, 234 send_ring_buffer_size); 235 236 hv_vmbus_ring_buffer_init( 237 &new_channel->inbound, 238 in, 239 recv_ring_buffer_size); 240 241 /* Create sysctl tree for this channel */ 242 vmbus_channel_sysctl_create(new_channel); 243 244 /** 245 * Establish the gpadl for the ring buffer 246 */ 247 new_channel->ring_buffer_gpadl_handle = 0; 248 249 ret = hv_vmbus_channel_establish_gpadl(new_channel, 250 new_channel->outbound.ring_buffer, 251 send_ring_buffer_size + recv_ring_buffer_size, 252 &new_channel->ring_buffer_gpadl_handle); 253 254 /* 255 * Open channel w/ the bufring GPADL on the target CPU. 256 */ 257 mh = vmbus_msghc_get(sc, sizeof(*req)); 258 if (mh == NULL) { 259 device_printf(sc->vmbus_dev, 260 "can not get msg hypercall for chopen(chan%u)\n", 261 new_channel->ch_id); 262 ret = ENXIO; 263 goto failed; 264 } 265 266 req = vmbus_msghc_dataptr(mh); 267 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_CHOPEN; 268 req->chm_chanid = new_channel->ch_id; 269 req->chm_openid = new_channel->ch_id; 270 req->chm_gpadl = new_channel->ring_buffer_gpadl_handle; 271 req->chm_vcpuid = new_channel->target_vcpu; 272 req->chm_rxbr_pgofs = send_ring_buffer_size >> PAGE_SHIFT; 273 if (user_data_len) 274 memcpy(req->chm_udata, user_data, user_data_len); 275 276 ret = vmbus_msghc_exec(sc, mh); 277 if (ret != 0) { 278 device_printf(sc->vmbus_dev, 279 "chopen(chan%u) msg hypercall exec failed: %d\n", 280 new_channel->ch_id, ret); 281 vmbus_msghc_put(sc, mh); 282 goto failed; 283 } 284 285 msg = vmbus_msghc_wait_result(sc, mh); 286 resp = (const struct vmbus_chanmsg_chopen_resp *)msg->msg_data; 287 status = resp->chm_status; 288 289 vmbus_msghc_put(sc, mh); 290 291 if (status == 0) { 292 if (bootverbose) { 293 device_printf(sc->vmbus_dev, "chan%u opened\n", 294 new_channel->ch_id); 295 } 296 return 0; 297 } 298 299 device_printf(sc->vmbus_dev, "failed to open chan%u\n", 300 new_channel->ch_id); 301 ret = ENXIO; 302 303failed: 304 atomic_clear_int(&new_channel->ch_stflags, VMBUS_CHAN_ST_OPENED); 305 return ret; 306} 307 308/** 309 * @brief Establish a GPADL for the specified buffer 310 */ 311int 312hv_vmbus_channel_establish_gpadl(struct hv_vmbus_channel *channel, 313 void *contig_buffer, uint32_t size, uint32_t *gpadl0) 314{ 315 struct vmbus_softc *sc = channel->vmbus_sc; 316 struct vmbus_msghc *mh; 317 struct vmbus_chanmsg_gpadl_conn *req; 318 const struct vmbus_message *msg; 319 size_t reqsz; 320 uint32_t gpadl, status; 321 int page_count, range_len, i, cnt, error; 322 uint64_t page_id, paddr; 323 324 /* 325 * Preliminary checks. 326 */ 327 328 KASSERT((size & PAGE_MASK) == 0, 329 ("invalid GPA size %u, not multiple page size", size)); 330 page_count = size >> PAGE_SHIFT; 331 332 paddr = hv_get_phys_addr(contig_buffer); 333 KASSERT((paddr & PAGE_MASK) == 0, 334 ("GPA is not page aligned %jx", (uintmax_t)paddr)); 335 page_id = paddr >> PAGE_SHIFT; 336 337 range_len = __offsetof(struct vmbus_gpa_range, gpa_page[page_count]); 338 /* 339 * We don't support multiple GPA ranges. 340 */ 341 if (range_len > UINT16_MAX) { 342 device_printf(sc->vmbus_dev, "GPA too large, %d pages\n", 343 page_count); 344 return EOPNOTSUPP; 345 } 346 347 /* 348 * Allocate GPADL id. 349 */ 350 gpadl = vmbus_gpadl_alloc(sc); 351 *gpadl0 = gpadl; 352 353 /* 354 * Connect this GPADL to the target channel. 355 * 356 * NOTE: 357 * Since each message can only hold small set of page 358 * addresses, several messages may be required to 359 * complete the connection. 360 */ 361 if (page_count > VMBUS_CHANMSG_GPADL_CONN_PGMAX) 362 cnt = VMBUS_CHANMSG_GPADL_CONN_PGMAX; 363 else 364 cnt = page_count; 365 page_count -= cnt; 366 367 reqsz = __offsetof(struct vmbus_chanmsg_gpadl_conn, 368 chm_range.gpa_page[cnt]); 369 mh = vmbus_msghc_get(sc, reqsz); 370 if (mh == NULL) { 371 device_printf(sc->vmbus_dev, 372 "can not get msg hypercall for gpadl->chan%u\n", 373 channel->ch_id); 374 return EIO; 375 } 376 377 req = vmbus_msghc_dataptr(mh); 378 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_GPADL_CONN; 379 req->chm_chanid = channel->ch_id; 380 req->chm_gpadl = gpadl; 381 req->chm_range_len = range_len; 382 req->chm_range_cnt = 1; 383 req->chm_range.gpa_len = size; 384 req->chm_range.gpa_ofs = 0; 385 for (i = 0; i < cnt; ++i) 386 req->chm_range.gpa_page[i] = page_id++; 387 388 error = vmbus_msghc_exec(sc, mh); 389 if (error) { 390 device_printf(sc->vmbus_dev, 391 "gpadl->chan%u msg hypercall exec failed: %d\n", 392 channel->ch_id, error); 393 vmbus_msghc_put(sc, mh); 394 return error; 395 } 396 397 while (page_count > 0) { 398 struct vmbus_chanmsg_gpadl_subconn *subreq; 399 400 if (page_count > VMBUS_CHANMSG_GPADL_SUBCONN_PGMAX) 401 cnt = VMBUS_CHANMSG_GPADL_SUBCONN_PGMAX; 402 else 403 cnt = page_count; 404 page_count -= cnt; 405 406 reqsz = __offsetof(struct vmbus_chanmsg_gpadl_subconn, 407 chm_gpa_page[cnt]); 408 vmbus_msghc_reset(mh, reqsz); 409 410 subreq = vmbus_msghc_dataptr(mh); 411 subreq->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_GPADL_SUBCONN; 412 subreq->chm_gpadl = gpadl; 413 for (i = 0; i < cnt; ++i) 414 subreq->chm_gpa_page[i] = page_id++; 415 416 vmbus_msghc_exec_noresult(mh); 417 } 418 KASSERT(page_count == 0, ("invalid page count %d", page_count)); 419 420 msg = vmbus_msghc_wait_result(sc, mh); 421 status = ((const struct vmbus_chanmsg_gpadl_connresp *) 422 msg->msg_data)->chm_status; 423 424 vmbus_msghc_put(sc, mh); 425 426 if (status != 0) { 427 device_printf(sc->vmbus_dev, "gpadl->chan%u failed: " 428 "status %u\n", channel->ch_id, status); 429 return EIO; 430 } else { 431 if (bootverbose) { 432 device_printf(sc->vmbus_dev, "gpadl->chan%u " 433 "succeeded\n", channel->ch_id); 434 } 435 } 436 return 0; 437} 438 439/* 440 * Disconnect the GPA from the target channel 441 */ 442int 443hv_vmbus_channel_teardown_gpdal(struct hv_vmbus_channel *chan, uint32_t gpadl) 444{ 445 struct vmbus_softc *sc = chan->vmbus_sc; 446 struct vmbus_msghc *mh; 447 struct vmbus_chanmsg_gpadl_disconn *req; 448 int error; 449 450 mh = vmbus_msghc_get(sc, sizeof(*req)); 451 if (mh == NULL) { 452 device_printf(sc->vmbus_dev, 453 "can not get msg hypercall for gpa x->chan%u\n", 454 chan->ch_id); 455 return EBUSY; 456 } 457 458 req = vmbus_msghc_dataptr(mh); 459 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_GPADL_DISCONN; 460 req->chm_chanid = chan->ch_id; 461 req->chm_gpadl = gpadl; 462 463 error = vmbus_msghc_exec(sc, mh); 464 if (error) { 465 device_printf(sc->vmbus_dev, 466 "gpa x->chan%u msg hypercall exec failed: %d\n", 467 chan->ch_id, error); 468 vmbus_msghc_put(sc, mh); 469 return error; 470 } 471 472 vmbus_msghc_wait_result(sc, mh); 473 /* Discard result; no useful information */ 474 vmbus_msghc_put(sc, mh); 475 476 return 0; 477} 478 479static void 480hv_vmbus_channel_close_internal(hv_vmbus_channel *channel) 481{ 482 struct vmbus_softc *sc = channel->vmbus_sc; 483 struct vmbus_msghc *mh; 484 struct vmbus_chanmsg_chclose *req; 485 struct taskqueue *rxq = channel->rxq; 486 int error; 487 488 /* TODO: stringent check */ 489 atomic_clear_int(&channel->ch_stflags, VMBUS_CHAN_ST_OPENED); 490 491 sysctl_ctx_free(&channel->ch_sysctl_ctx); 492 493 /* 494 * set rxq to NULL to avoid more requests be scheduled 495 */ 496 channel->rxq = NULL; 497 taskqueue_drain(rxq, &channel->channel_task); 498 channel->on_channel_callback = NULL; 499 500 /** 501 * Send a closing message 502 */ 503 504 mh = vmbus_msghc_get(sc, sizeof(*req)); 505 if (mh == NULL) { 506 device_printf(sc->vmbus_dev, 507 "can not get msg hypercall for chclose(chan%u)\n", 508 channel->ch_id); 509 return; 510 } 511 512 req = vmbus_msghc_dataptr(mh); 513 req->chm_hdr.chm_type = VMBUS_CHANMSG_TYPE_CHCLOSE; 514 req->chm_chanid = channel->ch_id; 515 516 error = vmbus_msghc_exec_noresult(mh); 517 vmbus_msghc_put(sc, mh); 518 519 if (error) { 520 device_printf(sc->vmbus_dev, 521 "chclose(chan%u) msg hypercall exec failed: %d\n", 522 channel->ch_id, error); 523 return; 524 } else if (bootverbose) { 525 device_printf(sc->vmbus_dev, "close chan%u\n", 526 channel->ch_id); 527 } 528 529 /* Tear down the gpadl for the channel's ring buffer */ 530 if (channel->ring_buffer_gpadl_handle) { 531 hv_vmbus_channel_teardown_gpdal(channel, 532 channel->ring_buffer_gpadl_handle); 533 } 534 535 /* TODO: Send a msg to release the childRelId */ 536 537 /* cleanup the ring buffers for this channel */ 538 hv_ring_buffer_cleanup(&channel->outbound); 539 hv_ring_buffer_cleanup(&channel->inbound); 540 541 contigfree(channel->ring_buffer_pages, channel->ring_buffer_size, 542 M_DEVBUF); 543} 544 545/* 546 * Caller should make sure that all sub-channels have 547 * been added to 'chan' and all to-be-closed channels 548 * are not being opened. 549 */ 550void 551hv_vmbus_channel_close(struct hv_vmbus_channel *chan) 552{ 553 int subchan_cnt; 554 555 if (!VMBUS_CHAN_ISPRIMARY(chan)) { 556 /* 557 * Sub-channel is closed when its primary channel 558 * is closed; done. 559 */ 560 return; 561 } 562 563 /* 564 * Close all sub-channels, if any. 565 */ 566 subchan_cnt = chan->ch_subchan_cnt; 567 if (subchan_cnt > 0) { 568 struct hv_vmbus_channel **subchan; 569 int i; 570 571 subchan = vmbus_get_subchan(chan, subchan_cnt); 572 for (i = 0; i < subchan_cnt; ++i) 573 hv_vmbus_channel_close_internal(subchan[i]); 574 vmbus_rel_subchan(subchan, subchan_cnt); 575 } 576 577 /* Then close the primary channel. */ 578 hv_vmbus_channel_close_internal(chan); 579} 580 581/** 582 * @brief Send the specified buffer on the given channel 583 */ 584int 585hv_vmbus_channel_send_packet( 586 hv_vmbus_channel* channel, 587 void* buffer, 588 uint32_t buffer_len, 589 uint64_t request_id, 590 hv_vmbus_packet_type type, 591 uint32_t flags) 592{ 593 int ret = 0; 594 hv_vm_packet_descriptor desc; 595 uint32_t packet_len; 596 uint64_t aligned_data; 597 uint32_t packet_len_aligned; 598 boolean_t need_sig; 599 hv_vmbus_sg_buffer_list buffer_list[3]; 600 601 packet_len = sizeof(hv_vm_packet_descriptor) + buffer_len; 602 packet_len_aligned = HV_ALIGN_UP(packet_len, sizeof(uint64_t)); 603 aligned_data = 0; 604 605 /* Setup the descriptor */ 606 desc.type = type; /* HV_VMBUS_PACKET_TYPE_DATA_IN_BAND; */ 607 desc.flags = flags; /* HV_VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED */ 608 /* in 8-bytes granularity */ 609 desc.data_offset8 = sizeof(hv_vm_packet_descriptor) >> 3; 610 desc.length8 = (uint16_t) (packet_len_aligned >> 3); 611 desc.transaction_id = request_id; 612 613 buffer_list[0].data = &desc; 614 buffer_list[0].length = sizeof(hv_vm_packet_descriptor); 615 616 buffer_list[1].data = buffer; 617 buffer_list[1].length = buffer_len; 618 619 buffer_list[2].data = &aligned_data; 620 buffer_list[2].length = packet_len_aligned - packet_len; 621 622 ret = hv_ring_buffer_write(&channel->outbound, buffer_list, 3, 623 &need_sig); 624 625 /* TODO: We should determine if this is optional */ 626 if (ret == 0 && need_sig) 627 vmbus_chan_send_event(channel); 628 629 return (ret); 630} 631 632/** 633 * @brief Send a range of single-page buffer packets using 634 * a GPADL Direct packet type 635 */ 636int 637hv_vmbus_channel_send_packet_pagebuffer( 638 hv_vmbus_channel* channel, 639 hv_vmbus_page_buffer page_buffers[], 640 uint32_t page_count, 641 void* buffer, 642 uint32_t buffer_len, 643 uint64_t request_id) 644{ 645 646 int ret = 0; 647 boolean_t need_sig; 648 uint32_t packet_len; 649 uint32_t page_buflen; 650 uint32_t packetLen_aligned; 651 hv_vmbus_sg_buffer_list buffer_list[4]; 652 hv_vmbus_channel_packet_page_buffer desc; 653 uint32_t descSize; 654 uint64_t alignedData = 0; 655 656 if (page_count > HV_MAX_PAGE_BUFFER_COUNT) 657 return (EINVAL); 658 659 /* 660 * Adjust the size down since hv_vmbus_channel_packet_page_buffer 661 * is the largest size we support 662 */ 663 descSize = __offsetof(hv_vmbus_channel_packet_page_buffer, range); 664 page_buflen = sizeof(hv_vmbus_page_buffer) * page_count; 665 packet_len = descSize + page_buflen + buffer_len; 666 packetLen_aligned = HV_ALIGN_UP(packet_len, sizeof(uint64_t)); 667 668 /* Setup the descriptor */ 669 desc.type = HV_VMBUS_PACKET_TYPE_DATA_USING_GPA_DIRECT; 670 desc.flags = HV_VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED; 671 /* in 8-bytes granularity */ 672 desc.data_offset8 = (descSize + page_buflen) >> 3; 673 desc.length8 = (uint16_t) (packetLen_aligned >> 3); 674 desc.transaction_id = request_id; 675 desc.range_count = page_count; 676 677 buffer_list[0].data = &desc; 678 buffer_list[0].length = descSize; 679 680 buffer_list[1].data = page_buffers; 681 buffer_list[1].length = page_buflen; 682 683 buffer_list[2].data = buffer; 684 buffer_list[2].length = buffer_len; 685 686 buffer_list[3].data = &alignedData; 687 buffer_list[3].length = packetLen_aligned - packet_len; 688 689 ret = hv_ring_buffer_write(&channel->outbound, buffer_list, 4, 690 &need_sig); 691 692 /* TODO: We should determine if this is optional */ 693 if (ret == 0 && need_sig) 694 vmbus_chan_send_event(channel); 695 696 return (ret); 697} 698 699/** 700 * @brief Send a multi-page buffer packet using a GPADL Direct packet type 701 */ 702int 703hv_vmbus_channel_send_packet_multipagebuffer( 704 hv_vmbus_channel* channel, 705 hv_vmbus_multipage_buffer* multi_page_buffer, 706 void* buffer, 707 uint32_t buffer_len, 708 uint64_t request_id) 709{ 710 711 int ret = 0; 712 uint32_t desc_size; 713 boolean_t need_sig; 714 uint32_t packet_len; 715 uint32_t packet_len_aligned; 716 uint32_t pfn_count; 717 uint64_t aligned_data = 0; 718 hv_vmbus_sg_buffer_list buffer_list[3]; 719 hv_vmbus_channel_packet_multipage_buffer desc; 720 721 pfn_count = 722 HV_NUM_PAGES_SPANNED( 723 multi_page_buffer->offset, 724 multi_page_buffer->length); 725 726 if ((pfn_count == 0) || (pfn_count > HV_MAX_MULTIPAGE_BUFFER_COUNT)) 727 return (EINVAL); 728 /* 729 * Adjust the size down since hv_vmbus_channel_packet_multipage_buffer 730 * is the largest size we support 731 */ 732 desc_size = 733 sizeof(hv_vmbus_channel_packet_multipage_buffer) - 734 ((HV_MAX_MULTIPAGE_BUFFER_COUNT - pfn_count) * 735 sizeof(uint64_t)); 736 packet_len = desc_size + buffer_len; 737 packet_len_aligned = HV_ALIGN_UP(packet_len, sizeof(uint64_t)); 738 739 /* 740 * Setup the descriptor 741 */ 742 desc.type = HV_VMBUS_PACKET_TYPE_DATA_USING_GPA_DIRECT; 743 desc.flags = HV_VMBUS_DATA_PACKET_FLAG_COMPLETION_REQUESTED; 744 desc.data_offset8 = desc_size >> 3; /* in 8-bytes granularity */ 745 desc.length8 = (uint16_t) (packet_len_aligned >> 3); 746 desc.transaction_id = request_id; 747 desc.range_count = 1; 748 749 desc.range.length = multi_page_buffer->length; 750 desc.range.offset = multi_page_buffer->offset; 751 752 memcpy(desc.range.pfn_array, multi_page_buffer->pfn_array, 753 pfn_count * sizeof(uint64_t)); 754 755 buffer_list[0].data = &desc; 756 buffer_list[0].length = desc_size; 757 758 buffer_list[1].data = buffer; 759 buffer_list[1].length = buffer_len; 760 761 buffer_list[2].data = &aligned_data; 762 buffer_list[2].length = packet_len_aligned - packet_len; 763 764 ret = hv_ring_buffer_write(&channel->outbound, buffer_list, 3, 765 &need_sig); 766 767 /* TODO: We should determine if this is optional */ 768 if (ret == 0 && need_sig) 769 vmbus_chan_send_event(channel); 770 771 return (ret); 772} 773 774/** 775 * @brief Retrieve the user packet on the specified channel 776 */ 777int 778hv_vmbus_channel_recv_packet( 779 hv_vmbus_channel* channel, 780 void* Buffer, 781 uint32_t buffer_len, 782 uint32_t* buffer_actual_len, 783 uint64_t* request_id) 784{ 785 int ret; 786 uint32_t user_len; 787 uint32_t packet_len; 788 hv_vm_packet_descriptor desc; 789 790 *buffer_actual_len = 0; 791 *request_id = 0; 792 793 ret = hv_ring_buffer_peek(&channel->inbound, &desc, 794 sizeof(hv_vm_packet_descriptor)); 795 if (ret != 0) 796 return (0); 797 798 packet_len = desc.length8 << 3; 799 user_len = packet_len - (desc.data_offset8 << 3); 800 801 *buffer_actual_len = user_len; 802 803 if (user_len > buffer_len) 804 return (EINVAL); 805 806 *request_id = desc.transaction_id; 807 808 /* Copy over the packet to the user buffer */ 809 ret = hv_ring_buffer_read(&channel->inbound, Buffer, user_len, 810 (desc.data_offset8 << 3)); 811 812 return (0); 813} 814 815/** 816 * @brief Retrieve the raw packet on the specified channel 817 */ 818int 819hv_vmbus_channel_recv_packet_raw( 820 hv_vmbus_channel* channel, 821 void* buffer, 822 uint32_t buffer_len, 823 uint32_t* buffer_actual_len, 824 uint64_t* request_id) 825{ 826 int ret; 827 uint32_t packetLen; 828 hv_vm_packet_descriptor desc; 829 830 *buffer_actual_len = 0; 831 *request_id = 0; 832 833 ret = hv_ring_buffer_peek( 834 &channel->inbound, &desc, 835 sizeof(hv_vm_packet_descriptor)); 836 837 if (ret != 0) 838 return (0); 839 840 packetLen = desc.length8 << 3; 841 *buffer_actual_len = packetLen; 842 843 if (packetLen > buffer_len) 844 return (ENOBUFS); 845 846 *request_id = desc.transaction_id; 847 848 /* Copy over the entire packet to the user buffer */ 849 ret = hv_ring_buffer_read(&channel->inbound, buffer, packetLen, 0); 850 851 return (0); 852} 853 854static void 855vmbus_chan_task(void *xchan, int pending __unused) 856{ 857 struct hv_vmbus_channel *chan = xchan; 858 void (*callback)(void *); 859 void *arg; 860 861 arg = chan->channel_callback_context; 862 callback = chan->on_channel_callback; 863 864 /* 865 * Optimize host to guest signaling by ensuring: 866 * 1. While reading the channel, we disable interrupts from 867 * host. 868 * 2. Ensure that we process all posted messages from the host 869 * before returning from this callback. 870 * 3. Once we return, enable signaling from the host. Once this 871 * state is set we check to see if additional packets are 872 * available to read. In this case we repeat the process. 873 * 874 * NOTE: Interrupt has been disabled in the ISR. 875 */ 876 for (;;) { 877 uint32_t left; 878 879 callback(arg); 880 881 left = hv_ring_buffer_read_end(&chan->inbound); 882 if (left == 0) { 883 /* No more data in RX bufring; done */ 884 break; 885 } 886 hv_ring_buffer_read_begin(&chan->inbound); 887 } 888} 889 890static void 891vmbus_chan_task_nobatch(void *xchan, int pending __unused) 892{ 893 struct hv_vmbus_channel *chan = xchan; 894 895 chan->on_channel_callback(chan->channel_callback_context); 896} 897 898static __inline void 899vmbus_event_flags_proc(struct vmbus_softc *sc, volatile u_long *event_flags, 900 int flag_cnt) 901{ 902 int f; 903 904 for (f = 0; f < flag_cnt; ++f) { 905 uint32_t chid_base; 906 u_long flags; 907 int chid_ofs; 908 909 if (event_flags[f] == 0) 910 continue; 911 912 flags = atomic_swap_long(&event_flags[f], 0); 913 chid_base = f << VMBUS_EVTFLAG_SHIFT; 914 915 while ((chid_ofs = ffsl(flags)) != 0) { 916 struct hv_vmbus_channel *channel; 917 918 --chid_ofs; /* NOTE: ffsl is 1-based */ 919 flags &= ~(1UL << chid_ofs); 920 921 channel = sc->vmbus_chmap[chid_base + chid_ofs]; 922 923 /* if channel is closed or closing */ 924 if (channel == NULL || channel->rxq == NULL) 925 continue; 926 927 if (channel->ch_flags & VMBUS_CHAN_FLAG_BATCHREAD) 928 hv_ring_buffer_read_begin(&channel->inbound); 929 taskqueue_enqueue(channel->rxq, &channel->channel_task); 930 } 931 } 932} 933 934void 935vmbus_event_proc(struct vmbus_softc *sc, int cpu) 936{ 937 struct vmbus_evtflags *eventf; 938 939 /* 940 * On Host with Win8 or above, the event page can be checked directly 941 * to get the id of the channel that has the pending interrupt. 942 */ 943 eventf = VMBUS_PCPU_GET(sc, event_flags, cpu) + VMBUS_SINT_MESSAGE; 944 vmbus_event_flags_proc(sc, eventf->evt_flags, 945 VMBUS_PCPU_GET(sc, event_flags_cnt, cpu)); 946} 947 948void 949vmbus_event_proc_compat(struct vmbus_softc *sc, int cpu) 950{ 951 struct vmbus_evtflags *eventf; 952 953 eventf = VMBUS_PCPU_GET(sc, event_flags, cpu) + VMBUS_SINT_MESSAGE; 954 if (atomic_testandclear_long(&eventf->evt_flags[0], 0)) { 955 vmbus_event_flags_proc(sc, sc->vmbus_rx_evtflags, 956 VMBUS_CHAN_MAX_COMPAT >> VMBUS_EVTFLAG_SHIFT); 957 } 958} 959 960static void 961vmbus_chan_update_evtflagcnt(struct vmbus_softc *sc, 962 const struct hv_vmbus_channel *chan) 963{ 964 volatile int *flag_cnt_ptr; 965 int flag_cnt; 966 967 flag_cnt = (chan->ch_id / VMBUS_EVTFLAG_LEN) + 1; 968 flag_cnt_ptr = VMBUS_PCPU_PTR(sc, event_flags_cnt, chan->target_cpu); 969 970 for (;;) { 971 int old_flag_cnt; 972 973 old_flag_cnt = *flag_cnt_ptr; 974 if (old_flag_cnt >= flag_cnt) 975 break; 976 if (atomic_cmpset_int(flag_cnt_ptr, old_flag_cnt, flag_cnt)) { 977 if (bootverbose) { 978 device_printf(sc->vmbus_dev, 979 "channel%u update cpu%d flag_cnt to %d\n", 980 chan->ch_id, 981 chan->target_cpu, flag_cnt); 982 } 983 break; 984 } 985 } 986} 987