Lines Matching defs:msg_st

135 static errval_t chan_open_call_tx(struct txq_msg_st* msg_st)
137 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
139 return xeon_phi_chan_open_call__tx(msg_st->queue->binding, TXQCONT(msg_st),
146 static errval_t chan_open_request_response_tx(struct txq_msg_st* msg_st)
148 return xeon_phi_chan_open_request_response__tx(msg_st->queue->binding,
149 TXQCONT(msg_st), msg_st->err);
152 static errval_t kill_response_tx(struct txq_msg_st* msg_st)
154 return xeon_phi_kill_response__tx(msg_st->queue->binding, TXQCONT(msg_st),
155 msg_st->err);
158 static errval_t dma_register_response_tx(struct txq_msg_st* msg_st)
160 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
162 return xeon_phi_dma_register_response__tx(msg_st->queue->binding, TXQCONT(msg_st),
164 msg_st->err);
167 static errval_t dma_memcpy_response_tx(struct txq_msg_st* msg_st)
169 return xeon_phi_dma_memcpy_response__tx(msg_st->queue->binding, TXQCONT(msg_st),
170 msg_st->err);
173 static errval_t get_nodeid_response_tx(struct txq_msg_st* msg_st)
175 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
177 return xeon_phi_get_nodeid_response__tx(msg_st->queue->binding, TXQCONT(msg_st),
181 static errval_t spawn_with_cap_response_tx(struct txq_msg_st* msg_st)
183 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
185 return xeon_phi_spawn_with_cap_response__tx(msg_st->queue->binding,
186 TXQCONT(msg_st),
188 msg_st->err);
191 static errval_t spawn_response_tx(struct txq_msg_st* msg_st)
193 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
195 return xeon_phi_spawn_response__tx(msg_st->queue->binding, TXQCONT(msg_st),
196 xphi_st->args.spawn.domainid, msg_st->err);
199 static errval_t domain_init_response_tx(struct txq_msg_st* msg_st)
201 return xeon_phi_domain_init_response__tx(msg_st->queue->binding, TXQCONT(msg_st),
202 msg_st->err);
205 static errval_t domain_register_response_tx(struct txq_msg_st* msg_st)
207 return xeon_phi_domain_register_response__tx(msg_st->queue->binding,
208 TXQCONT(msg_st), msg_st->err);
211 static errval_t domain_lookup_response_tx(struct txq_msg_st* msg_st)
213 struct xphi_svc_msg_st *st = (struct xphi_svc_msg_st *) msg_st;
214 return xeon_phi_domain_lookup_response__tx(msg_st->queue->binding,
215 TXQCONT(msg_st),
216 st->args.domain.domid, msg_st->err);
219 static errval_t domain_wait_response_tx(struct txq_msg_st* msg_st)
221 struct xphi_svc_msg_st *st = (struct xphi_svc_msg_st *) msg_st;
223 return xeon_phi_domain_wait_response__tx(msg_st->queue->binding, TXQCONT(msg_st),
224 st->args.domain.domid, msg_st->err);
227 static errval_t alloc_mem_response_tx(struct txq_msg_st* msg_st)
229 struct xphi_svc_msg_st *st = (struct xphi_svc_msg_st *) msg_st;
231 return xeon_phi_alloc_mem_response__tx(msg_st->queue->binding, TXQCONT(msg_st),
232 st->args.alloc.cap, msg_st->err);
250 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
251 if (msg_st == NULL) {
255 msg_st->send = domain_lookup_response_tx;
256 msg_st->cleanup = NULL;
258 struct xphi_svc_msg_st *st = (struct xphi_svc_msg_st *) msg_st;
260 msg_st->err = interphi_domain_lookup(&phi->topology[phi->id], (CONST_CAST)name,
262 txq_send(msg_st);
274 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
275 if (msg_st == NULL) {
281 msg_st->err = interphi_domain_wait(&phi->topology[phi->id], (CONST_CAST)name, svc_st);
282 if (err_is_fail(msg_st->err)) {
283 txq_send(msg_st);
296 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
297 if (msg_st == NULL) {
301 msg_st->send = domain_register_response_tx;
302 msg_st->cleanup = NULL;
309 msg_st->err = interphi_domain_register(&phi->topology[phi->id], name, domid);
311 msg_st->err = domain_register(name, domid);
313 txq_send(msg_st);
331 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
332 if (msg_st == NULL) {
336 msg_st->send = domain_init_response_tx;
337 msg_st->cleanup = NULL;
340 msg_st->err = XEON_PHI_ERR_CLIENT_REGISTER;
341 txq_send(msg_st);
348 msg_st->err = interphi_domain_register(&phi->topology[phi->id], svc_st->name,
354 msg_st->err = domain_register(svc_st->name, svc_st->domainid);
357 txq_send(msg_st);
383 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
384 if (msg_st == NULL) {
388 msg_st->send = chan_open_request_response_tx;
389 msg_st->cleanup = NULL;
396 msg_st->err = interphi_chan_open(node, domain, svc_st->domainid, usrdata,
399 txq_send(msg_st);
408 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
409 if (msg_st == NULL) {
413 msg_st->send = kill_response_tx;
414 msg_st->cleanup = NULL;
422 msg_st->err = interphi_kill(node, domainid);
424 txq_send(msg_st);
439 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
440 if (msg_st == NULL) {
444 msg_st->send = spawn_with_cap_response_tx;
445 msg_st->cleanup = NULL;
447 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
454 msg_st->err = interphi_spawn_with_cap(node, core, (CONST_CAST)cmdline, length,
457 txq_send(msg_st);
471 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
472 if (msg_st == NULL) {
476 msg_st->send = spawn_response_tx;
477 msg_st->cleanup = NULL;
479 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
486 msg_st->err = interphi_spawn(node, core, (CONST_CAST)cmdline, length, flags,
488 txq_send(msg_st);
497 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
498 if (msg_st == NULL) {
502 msg_st->send = dma_register_response_tx;
503 msg_st->cleanup = NULL;
505 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
509 msg_st->err = xdma_state_init(svc_st->phi, &svc_st->dma_mem_mgr);
510 if (err_is_fail(msg_st->err )) {
515 msg_st->err= xdma_register_region(svc_st->phi, svc_st->dma_mem_mgr, mem,
518 txq_send(msg_st);
526 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
527 if (msg_st == NULL) {
533 msg_st->send = dma_memcpy_response_tx;
534 msg_st->cleanup = NULL;
535 msg_st->err = SYS_ERR_OK;
538 msg_st->err = DMA_ERR_MEM_NOT_REGISTERED;
544 err = xdma_memcpy(svc_st->phi, svc_st->dma_mem_mgr, to, from, bytes, msg_st);
545 if (err_is_fail(msg_st->err)) {
546 msg_st->err = err;
554 txq_send(msg_st);
564 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
565 if (msg_st == NULL) {
569 msg_st->send = get_nodeid_response_tx;
570 msg_st->cleanup = NULL;
572 msg_st->err = SYS_ERR_OK;
574 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
581 msg_st->err = err;
582 txq_send(msg_st);
605 txq_send(msg_st);
612 struct txq_msg_st *msg_st = txq_msg_st_alloc(&svc_st->queue);
613 if (msg_st == NULL) {
617 msg_st->send = alloc_mem_response_tx;
618 msg_st->cleanup = NULL;
622 struct xphi_svc_msg_st *st = (struct xphi_svc_msg_st *)msg_st;
624 msg_st->err = interphi_alloc_mem(node, bytes, &st->args.alloc.cap);
625 if (err_is_ok(msg_st->err)) {
633 txq_send(msg_st);
781 struct txq_msg_st *msg_st = txq_msg_st_alloc(&st->queue);
782 if (msg_st == NULL) {
786 msg_st->send = chan_open_call_tx;
787 msg_st->cleanup = NULL;
789 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
796 txq_send(msg_st);
814 struct txq_msg_st *msg_st = txq_msg_st_alloc(&st->queue);
815 if (msg_st == NULL) {
819 msg_st->send = domain_wait_response_tx;
820 msg_st->cleanup = NULL;
822 struct xphi_svc_msg_st *xphi_st = (struct xphi_svc_msg_st *) msg_st;
826 txq_send(msg_st);