Lines Matching refs:task_params

48 static inline u32 calc_rw_task_size(struct nvmetcp_task_params *task_params,
54 io_size = task_params->tx_io_size;
56 io_size = task_params->rx_io_size;
64 static inline void init_sqe(struct nvmetcp_task_params *task_params,
68 if (!task_params->sqe)
71 memset(task_params->sqe, 0, sizeof(*task_params->sqe));
72 task_params->sqe->task_id = cpu_to_le16(task_params->itid);
79 SET_FIELD(task_params->sqe->contlen_cdbsize,
81 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_WQE_TYPE,
83 if (task_params->tx_io_size) {
84 if (task_params->send_write_incapsule)
85 buf_size = calc_rw_task_size(task_params, task_type);
94 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_NUM_SGES, num_sges);
95 SET_FIELD(task_params->sqe->contlen_cdbsize, NVMETCP_WQE_CONT_LEN, buf_size);
99 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_WQE_TYPE,
101 SET_FIELD(task_params->sqe->contlen_cdbsize,
106 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_WQE_TYPE,
109 if (task_params->tx_io_size) {
110 SET_FIELD(task_params->sqe->contlen_cdbsize, NVMETCP_WQE_CONT_LEN,
111 task_params->tx_io_size);
112 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_NUM_SGES,
119 SET_FIELD(task_params->sqe->flags, NVMETCP_WQE_WQE_TYPE,
130 struct nvmetcp_task_params *task_params,
133 context->ystorm_st_context.state.cccid = task_params->host_cccid;
135 context->ustorm_st_context.nvme_tcp_opaque_lo = cpu_to_le32(task_params->opq.lo);
136 context->ustorm_st_context.nvme_tcp_opaque_hi = cpu_to_le32(task_params->opq.hi);
141 init_default_nvmetcp_task(struct nvmetcp_task_params *task_params,
145 struct e5_nvmetcp_task_context *context = task_params->context;
150 init_nvmetcp_task_params(context, task_params,
178 context->mstorm_ag_context.task_cid = cpu_to_le16(task_params->conn_icid);
183 context->ustorm_st_context.cq_rss_number = task_params->cq_rss_number;
184 context->ustorm_ag_context.icid = cpu_to_le16(task_params->conn_icid);
216 init_rw_nvmetcp_task(struct nvmetcp_task_params *task_params,
221 struct e5_nvmetcp_task_context *context = task_params->context;
222 u32 task_size = calc_rw_task_size(task_params, task_type);
226 init_default_nvmetcp_task(task_params, pdu_header, nvme_cmd, task_type);
229 if (task_params->tx_io_size) {
244 } else if (task_params->rx_io_size) {
271 if (task_params->send_write_incapsule)
280 init_sqe(task_params, sgl_task_params, task_type);
284 init_common_initiator_read_task(struct nvmetcp_task_params *task_params,
289 init_rw_nvmetcp_task(task_params, NVMETCP_TASK_TYPE_HOST_READ,
293 void init_nvmetcp_host_read_task(struct nvmetcp_task_params *task_params,
298 init_common_initiator_read_task(task_params, (void *)cmd_pdu_header,
303 init_common_initiator_write_task(struct nvmetcp_task_params *task_params,
308 init_rw_nvmetcp_task(task_params, NVMETCP_TASK_TYPE_HOST_WRITE,
312 void init_nvmetcp_host_write_task(struct nvmetcp_task_params *task_params,
317 init_common_initiator_write_task(task_params, (void *)cmd_pdu_header,
322 init_common_login_request_task(struct nvmetcp_task_params *task_params,
327 struct e5_nvmetcp_task_context *context = task_params->context;
329 init_default_nvmetcp_task(task_params, (void *)login_req_pdu_header, NULL,
337 task_params->rx_io_size ?
341 task_params->tx_io_size ?
347 if (task_params->tx_io_size)
351 if (task_params->rx_io_size)
357 cpu_to_le32(task_params->rx_io_size ?
359 init_sqe(task_params, tx_sgl_task_params, NVMETCP_TASK_TYPE_INIT_CONN_REQUEST);
363 void init_nvmetcp_init_conn_req_task(struct nvmetcp_task_params *task_params,
368 init_common_login_request_task(task_params, init_conn_req_pdu_hdr,
372 void init_cleanup_task_nvmetcp(struct nvmetcp_task_params *task_params)
374 init_sqe(task_params, NULL, NVMETCP_TASK_TYPE_CLEANUP);