/linux-master/drivers/dma/ptdma/ |
H A D | ptdma-dev.c | 57 void pt_start_queue(struct pt_cmd_queue *cmd_q) argument 60 iowrite32(cmd_q->qcontrol | CMD_Q_RUN, cmd_q->reg_control); 63 void pt_stop_queue(struct pt_cmd_queue *cmd_q) argument 66 iowrite32(cmd_q->qcontrol & ~CMD_Q_RUN, cmd_q->reg_control); 69 static int pt_core_execute_cmd(struct ptdma_desc *desc, struct pt_cmd_queue *cmd_q) argument 72 u8 *q_desc = (u8 *)&cmd_q->qbase[cmd_q->qidx]; 80 spin_lock_irqsave(&cmd_q 100 pt_core_perform_passthru(struct pt_cmd_queue *cmd_q, struct pt_passthru_engine *pt_engine) argument 128 struct pt_cmd_queue *cmd_q = &cmd->pt->cmd_q; local 144 pt_check_status_trans(struct pt_device *pt, struct pt_cmd_queue *cmd_q) argument 167 struct pt_cmd_queue *cmd_q = &pt->cmd_q; local 179 struct pt_cmd_queue *cmd_q = &pt->cmd_q; local 280 struct pt_cmd_queue *cmd_q = &pt->cmd_q; local [all...] |
H A D | ptdma-debugfs.c | 57 struct pt_cmd_queue *cmd_q = s->private; local 60 if (!cmd_q) 63 seq_printf(s, " Pass-Thru: %ld\n", cmd_q->total_pt_ops); 65 regval = ioread32(cmd_q->reg_control + 0x000C); 87 struct pt_cmd_queue *cmd_q; local 99 cmd_q = &pt->cmd_q; 104 debugfs_create_file("stats", 0400, debugfs_q_instance, cmd_q,
|
H A D | ptdma-dmaengine.c | 53 struct pt_cmd_queue *cmd_q; local 59 cmd_q = &pt->cmd_q; 65 pt_cmd->ret = pt_core_perform_passthru(cmd_q, pt_engine); 174 desc->pt->cmd_q.int_en = !!(flags & DMA_PREP_INTERRUPT); 266 struct pt_cmd_queue *cmd_q = &pt->cmd_q; local 268 pt_check_status_trans(pt, cmd_q); 278 pt_stop_queue(&chan->pt->cmd_q); 291 pt_start_queue(&chan->pt->cmd_q); 306 struct pt_cmd_queue *cmd_q = &chan->pt->cmd_q; local [all...] |
H A D | ptdma.h | 251 struct pt_cmd_queue cmd_q; member in struct:pt_device 321 int pt_core_perform_passthru(struct pt_cmd_queue *cmd_q, 324 void pt_check_status_trans(struct pt_device *pt, struct pt_cmd_queue *cmd_q); 325 void pt_start_queue(struct pt_cmd_queue *cmd_q); 326 void pt_stop_queue(struct pt_cmd_queue *cmd_q); 330 iowrite32(0, pt->cmd_q.reg_control + 0x000C); 335 iowrite32(SUPPORTED_INTERRUPTS, pt->cmd_q.reg_control + 0x000C);
|
/linux-master/drivers/crypto/ccp/ |
H A D | ccp-dev-v3.c | 19 static u32 ccp_alloc_ksb(struct ccp_cmd_queue *cmd_q, unsigned int count) argument 22 struct ccp_device *ccp = cmd_q->ccp; 50 static void ccp_free_ksb(struct ccp_cmd_queue *cmd_q, unsigned int start, argument 53 struct ccp_device *ccp = cmd_q->ccp; 69 static unsigned int ccp_get_free_slots(struct ccp_cmd_queue *cmd_q) argument 71 return CMD_Q_DEPTH(ioread32(cmd_q->reg_status)); 76 struct ccp_cmd_queue *cmd_q = op->cmd_q; local 77 struct ccp_device *ccp = cmd_q->ccp; 87 cmd_q 327 struct ccp_cmd_queue *cmd_q; local 372 struct ccp_cmd_queue *cmd_q; local 519 struct ccp_cmd_queue *cmd_q; local [all...] |
H A D | ccp-dev-v5.c | 25 static u32 ccp_lsb_alloc(struct ccp_cmd_queue *cmd_q, unsigned int count) argument 31 if (cmd_q->lsb >= 0) { 32 start = (u32)bitmap_find_next_zero_area(cmd_q->lsbmap, 36 bitmap_set(cmd_q->lsbmap, start, count); 37 return start + cmd_q->lsb * LSB_SIZE; 42 ccp = cmd_q->ccp; 70 static void ccp_lsb_free(struct ccp_cmd_queue *cmd_q, unsigned int start, argument 76 if (cmd_q->lsb == start) { 78 bitmap_clear(cmd_q->lsbmap, start, count); 81 struct ccp_device *ccp = cmd_q 207 ccp5_get_free_slots(struct ccp_cmd_queue *cmd_q) argument 221 ccp5_do_cmd(struct ccp5_desc *desc, struct ccp_cmd_queue *cmd_q) argument 587 ccp_find_lsb_regions(struct ccp_cmd_queue *cmd_q, u64 status) argument 628 struct ccp_cmd_queue *cmd_q = &ccp->cmd_q[i]; local 743 struct ccp_cmd_queue *cmd_q = &ccp->cmd_q[i]; local 782 struct ccp_cmd_queue *cmd_q; local 1008 struct ccp_cmd_queue *cmd_q; local [all...] |
H A D | ccp-debugfs.c | 115 struct ccp_cmd_queue *cmd_q = &ccp->cmd_q[i]; local 117 total_ops += cmd_q->total_ops; 118 total_aes_ops += cmd_q->total_aes_ops; 119 total_xts_aes_ops += cmd_q->total_xts_aes_ops; 120 total_3des_ops += cmd_q->total_3des_ops; 121 total_sha_ops += cmd_q->total_sha_ops; 122 total_rsa_ops += cmd_q->total_rsa_ops; 123 total_pt_ops += cmd_q->total_pt_ops; 124 total_ecc_ops += cmd_q 158 ccp5_debugfs_reset_queue_stats(struct ccp_cmd_queue *cmd_q) argument 194 struct ccp_cmd_queue *cmd_q = filp->private_data; local 249 struct ccp_cmd_queue *cmd_q = filp->private_data; local 284 struct ccp_cmd_queue *cmd_q; local [all...] |
H A D | ccp-ops.c | 153 struct ccp_cmd_queue *cmd_q, 162 wa->dev = cmd_q->ccp->dev; 166 wa->dma_pool = cmd_q->dma_pool; 265 static void ccp_free_data(struct ccp_data *data, struct ccp_cmd_queue *cmd_q) argument 271 static int ccp_init_data(struct ccp_data *data, struct ccp_cmd_queue *cmd_q, argument 280 ret = ccp_init_sg_workarea(&data->sg_wa, cmd_q->ccp->dev, sg, sg_len, 285 ret = ccp_init_dm_workarea(&data->dm_wa, cmd_q, dm_len, dir); 292 ccp_free_data(data, cmd_q); 429 static int ccp_copy_to_from_sb(struct ccp_cmd_queue *cmd_q, argument 437 op.cmd_q 152 ccp_init_dm_workarea(struct ccp_dm_workarea *wa, struct ccp_cmd_queue *cmd_q, unsigned int len, enum dma_data_direction dir) argument 461 ccp_copy_to_sb(struct ccp_cmd_queue *cmd_q, struct ccp_dm_workarea *wa, u32 jobid, u32 sb, u32 byte_swap) argument 468 ccp_copy_from_sb(struct ccp_cmd_queue *cmd_q, struct ccp_dm_workarea *wa, u32 jobid, u32 sb, u32 byte_swap) argument 476 ccp_run_aes_cmac_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 632 ccp_run_aes_gcm_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 916 ccp_run_aes_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 1093 ccp_run_xts_aes_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 1293 ccp_run_des3_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 1489 ccp_run_sha_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 1835 ccp_run_rsa_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 1967 ccp_run_passthru_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 2099 ccp_run_passthru_nomap_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 2163 ccp_run_ecc_mm_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 2270 ccp_run_ecc_pm_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 2440 ccp_run_ecc_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument 2466 ccp_run_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd) argument [all...] |
H A D | ccp-dev.c | 327 if (ccp->cmd_q[i].active) 339 wake_up_process(ccp->cmd_q[i].kthread); 361 if (ccp->cmd_q[i].active) 371 wake_up_process(ccp->cmd_q[i].kthread); 374 static struct ccp_cmd *ccp_dequeue_cmd(struct ccp_cmd_queue *cmd_q) argument 376 struct ccp_device *ccp = cmd_q->ccp; 383 cmd_q->active = 0; 386 cmd_q->suspended = 1; 395 cmd_q->active = 1; 436 struct ccp_cmd_queue *cmd_q local [all...] |
H A D | ccp-dev.h | 380 struct ccp_cmd_queue cmd_q[MAX_HW_QUEUES]; member in struct:ccp_device 532 struct ccp_cmd_queue *cmd_q; member in struct:ccp_op 641 int ccp_run_cmd(struct ccp_cmd_queue *cmd_q, struct ccp_cmd *cmd);
|
/linux-master/net/bluetooth/ |
H A D | hci_request.c | 37 skb_queue_head_init(&req->cmd_q); 44 skb_queue_purge(&req->cmd_q); 59 bt_dev_dbg(hdev, "length %u", skb_queue_len(&req->cmd_q)); 65 skb_queue_purge(&req->cmd_q); 70 if (skb_queue_empty(&req->cmd_q)) 73 skb = skb_peek_tail(&req->cmd_q); 81 spin_lock_irqsave(&hdev->cmd_q.lock, flags); 82 skb_queue_splice_tail(&req->cmd_q, &hdev->cmd_q); 83 spin_unlock_irqrestore(&hdev->cmd_q [all...] |
H A D | hci_request.h | 38 struct sk_buff_head cmd_q; member in struct:hci_request
|
H A D | hci_core.c | 596 skb_queue_purge(&hdev->cmd_q); 2604 skb_queue_head_init(&hdev->cmd_q); 3123 skb_queue_tail(&hdev->cmd_q, skb); 4025 skb = skb_peek(&hdev->cmd_q); 4050 skb_queue_head(&hdev->cmd_q, skb); 4092 * command queue (hdev->cmd_q). 4105 spin_lock_irqsave(&hdev->cmd_q.lock, flags); 4106 while ((skb = __skb_dequeue(&hdev->cmd_q))) { 4108 __skb_queue_head(&hdev->cmd_q, skb); 4118 spin_unlock_irqrestore(&hdev->cmd_q [all...] |
H A D | hci_sync.c | 109 if (skb_queue_empty(&req->cmd_q)) 114 skb_queue_tail(&req->cmd_q, skb); 123 bt_dev_dbg(hdev, "length %u", skb_queue_len(&req->cmd_q)); 129 skb_queue_purge(&req->cmd_q); 134 if (skb_queue_empty(&req->cmd_q)) 137 skb = skb_peek_tail(&req->cmd_q); 141 spin_lock_irqsave(&hdev->cmd_q.lock, flags); 142 skb_queue_splice_tail(&req->cmd_q, &hdev->cmd_q); 143 spin_unlock_irqrestore(&hdev->cmd_q [all...] |
/linux-master/drivers/scsi/ibmvscsi_tgt/ |
H A D | ibmvscsi_tgt.c | 278 bytes = vscsi->cmd_q.size * PAGE_SIZE; 279 memset(vscsi->cmd_q.base_addr, 0, bytes); 280 vscsi->cmd_q.index = 0; 363 crq = ibmvscsis_cmd_q_dequeue(vscsi->cmd_q.mask, &vscsi->cmd_q.index, 364 vscsi->cmd_q.base_addr); 377 crq = ibmvscsis_cmd_q_dequeue(vscsi->cmd_q.mask, 378 &vscsi->cmd_q.index, 379 vscsi->cmd_q.base_addr); 896 bytes = vscsi->cmd_q [all...] |
H A D | ibmvscsi_tgt.h | 268 struct cmd_queue cmd_q; member in struct:scsi_info
|
/linux-master/drivers/platform/olpc/ |
H A D | olpc-ec.c | 43 struct list_head cmd_q; member in struct:olpc_ec_priv 85 if (!list_empty(&ec->cmd_q)) { 86 desc = list_first_entry(&ec->cmd_q, struct ec_cmd_desc, node); 120 list_add_tail(&desc->node, &ec->cmd_q); 422 INIT_LIST_HEAD(&ec->cmd_q);
|
/linux-master/drivers/media/platform/amphion/ |
H A D | vpu_cmds.c | 150 list_for_each_entry_safe(cmd, tmp, &inst->cmd_q, list) { 184 list_add_tail(&cmd->list, &inst->cmd_q); 244 list_for_each_entry_safe(cmd, tmp, &inst->cmd_q, list) {
|
H A D | vpu.h | 253 struct list_head cmd_q; member in struct:vpu_inst
|
/linux-master/net/nfc/nci/ |
H A D | core.c | 541 skb_queue_purge(&ndev->cmd_q); 583 skb_queue_purge(&ndev->cmd_q); 1256 skb_queue_head_init(&ndev->cmd_q); 1385 skb_queue_tail(&ndev->cmd_q, skb); 1568 skb = skb_dequeue(&ndev->cmd_q);
|
H A D | rsp.c | 426 if (!skb_queue_empty(&ndev->cmd_q))
|
/linux-master/drivers/net/ethernet/brocade/bna/ |
H A D | bfa_ioc.c | 2120 INIT_LIST_HEAD(&mod->cmd_q); 2140 if (list_empty(&mod->cmd_q)) 2153 cmd = list_first_entry(&mod->cmd_q, struct bfa_mbox_cmd, qe); 2175 while (!list_empty(&mod->cmd_q)) { 2176 cmd = list_first_entry(&mod->cmd_q, struct bfa_mbox_cmd, qe); 2662 if (!list_empty(&mod->cmd_q)) { 2663 list_add_tail(&cmd->qe, &mod->cmd_q); 2672 list_add_tail(&cmd->qe, &mod->cmd_q);
|
H A D | bfa_ioc.h | 109 struct list_head cmd_q; /*!< pending mbox queue */ member in struct:bfa_ioc_mbox_mod
|
/linux-master/drivers/scsi/ |
H A D | sg.c | 153 char cmd_q; /* 1 -> allow command queuing, 0 -> don't */ member in struct:sg_fd 729 sfp->cmd_q = 1; /* when sg_io_hdr seen, set command queuing on */ 1054 sfp->cmd_q = val ? 1 : 0; 1057 return put_user((int) sfp->cmd_q, ip); 1197 else if (!sfp->cmd_q) { 2111 if (!sfp->cmd_q) 2179 sfp->cmd_q = SG_DEF_COMMAND_Q; 2548 seq_printf(s, " cmd_q=%d f_packid=%d k_orphan=%d closed=0\n", 2549 (int) fp->cmd_q, (int) fp->force_packid,
|
/linux-master/drivers/net/wireless/intel/iwlwifi/mvm/ |
H A D | utils.c | 720 bool tdls, bool cmd_q) 724 unsigned int default_timeout = cmd_q ? 746 if (cmd_q) 718 iwl_mvm_get_wd_timeout(struct iwl_mvm *mvm, struct ieee80211_vif *vif, bool tdls, bool cmd_q) argument
|