/linux-master/drivers/net/wireless/ti/wlcore/ |
H A D | acx.c | 1492 conf_queues = wl->conf.rx_streaming.queues; 1500 * Skip non-changed queues, to avoid redundant acxs. 1501 * this check assumes conf.rx_streaming.queues can't
|
/linux-master/drivers/scsi/arm/ |
H A D | acornscsi.h | 302 unsigned int queues; member in struct:acornscsi_hostdata::__anon593 317 } queues; member in struct:acornscsi_hostdata
|
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_xive.c | 75 * in the queues. 161 q = &xc->queues[prio]; 397 struct xive_q *q = &xc->queues[prio]; 870 struct xive_q *q = &xc->queues[prio]; 940 struct xive_q *q = &xc->queues[prio]; 1021 q = &xc->queues[prio]; 1036 q = &xc->queues[prio]; 1121 * Synchronize hardware to sensure the queues are updated when 1241 * interrupts out of remote processor queues. 1284 /* First, check provisioning of queues */ [all...] |
H A D | book3s_xive.h | 116 * For state save, we lazily scan the queues on the first interrupt 131 /* Which queues (priorities) are in use by the guest */ 178 /* Each VP has 8 queues though we only provision some */ 179 struct xive_q queues[KVMPPC_XIVE_Q_COUNT]; member in struct:kvmppc_xive_vcpu
|
H A D | book3s_xive_native.c | 47 struct xive_q *q = &xc->queues[prio]; 110 /* Free the queues */ 189 /* TODO: reset all queues to a clean state ? */ 600 q = &xc->queues[priority]; 745 q = &xc->queues[priority]; 901 struct xive_q *q = &xc->queues[prio];
|
/linux-master/drivers/block/null_blk/ |
H A D | main.c | 90 MODULE_PARM_DESC(submit_queues, "Number of submission queues"); 94 MODULE_PARM_DESC(poll_queues, "Number of IOPOLL submission queues"); 186 MODULE_PARM_DESC(shared_tag_bitmap, "Use shared tag bitmap for all submission queues for blk-mq"); 214 MODULE_PARM_DESC(use_per_node_hctx, "Use per-node allocation for hardware context queues. Default: false"); 375 * Make sure that null_init_hctx() does not access nullb->queues[] past 1444 * number of hardware queues are prepared. If block layer failed 1445 * to prepare them, use previous numbers of submit queues and 1446 * poll queues to map queues. 1639 nq = &nullb->queues[hctx_id [all...] |
H A D | null_blk.h | 85 unsigned int submit_queues; /* number of submission queues */ 86 unsigned int prev_submit_queues; /* number of submission queues before change */ 87 unsigned int poll_queues; /* number of IOPOLL submission queues */ 88 unsigned int prev_poll_queues; /* number of IOPOLL submission queues before change */ 122 struct nullb_queue *queues; member in struct:nullb
|
/linux-master/drivers/crypto/ccp/ |
H A D | ccp-dev-v5.c | 590 int queues = 0; local 601 queues = bitmap_weight(cmd_q->lsbmask, MAX_LSB_CNT); 603 cmd_q->id, queues); 605 return queues ? 0 : -EINVAL; 660 * find an LSB that can be assigned to the queue. If there are N queues that 663 * If we have fewer LSBs than queues, all LSB regions become shared resources. 686 * Brute force search to start with the queues that are more 790 /* Find available queues */ 868 dev_notice(dev, "no command queues available\n"); 873 /* Turn off the queues an [all...] |
/linux-master/drivers/crypto/hisilicon/sec/ |
H A D | sec_drv.c | 639 if (!info->queues[i].in_use) { 640 info->queues[i].in_use = true; 644 return &info->queues[i]; 976 /* Same QoS for all queues */ 1236 ret = sec_queue_config(info, &info->queues[i], i); 1240 ret = sec_queue_irq_init(&info->queues[i]); 1242 sec_queue_unconfig(info, &info->queues[i]); 1266 sec_queue_irq_uninit(&info->queues[j]); 1267 sec_queue_unconfig(info, &info->queues[j]); 1285 sec_queue_irq_uninit(&info->queues[ [all...] |
H A D | sec_drv.h | 327 * use of the hardware queues. 328 * @havesoftqueue: A flag to say we have a queues - as we may need one for the 397 * struct sec_dev_info: The full SEC unit comprising queues and processors. 402 * @queues: The 16 queues that this SEC instance provides. 412 struct sec_queue queues[SEC_Q_NUM]; member in struct:sec_dev_info
|
/linux-master/drivers/dma/ |
H A D | fsl-qdma.c | 617 * all queues. 887 * all queues. 1121 u32 len, chans, queues; local 1166 ret = of_property_read_u32(np, "fsl,dma-queues", &queues); 1168 dev_err(&pdev->dev, "Can't get queues.\n"); 1174 fsl_qdma->n_queues = queues;
|
/linux-master/drivers/dma/ti/ |
H A D | cppi41.c | 320 /* not set all bits for queues less than Q 93 */ 950 const struct chan_queues *queues; local 965 queues = cdd->queues_tx; 967 queues = cdd->queues_rx; 974 cchan->q_num = queues[cchan->port_num].submit; 975 cchan->q_comp_num = queues[cchan->port_num].complete;
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_chardev.c | 346 * on unmapped queues for usermode queue oversubscription (no aggregated doorbell) 1694 if (!list_empty(&p->pqm.queues)) { 1695 pr_debug("Process has user queues running\n"); 2066 /* Confirm all process queues are evicted */ 2068 pr_err("Cannot dump process when queues are not in evicted state\n"); 2561 * Set the process to evicted state to avoid running any new queues before all the memory 2615 pr_err("Failed to unpause queues ret:%d\n", ret);
|
H A D | kfd_debug.c | 51 list_for_each_entry(pqn, &pqm->queues, process_queue_list) { 160 list_for_each_entry(pqn, &pqm->queues, 222 list_for_each_entry(pqn, &pqm->queues, 327 list_for_each_entry(pqn, &pqm->queues, process_queue_list) { 336 list_for_each_entry(pqn, &pqm->queues, process_queue_list) 565 pr_debug("Resumed %d queues\n", resume_count); 624 list_for_each_entry(pqn, &pqm->queues, process_queue_list) { 1103 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
|
H A D | kfd_device_queue_manager.c | 134 /* Mask out the reserved queues */ 294 list_for_each_entry(cur, &dqm->queues, list) { 364 /* For SDMA queues on SOC15 with 8-byte doorbell, use static 389 /* For CP queues on SOC15 */ 547 pr_warn("Can't create new usermode queue because %d queues were already created\n", 560 * Eviction state logic: mark all queues as evicted, even ones 561 * not currently active. Restoring inactive queues later only 634 pr_debug("Total of %d queues are accountable so far\n", 965 * called on multiple queues in a loop, so rather than locking/unlocking 1011 * called on multiple queues i [all...] |
H A D | kfd_device_queue_manager.h | 91 * @exeute_queues: Dispatches the queues list to the H/W. 99 * @start: Initializes the resources/modules the device needs for queues 116 * @process_termination: Clears all process queues belongs to that device. 118 * @evict_process_queues: Evict all active queues of a process 120 * @restore_process_queues: Restore all evicted queues of a process 125 * @reset_queues: reset queues which consume RAS poison 217 * This struct is a base class for the kfd queues scheduler in the 221 * concrete device. This class is the only class in the queues scheduler 234 struct list_head queues; member in struct:device_queue_manager
|
H A D | kfd_packet_manager.c | 122 struct list_head *queues, 145 pr_debug("Building runlist ib process count: %d queues count %d\n", 149 list_for_each_entry(cur, queues, list) { 206 pr_debug("Finished map process and queues to runlist\n"); 121 pm_create_runlist_ib(struct packet_manager *pm, struct list_head *queues, uint64_t *rl_gpu_addr, size_t *rl_size_bytes) argument
|
H A D | kfd_priv.h | 85 * definitions for Kaveri. In Kaveri only the first ME queues participates 149 * Kernel module parameter to specify maximum number of supported queues per 191 * Don't evict process queues on vm fault 388 * @KFD_UNMAP_QUEUES_FILTER_ALL_QUEUES: Preempts all queues in the 389 * running queues list. 391 * @KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES: Preempts all non-static queues 394 * @KFD_UNMAP_QUEUES_FILTER_BY_PASID: Preempts queues that belongs to 446 * @priority: Defines the queue priority relative to other queues in the 451 * currently all queues are initialized with the highest priority. 470 * @is_evicted: Defines if the queue is evicted. Only active queues 636 struct list_head queues; member in struct:process_queue_manager [all...] |
H A D | kfd_process_queue_manager.c | 37 list_for_each_entry(pqn, &pqm->queues, process_queue_list) { 71 pr_info("Cannot open more queues for process with pasid 0x%x\n", 166 INIT_LIST_HEAD(&pqm->queues); 209 list_for_each_entry_safe(pqn, next, &pqm->queues, process_queue_list) { 290 * On GFX 9.4.3, increase the number of queues that 306 * for debug process, verify that it is within the static queues limit 339 /* SDMA queues are always allocated statically no matter 420 list_add(&pqn->process_queue_list, &pqm->queues); 436 /* check if queues list is empty unregister process from device */ 643 list_for_each_entry(pqn, &pqm->queues, process_queue_lis [all...] |
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_context.c | 170 * pvr_context_destroy_queues() - Destroy all queues attached to a context. 171 * @ctx: Context to destroy queues on. 174 * It releases all resources attached to the queues bound to this context. 180 pvr_queue_destroy(ctx->queues.fragment); 181 pvr_queue_destroy(ctx->queues.geometry); 184 pvr_queue_destroy(ctx->queues.compute); 187 pvr_queue_destroy(ctx->queues.transfer); 193 * pvr_context_create_queues() - Create all queues attached to a context. 194 * @ctx: Context to create queues on. 210 ctx->queues [all...] |
H A D | pvr_context.h | 64 * @faulty: Set to 1 when the context queues had unfinished job when 72 /** @queues: Union containing all kind of queues. */ 87 } queues; member in struct:pvr_context 95 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.geometry : NULL; 97 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.fragment : NULL; 99 return ctx->type == DRM_PVR_CTX_TYPE_COMPUTE ? ctx->queues.compute : NULL; 101 return ctx->type == DRM_PVR_CTX_TYPE_TRANSFER_FRAG ? ctx->queues.transfer : NULL;
|
H A D | pvr_device.c | 128 * active queues and calls pvr_queue_process() on them. 135 mutex_lock(&pvr_dev->queues.lock); 137 /* Move all active queues to a temporary list. Queues that remain 139 * the queues.active list by pvr_queue_process(). 141 list_splice_init(&pvr_dev->queues.active, &active_queues); 146 mutex_unlock(&pvr_dev->queues.lock);
|
H A D | pvr_device.h | 193 * @queues: Queue-related fields. 196 /** @queues.active: Active queue list. */ 199 /** @queues.idle: Idle queue list. */ 202 /** @queues.lock: Lock protecting access to the active/idle 205 } queues; member in struct:pvr_device
|
H A D | pvr_power.c | 162 mutex_lock(&pvr_dev->queues.lock); 163 has_active_contexts = list_empty(&pvr_dev->queues.active); 164 mutex_unlock(&pvr_dev->queues.lock);
|
H A D | pvr_queue.c | 523 job->ctx->queues.fragment); 564 lockdep_assert_held(&pvr_dev->queues.lock); 574 list_move_tail(&queue->node, &pvr_dev->queues.idle); 576 list_move_tail(&queue->node, &pvr_dev->queues.active); 596 mutex_lock(&pvr_dev->queues.lock); 598 mutex_unlock(&pvr_dev->queues.lock); 730 struct pvr_queue *geom_queue = job->ctx->queues.geometry; 731 struct pvr_queue *frag_queue = job->ctx->queues.fragment; 765 /* Make sure we CPU-signal the UFO object, so other queues don't get 819 mutex_lock(&pvr_dev->queues [all...] |