Searched refs:queues (Results 1 - 25 of 217) sorted by path

123456789

/linux-master/drivers/net/wireless/ti/wlcore/
H A Dacx.c1492 conf_queues = wl->conf.rx_streaming.queues;
1500 * Skip non-changed queues, to avoid redundant acxs.
1501 * this check assumes conf.rx_streaming.queues can't
/linux-master/drivers/scsi/arm/
H A Dacornscsi.h302 unsigned int queues; member in struct:acornscsi_hostdata::__anon593
317 } queues; member in struct:acornscsi_hostdata
/linux-master/arch/powerpc/kvm/
H A Dbook3s_xive.c75 * in the queues.
161 q = &xc->queues[prio];
397 struct xive_q *q = &xc->queues[prio];
870 struct xive_q *q = &xc->queues[prio];
940 struct xive_q *q = &xc->queues[prio];
1021 q = &xc->queues[prio];
1036 q = &xc->queues[prio];
1121 * Synchronize hardware to sensure the queues are updated when
1241 * interrupts out of remote processor queues.
1284 /* First, check provisioning of queues */
[all...]
H A Dbook3s_xive.h116 * For state save, we lazily scan the queues on the first interrupt
131 /* Which queues (priorities) are in use by the guest */
178 /* Each VP has 8 queues though we only provision some */
179 struct xive_q queues[KVMPPC_XIVE_Q_COUNT]; member in struct:kvmppc_xive_vcpu
H A Dbook3s_xive_native.c47 struct xive_q *q = &xc->queues[prio];
110 /* Free the queues */
189 /* TODO: reset all queues to a clean state ? */
600 q = &xc->queues[priority];
745 q = &xc->queues[priority];
901 struct xive_q *q = &xc->queues[prio];
/linux-master/drivers/block/null_blk/
H A Dmain.c90 MODULE_PARM_DESC(submit_queues, "Number of submission queues");
94 MODULE_PARM_DESC(poll_queues, "Number of IOPOLL submission queues");
186 MODULE_PARM_DESC(shared_tag_bitmap, "Use shared tag bitmap for all submission queues for blk-mq");
214 MODULE_PARM_DESC(use_per_node_hctx, "Use per-node allocation for hardware context queues. Default: false");
375 * Make sure that null_init_hctx() does not access nullb->queues[] past
1444 * number of hardware queues are prepared. If block layer failed
1445 * to prepare them, use previous numbers of submit queues and
1446 * poll queues to map queues.
1639 nq = &nullb->queues[hctx_id
[all...]
H A Dnull_blk.h85 unsigned int submit_queues; /* number of submission queues */
86 unsigned int prev_submit_queues; /* number of submission queues before change */
87 unsigned int poll_queues; /* number of IOPOLL submission queues */
88 unsigned int prev_poll_queues; /* number of IOPOLL submission queues before change */
122 struct nullb_queue *queues; member in struct:nullb
/linux-master/drivers/crypto/ccp/
H A Dccp-dev-v5.c590 int queues = 0; local
601 queues = bitmap_weight(cmd_q->lsbmask, MAX_LSB_CNT);
603 cmd_q->id, queues);
605 return queues ? 0 : -EINVAL;
660 * find an LSB that can be assigned to the queue. If there are N queues that
663 * If we have fewer LSBs than queues, all LSB regions become shared resources.
686 * Brute force search to start with the queues that are more
790 /* Find available queues */
868 dev_notice(dev, "no command queues available\n");
873 /* Turn off the queues an
[all...]
/linux-master/drivers/crypto/hisilicon/sec/
H A Dsec_drv.c639 if (!info->queues[i].in_use) {
640 info->queues[i].in_use = true;
644 return &info->queues[i];
976 /* Same QoS for all queues */
1236 ret = sec_queue_config(info, &info->queues[i], i);
1240 ret = sec_queue_irq_init(&info->queues[i]);
1242 sec_queue_unconfig(info, &info->queues[i]);
1266 sec_queue_irq_uninit(&info->queues[j]);
1267 sec_queue_unconfig(info, &info->queues[j]);
1285 sec_queue_irq_uninit(&info->queues[
[all...]
H A Dsec_drv.h327 * use of the hardware queues.
328 * @havesoftqueue: A flag to say we have a queues - as we may need one for the
397 * struct sec_dev_info: The full SEC unit comprising queues and processors.
402 * @queues: The 16 queues that this SEC instance provides.
412 struct sec_queue queues[SEC_Q_NUM]; member in struct:sec_dev_info
/linux-master/drivers/dma/
H A Dfsl-qdma.c617 * all queues.
887 * all queues.
1121 u32 len, chans, queues; local
1166 ret = of_property_read_u32(np, "fsl,dma-queues", &queues);
1168 dev_err(&pdev->dev, "Can't get queues.\n");
1174 fsl_qdma->n_queues = queues;
/linux-master/drivers/dma/ti/
H A Dcppi41.c320 /* not set all bits for queues less than Q 93 */
950 const struct chan_queues *queues; local
965 queues = cdd->queues_tx;
967 queues = cdd->queues_rx;
974 cchan->q_num = queues[cchan->port_num].submit;
975 cchan->q_comp_num = queues[cchan->port_num].complete;
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_chardev.c346 * on unmapped queues for usermode queue oversubscription (no aggregated doorbell)
1694 if (!list_empty(&p->pqm.queues)) {
1695 pr_debug("Process has user queues running\n");
2066 /* Confirm all process queues are evicted */
2068 pr_err("Cannot dump process when queues are not in evicted state\n");
2561 * Set the process to evicted state to avoid running any new queues before all the memory
2615 pr_err("Failed to unpause queues ret:%d\n", ret);
H A Dkfd_debug.c51 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
160 list_for_each_entry(pqn, &pqm->queues,
222 list_for_each_entry(pqn, &pqm->queues,
327 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
336 list_for_each_entry(pqn, &pqm->queues, process_queue_list)
565 pr_debug("Resumed %d queues\n", resume_count);
624 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
1103 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
H A Dkfd_device_queue_manager.c134 /* Mask out the reserved queues */
294 list_for_each_entry(cur, &dqm->queues, list) {
364 /* For SDMA queues on SOC15 with 8-byte doorbell, use static
389 /* For CP queues on SOC15 */
547 pr_warn("Can't create new usermode queue because %d queues were already created\n",
560 * Eviction state logic: mark all queues as evicted, even ones
561 * not currently active. Restoring inactive queues later only
634 pr_debug("Total of %d queues are accountable so far\n",
965 * called on multiple queues in a loop, so rather than locking/unlocking
1011 * called on multiple queues i
[all...]
H A Dkfd_device_queue_manager.h91 * @exeute_queues: Dispatches the queues list to the H/W.
99 * @start: Initializes the resources/modules the device needs for queues
116 * @process_termination: Clears all process queues belongs to that device.
118 * @evict_process_queues: Evict all active queues of a process
120 * @restore_process_queues: Restore all evicted queues of a process
125 * @reset_queues: reset queues which consume RAS poison
217 * This struct is a base class for the kfd queues scheduler in the
221 * concrete device. This class is the only class in the queues scheduler
234 struct list_head queues; member in struct:device_queue_manager
H A Dkfd_packet_manager.c122 struct list_head *queues,
145 pr_debug("Building runlist ib process count: %d queues count %d\n",
149 list_for_each_entry(cur, queues, list) {
206 pr_debug("Finished map process and queues to runlist\n");
121 pm_create_runlist_ib(struct packet_manager *pm, struct list_head *queues, uint64_t *rl_gpu_addr, size_t *rl_size_bytes) argument
H A Dkfd_priv.h85 * definitions for Kaveri. In Kaveri only the first ME queues participates
149 * Kernel module parameter to specify maximum number of supported queues per
191 * Don't evict process queues on vm fault
388 * @KFD_UNMAP_QUEUES_FILTER_ALL_QUEUES: Preempts all queues in the
389 * running queues list.
391 * @KFD_UNMAP_QUEUES_FILTER_DYNAMIC_QUEUES: Preempts all non-static queues
394 * @KFD_UNMAP_QUEUES_FILTER_BY_PASID: Preempts queues that belongs to
446 * @priority: Defines the queue priority relative to other queues in the
451 * currently all queues are initialized with the highest priority.
470 * @is_evicted: Defines if the queue is evicted. Only active queues
636 struct list_head queues; member in struct:process_queue_manager
[all...]
H A Dkfd_process_queue_manager.c37 list_for_each_entry(pqn, &pqm->queues, process_queue_list) {
71 pr_info("Cannot open more queues for process with pasid 0x%x\n",
166 INIT_LIST_HEAD(&pqm->queues);
209 list_for_each_entry_safe(pqn, next, &pqm->queues, process_queue_list) {
290 * On GFX 9.4.3, increase the number of queues that
306 * for debug process, verify that it is within the static queues limit
339 /* SDMA queues are always allocated statically no matter
420 list_add(&pqn->process_queue_list, &pqm->queues);
436 /* check if queues list is empty unregister process from device */
643 list_for_each_entry(pqn, &pqm->queues, process_queue_lis
[all...]
/linux-master/drivers/gpu/drm/imagination/
H A Dpvr_context.c170 * pvr_context_destroy_queues() - Destroy all queues attached to a context.
171 * @ctx: Context to destroy queues on.
174 * It releases all resources attached to the queues bound to this context.
180 pvr_queue_destroy(ctx->queues.fragment);
181 pvr_queue_destroy(ctx->queues.geometry);
184 pvr_queue_destroy(ctx->queues.compute);
187 pvr_queue_destroy(ctx->queues.transfer);
193 * pvr_context_create_queues() - Create all queues attached to a context.
194 * @ctx: Context to create queues on.
210 ctx->queues
[all...]
H A Dpvr_context.h64 * @faulty: Set to 1 when the context queues had unfinished job when
72 /** @queues: Union containing all kind of queues. */
87 } queues; member in struct:pvr_context
95 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.geometry : NULL;
97 return ctx->type == DRM_PVR_CTX_TYPE_RENDER ? ctx->queues.fragment : NULL;
99 return ctx->type == DRM_PVR_CTX_TYPE_COMPUTE ? ctx->queues.compute : NULL;
101 return ctx->type == DRM_PVR_CTX_TYPE_TRANSFER_FRAG ? ctx->queues.transfer : NULL;
H A Dpvr_device.c128 * active queues and calls pvr_queue_process() on them.
135 mutex_lock(&pvr_dev->queues.lock);
137 /* Move all active queues to a temporary list. Queues that remain
139 * the queues.active list by pvr_queue_process().
141 list_splice_init(&pvr_dev->queues.active, &active_queues);
146 mutex_unlock(&pvr_dev->queues.lock);
H A Dpvr_device.h193 * @queues: Queue-related fields.
196 /** @queues.active: Active queue list. */
199 /** @queues.idle: Idle queue list. */
202 /** @queues.lock: Lock protecting access to the active/idle
205 } queues; member in struct:pvr_device
H A Dpvr_power.c162 mutex_lock(&pvr_dev->queues.lock);
163 has_active_contexts = list_empty(&pvr_dev->queues.active);
164 mutex_unlock(&pvr_dev->queues.lock);
H A Dpvr_queue.c523 job->ctx->queues.fragment);
564 lockdep_assert_held(&pvr_dev->queues.lock);
574 list_move_tail(&queue->node, &pvr_dev->queues.idle);
576 list_move_tail(&queue->node, &pvr_dev->queues.active);
596 mutex_lock(&pvr_dev->queues.lock);
598 mutex_unlock(&pvr_dev->queues.lock);
730 struct pvr_queue *geom_queue = job->ctx->queues.geometry;
731 struct pvr_queue *frag_queue = job->ctx->queues.fragment;
765 /* Make sure we CPU-signal the UFO object, so other queues don't get
819 mutex_lock(&pvr_dev->queues
[all...]

Completed in 563 milliseconds

123456789