Searched refs:queue (Results 26 - 50 of 1368) sorted by relevance

1234567891011>>

/linux-master/drivers/nvme/host/
H A Dtcp.c105 struct nvme_tcp_queue *queue; member in struct:nvme_tcp_request
204 static int nvme_tcp_try_send(struct nvme_tcp_queue *queue);
211 static inline int nvme_tcp_queue_id(struct nvme_tcp_queue *queue) argument
213 return queue - queue->ctrl->queues;
224 static inline struct blk_mq_tags *nvme_tcp_tagset(struct nvme_tcp_queue *queue) argument
226 u32 queue_idx = nvme_tcp_queue_id(queue);
229 return queue->ctrl->admin_tag_set.tags[queue_idx];
230 return queue->ctrl->tag_set.tags[queue_idx - 1];
233 static inline u8 nvme_tcp_hdgst_len(struct nvme_tcp_queue *queue) argument
238 nvme_tcp_ddgst_len(struct nvme_tcp_queue *queue) argument
353 nvme_tcp_send_all(struct nvme_tcp_queue *queue) argument
363 nvme_tcp_queue_more(struct nvme_tcp_queue *queue) argument
372 struct nvme_tcp_queue *queue = req->queue; local
393 nvme_tcp_process_req_list(struct nvme_tcp_queue *queue) argument
405 nvme_tcp_fetch_request(struct nvme_tcp_queue *queue) argument
451 nvme_tcp_verify_hdgst(struct nvme_tcp_queue *queue, void *pdu, size_t pdu_len) argument
478 nvme_tcp_check_ddgst(struct nvme_tcp_queue *queue, void *pdu) argument
514 struct nvme_tcp_queue *queue = &ctrl->queues[queue_idx]; local
535 struct nvme_tcp_queue *queue = &ctrl->queues[hctx_idx + 1]; local
545 struct nvme_tcp_queue *queue = &ctrl->queues[0]; local
552 nvme_tcp_recv_state(struct nvme_tcp_queue *queue) argument
559 nvme_tcp_init_recv_ctx(struct nvme_tcp_queue *queue) argument
577 nvme_tcp_process_nvme_cqe(struct nvme_tcp_queue *queue, struct nvme_completion *cqe) argument
603 nvme_tcp_handle_c2h_data(struct nvme_tcp_queue *queue, struct nvme_tcp_data_pdu *pdu) argument
637 nvme_tcp_handle_comp(struct nvme_tcp_queue *queue, struct nvme_tcp_rsp_pdu *pdu) argument
662 struct nvme_tcp_queue *queue = req->queue; local
693 nvme_tcp_handle_r2t(struct nvme_tcp_queue *queue, struct nvme_tcp_r2t_pdu *pdu) argument
742 nvme_tcp_recv_pdu(struct nvme_tcp_queue *queue, struct sk_buff *skb, unsigned int *offset, size_t *len) argument
800 nvme_tcp_recv_data(struct nvme_tcp_queue *queue, struct sk_buff *skb, unsigned int *offset, size_t *len) argument
871 nvme_tcp_recv_ddgst(struct nvme_tcp_queue *queue, struct sk_buff *skb, unsigned int *offset, size_t *len) argument
919 struct nvme_tcp_queue *queue = desc->arg.data; local
954 struct nvme_tcp_queue *queue; local
968 struct nvme_tcp_queue *queue; local
981 struct nvme_tcp_queue *queue; local
1007 nvme_tcp_done_send_req(struct nvme_tcp_queue *queue) argument
1027 struct nvme_tcp_queue *queue = req->queue; local
1090 struct nvme_tcp_queue *queue = req->queue; local
1131 struct nvme_tcp_queue *queue = req->queue; local
1165 struct nvme_tcp_queue *queue = req->queue; local
1196 nvme_tcp_try_send(struct nvme_tcp_queue *queue) argument
1246 nvme_tcp_try_recv(struct nvme_tcp_queue *queue) argument
1264 struct nvme_tcp_queue *queue = local
1295 nvme_tcp_free_crypto(struct nvme_tcp_queue *queue) argument
1304 nvme_tcp_alloc_crypto(struct nvme_tcp_queue *queue) argument
1339 struct nvme_tcp_queue *queue = &ctrl->queues[0]; local
1356 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; local
1378 nvme_tcp_init_connection(struct nvme_tcp_queue *queue) argument
1506 nvme_tcp_admin_queue(struct nvme_tcp_queue *queue) argument
1511 nvme_tcp_default_queue(struct nvme_tcp_queue *queue) argument
1520 nvme_tcp_read_queue(struct nvme_tcp_queue *queue) argument
1531 nvme_tcp_poll_queue(struct nvme_tcp_queue *queue) argument
1544 nvme_tcp_set_queue_io_cpu(struct nvme_tcp_queue *queue) argument
1565 struct nvme_tcp_queue *queue = data; local
1592 nvme_tcp_start_tls(struct nvme_ctrl *nctrl, struct nvme_tcp_queue *queue, key_serial_t pskid) argument
1644 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; local
1796 nvme_tcp_restore_sock_ops(struct nvme_tcp_queue *queue) argument
1808 __nvme_tcp_stop_queue(struct nvme_tcp_queue *queue) argument
1818 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; local
1829 nvme_tcp_setup_sock_ops(struct nvme_tcp_queue *queue) argument
1848 struct nvme_tcp_queue *queue = &ctrl->queues[idx]; local
2374 nvme_tcp_set_sg_inline(struct nvme_tcp_queue *queue, struct nvme_command *c, u32 data_len) argument
2398 struct nvme_tcp_queue *queue = &ctrl->queues[0]; local
2471 nvme_tcp_map_data(struct nvme_tcp_queue *queue, struct request *rq) argument
2496 struct nvme_tcp_queue *queue = req->queue; local
2547 struct nvme_tcp_queue *queue = hctx->driver_data; local
2557 struct nvme_tcp_queue *queue = hctx->driver_data; local
2586 struct nvme_tcp_queue *queue = hctx->driver_data; local
2602 struct nvme_tcp_queue *queue = &to_tcp_ctrl(ctrl)->queues[0]; local
[all...]
/linux-master/drivers/net/wireless/ralink/rt2x00/
H A Drt2x00queue.c12 Abstract: rt2x00 queue specific routines.
25 struct data_queue *queue = entry->queue; local
26 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev;
37 frame_size = queue->data_size + queue->desc_size + queue->winfo_size;
95 struct device *dev = entry->queue->rt2x00dev->dev;
112 struct device *dev = entry->queue->rt2x00dev->dev;
488 struct rt2x00_dev *rt2x00dev = entry->queue
529 struct data_queue *queue = entry->queue; local
540 rt2x00queue_kick_tx_queue(struct data_queue *queue, struct txentry_desc *txdesc) argument
599 rt2x00queue_write_tx_frame(struct data_queue *queue, struct sk_buff *skb, struct ieee80211_sta *sta, bool local) argument
784 rt2x00queue_for_each_entry(struct data_queue *queue, enum queue_index start, enum queue_index end, void *data, bool (*fn)(struct queue_entry *entry, void *data)) argument
839 rt2x00queue_get_entry(struct data_queue *queue, enum queue_index index) argument
863 struct data_queue *queue = entry->queue; local
890 rt2x00queue_pause_queue_nocheck(struct data_queue *queue) argument
907 rt2x00queue_pause_queue(struct data_queue *queue) argument
918 rt2x00queue_unpause_queue(struct data_queue *queue) argument
949 rt2x00queue_start_queue(struct data_queue *queue) argument
969 rt2x00queue_stop_queue(struct data_queue *queue) argument
986 rt2x00queue_flush_queue(struct data_queue *queue, bool drop) argument
1025 struct data_queue *queue; local
1040 struct data_queue *queue; local
1059 struct data_queue *queue; local
1068 rt2x00queue_reset(struct data_queue *queue) argument
1086 struct data_queue *queue; local
1097 rt2x00queue_alloc_entries(struct data_queue *queue) argument
1134 rt2x00queue_free_skbs(struct data_queue *queue) argument
1146 rt2x00queue_alloc_rxskbs(struct data_queue *queue) argument
1163 struct data_queue *queue; local
1202 struct data_queue *queue; local
1212 rt2x00queue_init(struct rt2x00_dev *rt2x00dev, struct data_queue *queue, enum data_queue_qid qid) argument
1233 struct data_queue *queue; local
[all...]
H A Drt2x00mmio.c51 struct data_queue *queue = rt2x00dev->rx; local
58 entry = rt2x00queue_get_entry(queue, Q_INDEX);
69 skbdesc->desc_len = entry->queue->desc_size;
88 void rt2x00mmio_flush_queue(struct data_queue *queue, bool drop) argument
92 for (i = 0; !rt2x00queue_empty(queue) && i < 10; i++)
101 struct data_queue *queue)
112 queue->limit * queue->desc_size, &dma,
118 * Initialize all queue entries to contain valid addresses.
120 for (i = 0; i < queue
100 rt2x00mmio_alloc_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument
129 rt2x00mmio_free_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument
144 struct data_queue *queue; local
180 struct data_queue *queue; local
[all...]
/linux-master/drivers/net/wireless/broadcom/b43legacy/
H A Dpio.h43 struct b43legacy_pioqueue *queue; member in struct:b43legacy_pio_txpacket
49 (packet)->queue->tx_packets_cache))
74 /* Packets on the txrunning queue are completely
84 u16 b43legacy_pio_read(struct b43legacy_pioqueue *queue, argument
87 return b43legacy_read16(queue->dev, queue->mmio_base + offset);
91 void b43legacy_pio_write(struct b43legacy_pioqueue *queue, argument
94 b43legacy_write16(queue->dev, queue->mmio_base + offset, value);
105 void b43legacy_pio_rx(struct b43legacy_pioqueue *queue);
137 b43legacy_pio_rx(struct b43legacy_pioqueue *queue) argument
141 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue) argument
145 b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue) argument
[all...]
/linux-master/drivers/net/ethernet/ibm/ehea/
H A Dehea_qmr.h196 static inline void *hw_qeit_calc(struct hw_queue *queue, u64 q_offset) argument
200 if (q_offset >= queue->queue_length)
201 q_offset -= queue->queue_length;
202 current_page = (queue->queue_pages)[q_offset >> EHEA_PAGESHIFT];
206 static inline void *hw_qeit_get(struct hw_queue *queue) argument
208 return hw_qeit_calc(queue, queue->current_q_offset);
211 static inline void hw_qeit_inc(struct hw_queue *queue) argument
213 queue->current_q_offset += queue
221 hw_qeit_get_inc(struct hw_queue *queue) argument
228 hw_qeit_get_inc_valid(struct hw_queue *queue) argument
245 hw_qeit_get_valid(struct hw_queue *queue) argument
261 hw_qeit_reset(struct hw_queue *queue) argument
267 hw_qeit_eq_get_inc(struct hw_queue *queue) argument
281 hw_eqit_eq_get_inc_valid(struct hw_queue *queue) argument
295 struct hw_queue *queue; local
310 struct hw_queue *queue = &my_qp->hw_squeue; local
327 struct hw_queue *queue = &qp->hw_rqueue1; local
[all...]
/linux-master/drivers/gpu/drm/imagination/
H A Dpvr_queue.c86 WARN(1, "Invalid queue type");
116 pvr_context_put(fence->queue->ctx);
125 switch (fence->queue->type) {
139 WARN(1, "Invalid queue type");
148 switch (fence->queue->type) {
162 WARN(1, "Invalid queue type");
254 * @queue: The queue this fence belongs to.
259 * pvr_queue_fence::queue field too.
263 struct pvr_queue *queue,
262 pvr_queue_fence_init(struct dma_fence *f, struct pvr_queue *queue, const struct dma_fence_ops *fence_ops, struct pvr_queue_fence_ctx *fence_ctx) argument
287 pvr_queue_cccb_fence_init(struct dma_fence *fence, struct pvr_queue *queue) argument
305 pvr_queue_job_fence_init(struct dma_fence *fence, struct pvr_queue *queue) argument
388 pvr_queue_get_job_cccb_fence(struct pvr_queue *queue, struct pvr_job *job) argument
441 pvr_queue_get_job_kccb_fence(struct pvr_queue *queue, struct pvr_job *job) argument
461 pvr_queue_get_paired_frag_job_dep(struct pvr_queue *queue, struct pvr_job *job) argument
499 struct pvr_queue *queue = container_of(s_entity, struct pvr_queue, entity); local
560 pvr_queue_update_active_state_locked(struct pvr_queue *queue) argument
592 pvr_queue_update_active_state(struct pvr_queue *queue) argument
603 struct pvr_queue *queue = container_of(job->base.sched, struct pvr_queue, scheduler); local
745 struct pvr_queue *queue = container_of(job->base.sched, local
756 pvr_queue_stop(struct pvr_queue *queue, struct pvr_job *bad_job) argument
761 pvr_queue_start(struct pvr_queue *queue) argument
802 struct pvr_queue *queue = container_of(sched, struct pvr_queue, scheduler); local
902 pvr_queue_signal_done_fences(struct pvr_queue *queue) argument
932 pvr_queue_check_job_waiting_for_cccb_space(struct pvr_queue *queue) argument
985 pvr_queue_process(struct pvr_queue *queue) argument
994 get_dm_type(struct pvr_queue *queue) argument
1018 init_fw_context(struct pvr_queue *queue, void *fw_ctx_map) argument
1049 pvr_queue_cleanup_fw_context(struct pvr_queue *queue) argument
1075 struct pvr_queue *queue; local
1151 struct pvr_queue *queue = container_of(job->base.sched, struct pvr_queue, scheduler); local
1163 struct pvr_queue *queue = priv; local
1214 struct pvr_queue *queue; local
1327 struct pvr_queue *queue; local
1339 struct pvr_queue *queue; local
1358 pvr_queue_kill(struct pvr_queue *queue) argument
1372 pvr_queue_destroy(struct pvr_queue *queue) argument
[all...]
/linux-master/drivers/net/ethernet/engleder/
H A Dtsnep_xdp.c24 struct tsnep_queue *queue; local
31 queue = &adapter->queue[queue_id];
32 if (queue->rx->queue_index != queue_id ||
33 queue->tx->queue_index != queue_id) {
35 "XSK support only for TX/RX queue pairs\n");
48 retval = tsnep_enable_xsk(queue, pool);
61 struct tsnep_queue *queue; local
71 queue = &adapter->queue[queue_i
[all...]
/linux-master/drivers/net/xen-netback/
H A Dnetback.c58 /* The time that packets can stay on the guest Rx internal queue
106 static void xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx,
109 static void make_tx_response(struct xenvif_queue *queue,
114 static void xenvif_idx_unmap(struct xenvif_queue *queue, u16 pending_idx);
116 static inline int tx_work_todo(struct xenvif_queue *queue);
118 static inline unsigned long idx_to_pfn(struct xenvif_queue *queue, argument
121 return page_to_pfn(queue->mmap_pages[idx]);
124 static inline unsigned long idx_to_kaddr(struct xenvif_queue *queue, argument
127 return (unsigned long)pfn_to_kaddr(idx_to_pfn(queue, idx));
160 void xenvif_kick_thread(struct xenvif_queue *queue) argument
165 xenvif_napi_schedule_or_enable_events(struct xenvif_queue *queue) argument
179 tx_add_credit(struct xenvif_queue *queue) argument
200 struct xenvif_queue *queue = from_timer(queue, t, credit_timeout); local
205 xenvif_tx_err(struct xenvif_queue *queue, struct xen_netif_tx_request *txp, unsigned int extra_count, RING_IDX end) argument
230 xenvif_count_requests(struct xenvif_queue *queue, struct xen_netif_tx_request *first, unsigned int extra_count, struct xen_netif_tx_request *txp, int work_to_do) argument
339 xenvif_tx_create_map_op(struct xenvif_queue *queue, u16 pending_idx, struct xen_netif_tx_request *txp, unsigned int extra_count, struct gnttab_map_grant_ref *mop) argument
374 xenvif_get_requests(struct xenvif_queue *queue, struct sk_buff *skb, struct xen_netif_tx_request *first, struct xen_netif_tx_request *txfrags, unsigned *copy_ops, unsigned *map_ops, unsigned int frag_overflow, struct sk_buff *nskb, unsigned int extra_count, unsigned int data_len) argument
522 xenvif_grant_handle_set(struct xenvif_queue *queue, u16 pending_idx, grant_handle_t handle) argument
536 xenvif_grant_handle_reset(struct xenvif_queue *queue, u16 pending_idx) argument
549 xenvif_tx_check_gop(struct xenvif_queue *queue, struct sk_buff *skb, struct gnttab_map_grant_ref **gopp_map, struct gnttab_copy **gopp_copy) argument
686 xenvif_fill_frags(struct xenvif_queue *queue, struct sk_buff *skb) argument
724 xenvif_get_extras(struct xenvif_queue *queue, struct xen_netif_extra_info *extras, unsigned int *extra_count, int work_to_do) argument
787 checksum_setup(struct xenvif_queue *queue, struct sk_buff *skb) argument
809 tx_credit_exceeded(struct xenvif_queue *queue, unsigned size) argument
914 xenvif_tx_build_gops(struct xenvif_queue *queue, int budget, unsigned *copy_ops, unsigned *map_ops) argument
1110 xenvif_handle_frag_list(struct xenvif_queue *queue, struct sk_buff *skb) argument
1170 xenvif_tx_submit(struct xenvif_queue *queue) argument
1287 struct xenvif_queue *queue = ubuf_to_queue(ubuf); local
1315 xenvif_tx_dealloc_action(struct xenvif_queue *queue) argument
1383 xenvif_tx_action(struct xenvif_queue *queue, int budget) argument
1418 _make_tx_response(struct xenvif_queue *queue, const struct xen_netif_tx_request *txp, unsigned int extra_count, s8 status) argument
1436 push_tx_responses(struct xenvif_queue *queue) argument
1445 xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx, s8 status) argument
1471 make_tx_response(struct xenvif_queue *queue, const struct xen_netif_tx_request *txp, unsigned int extra_count, s8 status) argument
1486 xenvif_idx_unmap(struct xenvif_queue *queue, u16 pending_idx) argument
1511 tx_work_todo(struct xenvif_queue *queue) argument
1519 tx_dealloc_work_todo(struct xenvif_queue *queue) argument
1524 xenvif_unmap_frontend_data_rings(struct xenvif_queue *queue) argument
1534 xenvif_map_frontend_data_rings(struct xenvif_queue *queue, grant_ref_t tx_ring_ref, grant_ref_t rx_ring_ref) argument
1581 xenvif_dealloc_kthread_should_stop(struct xenvif_queue *queue) argument
1592 struct xenvif_queue *queue = data; local
[all...]
H A Dinterface.c44 /* Number of bytes allowed on the internal guest Rx queue. */
52 void xenvif_skb_zerocopy_prepare(struct xenvif_queue *queue, argument
56 atomic_inc(&queue->inflight_packets);
59 void xenvif_skb_zerocopy_complete(struct xenvif_queue *queue) argument
61 atomic_dec(&queue->inflight_packets);
67 wake_up(&queue->dealloc_wq);
77 static bool xenvif_handle_tx_interrupt(struct xenvif_queue *queue) argument
81 rc = RING_HAS_UNCONSUMED_REQUESTS(&queue->tx);
83 napi_schedule(&queue->napi);
89 struct xenvif_queue *queue local
105 struct xenvif_queue *queue = local
132 xenvif_handle_rx_interrupt(struct xenvif_queue *queue) argument
144 struct xenvif_queue *queue = dev_id; local
160 struct xenvif_queue *queue = dev_id; local
208 struct xenvif_queue *queue = NULL; local
274 struct xenvif_queue *queue = NULL; local
306 struct xenvif_queue *queue = NULL; local
322 struct xenvif_queue *queue = NULL; local
561 xenvif_init_queue(struct xenvif_queue *queue) argument
671 xenvif_disconnect_queue(struct xenvif_queue *queue) argument
703 xenvif_connect_data(struct xenvif_queue *queue, unsigned long tx_ring_ref, unsigned long rx_ring_ref, unsigned int tx_evtchn, unsigned int rx_evtchn) argument
804 struct xenvif_queue *queue = NULL; local
838 xenvif_deinit_queue(struct xenvif_queue *queue) argument
[all...]
/linux-master/drivers/net/
H A Dxen-netfront.c94 /* IRQ name is queue name with "-tx" or "-rx" appended */
166 /* Multi-queue support */
218 static struct sk_buff *xennet_get_rx_skb(struct netfront_queue *queue, argument
222 struct sk_buff *skb = queue->rx_skbs[i];
223 queue->rx_skbs[i] = NULL;
227 static grant_ref_t xennet_get_rx_ref(struct netfront_queue *queue, argument
231 grant_ref_t ref = queue->grant_rx_ref[i];
232 queue->grant_rx_ref[i] = INVALID_GRANT_REF;
248 struct netfront_queue *queue = from_timer(queue, local
252 netfront_tx_slot_available(struct netfront_queue *queue) argument
258 xennet_maybe_wake_tx(struct netfront_queue *queue) argument
270 xennet_alloc_one_rx_buffer(struct netfront_queue *queue) argument
298 xennet_alloc_rx_buffers(struct netfront_queue *queue) argument
365 struct netfront_queue *queue = NULL; local
389 xennet_tx_buf_gc(struct netfront_queue *queue) argument
465 struct netfront_queue *queue; member in struct:xennet_gnttab_make_txreq
482 struct netfront_queue *queue = info->queue; local
603 xennet_mark_tx_pending(struct netfront_queue *queue) argument
612 xennet_xdp_xmit_one(struct net_device *dev, struct netfront_queue *queue, struct xdp_frame *xdpf) argument
650 struct netfront_queue *queue = NULL; local
719 struct netfront_queue *queue = NULL; local
872 struct netfront_queue *queue; local
886 struct netfront_queue *queue = &info->queues[i]; local
903 xennet_set_rx_rsp_cons(struct netfront_queue *queue, RING_IDX val) argument
913 xennet_move_rx_slot(struct netfront_queue *queue, struct sk_buff *skb, grant_ref_t ref) argument
926 xennet_get_extras(struct netfront_queue *queue, struct xen_netif_extra_info *extras, RING_IDX rp) argument
968 xennet_run_xdp(struct netfront_queue *queue, struct page *pdata, struct xen_netif_rx_response *rx, struct bpf_prog *prog, struct xdp_buff *xdp, bool *need_xdp_flush) argument
1015 xennet_get_responses(struct netfront_queue *queue, struct netfront_rx_info *rinfo, RING_IDX rp, struct sk_buff_head *list, bool *need_xdp_flush) argument
1158 xennet_fill_frags(struct netfront_queue *queue, struct sk_buff *skb, struct sk_buff_head *list) argument
1221 handle_incoming_queue(struct netfront_queue *queue, struct sk_buff_head *rxq) argument
1259 struct netfront_queue *queue = container_of(napi, struct netfront_queue, napi); local
1417 xennet_release_tx_bufs(struct netfront_queue *queue) argument
1439 xennet_release_rx_bufs(struct netfront_queue *queue) argument
1508 xennet_handle_tx(struct netfront_queue *queue, unsigned int *eoi) argument
1533 xennet_handle_rx(struct netfront_queue *queue, unsigned int *eoi) argument
1817 struct netfront_queue *queue = &info->queues[i]; local
1899 setup_netfront_single(struct netfront_queue *queue) argument
1925 setup_netfront_split(struct netfront_queue *queue) argument
1969 setup_netfront(struct xenbus_device *dev, struct netfront_queue *queue, unsigned int feature_split_evtchn) argument
2020 xennet_init_queue(struct netfront_queue *queue) argument
2076 write_queue_xenstore_keys(struct netfront_queue *queue, struct xenbus_transaction *xbt, int write_hierarchical) argument
2160 xennet_create_page_pool(struct netfront_queue *queue) argument
2215 struct netfront_queue *queue = &info->queues[i]; local
2260 struct netfront_queue *queue = NULL; local
2424 struct netfront_queue *queue = NULL; local
[all...]
/linux-master/block/
H A Dblk-mq-cpumap.c3 * CPU <-> hardware queue mapping helpers
21 unsigned int queue, cpu; local
30 for (queue = 0; queue < qmap->nr_queues; queue++) {
31 for_each_cpu(cpu, &masks[queue])
32 qmap->mq_map[cpu] = qmap->queue_offset + queue;
39 * blk_mq_hw_queue_to_node - Look up the memory node for a hardware queue index
40 * @qmap: CPU to hardware queue map.
41 * @index: hardware queue inde
[all...]
/linux-master/drivers/md/dm-vdo/
H A Dfunnel-queue.h13 * A funnel queue is a simple (almost) lock-free queue that accepts entries from multiple threads
19 * mechanism to ensure that only one thread is consuming from the queue. If more than one thread
20 * attempts to consume from the queue, the resulting behavior is undefined. Clients must not
21 * directly access or manipulate the internals of the queue, which are only exposed for the purpose
25 * the queue entries, and pointers to those structures are used exclusively by the queue. No macros
26 * are defined to template the queue, so the offset of the funnel_queue_entry in the records placed
27 * in the queue must all be the same so the client can derive their structure pointer from the
31 * soon as they are returned since this queue i
82 vdo_funnel_queue_put(struct funnel_queue *queue, struct funnel_queue_entry *entry) argument
[all...]
H A Dfunnel-workqueue.h33 void vdo_enqueue_work_queue(struct vdo_work_queue *queue, struct vdo_completion *completion);
35 void vdo_finish_work_queue(struct vdo_work_queue *queue);
37 void vdo_free_work_queue(struct vdo_work_queue *queue);
39 void vdo_dump_work_queue(struct vdo_work_queue *queue);
46 struct vdo_thread *vdo_get_work_queue_owner(struct vdo_work_queue *queue);
48 bool __must_check vdo_work_queue_type_is(struct vdo_work_queue *queue,
/linux-master/drivers/net/wireless/ath/ath5k/
H A Dqcu.c35 * basically we have 10 queues to play with. Each queue has a matching
36 * QCU that controls when the queue will get triggered and multiple QCUs
39 * and DCUs allowing us to have different DFS settings for each queue.
41 * When a frame goes into a TX queue, QCU decides when it'll trigger a
43 * it's buffer or -if it's a beacon queue- if it's time to fire up the queue
58 * ath5k_hw_num_tx_pending() - Get number of pending frames for a given queue
60 * @queue: One of enum ath5k_tx_queue_id
63 ath5k_hw_num_tx_pending(struct ath5k_hw *ah, unsigned int queue) argument
66 AR5K_ASSERT_ENTRY(queue, a
94 ath5k_hw_release_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument
138 ath5k_hw_get_tx_queueprops(struct ath5k_hw *ah, int queue, struct ath5k_txq_info *queue_info) argument
154 ath5k_hw_set_tx_queueprops(struct ath5k_hw *ah, int queue, const struct ath5k_txq_info *qinfo) argument
206 unsigned int queue; local
281 ath5k_hw_set_tx_retry_limits(struct ath5k_hw *ah, unsigned int queue) argument
324 ath5k_hw_reset_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument
[all...]
/linux-master/drivers/soc/ixp4xx/
H A Dixp4xx-qmgr.c29 void qmgr_put_entry(unsigned int queue, u32 val) argument
32 BUG_ON(!qmgr_queue_descs[queue]); /* not yet requested */
35 qmgr_queue_descs[queue], queue, val);
37 __raw_writel(val, &qmgr_regs->acc[queue][0]);
40 u32 qmgr_get_entry(unsigned int queue) argument
43 val = __raw_readl(&qmgr_regs->acc[queue][0]);
45 BUG_ON(!qmgr_queue_descs[queue]); /* not yet requested */
48 qmgr_queue_descs[queue], queue, va
53 __qmgr_get_stat1(unsigned int queue) argument
59 __qmgr_get_stat2(unsigned int queue) argument
72 qmgr_stat_empty(unsigned int queue) argument
84 qmgr_stat_below_low_watermark(unsigned int queue) argument
98 qmgr_stat_full(unsigned int queue) argument
112 qmgr_stat_overflow(unsigned int queue) argument
117 qmgr_set_irq(unsigned int queue, int src, void (*handler)(void *pdev), void *pdev) argument
205 qmgr_enable_irq(unsigned int queue) argument
217 qmgr_disable_irq(unsigned int queue) argument
239 qmgr_request_queue(unsigned int queue, unsigned int len , unsigned int nearly_empty_watermark, unsigned int nearly_full_watermark, const char *desc_format, const char* name) argument
329 qmgr_release_queue(unsigned int queue) argument
[all...]
/linux-master/net/sunrpc/
H A Dsched.c91 * queue->lock and bh_disabled in order to avoid races within
95 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) argument
101 if (list_empty(&queue->timer_list.list))
102 cancel_delayed_work(&queue->timer_list.dwork);
106 rpc_set_queue_timer(struct rpc_wait_queue *queue, unsigned long expires) argument
109 queue->timer_list.expires = expires;
114 mod_delayed_work(rpciod_workqueue, &queue->timer_list.dwork, expires);
121 __rpc_add_timer(struct rpc_wait_queue *queue, struct rpc_task *task, argument
125 if (list_empty(&queue->timer_list.list) || time_before(timeout, queue
130 rpc_set_waitqueue_priority(struct rpc_wait_queue *queue, int priority) argument
138 rpc_reset_waitqueue_priority(struct rpc_wait_queue *queue) argument
193 __rpc_add_wait_queue_priority(struct rpc_wait_queue *queue, struct rpc_task *task, unsigned char queue_priority) argument
205 __rpc_add_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task, unsigned char queue_priority) argument
233 __rpc_remove_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task) argument
243 __rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname, unsigned char nr_queues) argument
259 rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname) argument
265 rpc_init_wait_queue(struct rpc_wait_queue *queue, const char *qname) argument
271 rpc_destroy_wait_queue(struct rpc_wait_queue *queue) argument
506 __rpc_do_wake_up_task_on_wq(struct workqueue_struct *wq, struct rpc_wait_queue *queue, struct rpc_task *task) argument
527 rpc_wake_up_task_on_wq_queue_action_locked(struct workqueue_struct *wq, struct rpc_wait_queue *queue, struct rpc_task *task, bool (*action)(struct rpc_task *, void *), void *data) argument
546 rpc_wake_up_task_queue_locked(struct rpc_wait_queue *queue, struct rpc_task *task) argument
556 rpc_wake_up_queued_task(struct rpc_wait_queue *queue, struct rpc_task *task) argument
573 rpc_wake_up_task_queue_set_status_locked(struct rpc_wait_queue *queue, struct rpc_task *task, int status) argument
590 rpc_wake_up_queued_task_set_status(struct rpc_wait_queue *queue, struct rpc_task *task, int status) argument
603 __rpc_find_next_queued_priority(struct rpc_wait_queue *queue) argument
650 __rpc_find_next_queued(struct rpc_wait_queue *queue) argument
662 rpc_wake_up_first_on_wq(struct workqueue_struct *wq, struct rpc_wait_queue *queue, bool (*func)(struct rpc_task *, void *), void *data) argument
681 rpc_wake_up_first(struct rpc_wait_queue *queue, bool (*func)(struct rpc_task *, void *), void *data) argument
696 rpc_wake_up_next(struct rpc_wait_queue *queue) argument
707 rpc_wake_up_locked(struct rpc_wait_queue *queue) argument
725 rpc_wake_up(struct rpc_wait_queue *queue) argument
738 rpc_wake_up_status_locked(struct rpc_wait_queue *queue, int status) argument
757 rpc_wake_up_status(struct rpc_wait_queue *queue, int status) argument
767 struct rpc_wait_queue *queue = container_of(work, local
857 struct rpc_wait_queue *queue; local
874 struct rpc_wait_queue *queue; local
911 struct rpc_wait_queue *queue; local
[all...]
/linux-master/arch/powerpc/boot/dts/fsl/
H A Dpq3-etsec2-grp2-0.dtsi36 queue-group@b4000 {
H A Dpq3-etsec2-grp2-1.dtsi36 queue-group@b5000 {
H A Dpq3-etsec2-grp2-2.dtsi36 queue-group@b6000 {
/linux-master/drivers/staging/rtl8723bs/include/
H A Dcmd_osdep.h15 int _rtw_enqueue_cmd(struct __queue *queue, struct cmd_obj *obj);
16 extern struct cmd_obj *_rtw_dequeue_cmd(struct __queue *queue);
/linux-master/drivers/net/wireless/st/cw1200/
H A Dqueue.h3 * O(1) TX queue with built-in allocator for ST-Ericsson CW1200 drivers
33 struct list_head queue; member in struct:cw1200_queue
39 spinlock_t lock; /* Protect queue entry */
68 int cw1200_queue_init(struct cw1200_queue *queue,
73 int cw1200_queue_clear(struct cw1200_queue *queue);
75 void cw1200_queue_deinit(struct cw1200_queue *queue);
77 size_t cw1200_queue_get_num_queued(struct cw1200_queue *queue,
79 int cw1200_queue_put(struct cw1200_queue *queue,
82 int cw1200_queue_get(struct cw1200_queue *queue,
87 int cw1200_queue_requeue(struct cw1200_queue *queue, u3
[all...]
/linux-master/drivers/scsi/arm/
H A DMakefile8 obj-$(CONFIG_SCSI_ACORNSCSI_3) += acornscsi_mod.o queue.o msgqueue.o
9 obj-$(CONFIG_SCSI_ARXESCSI) += arxescsi.o fas216.o queue.o msgqueue.o
11 obj-$(CONFIG_SCSI_CUMANA_2) += cumana_2.o fas216.o queue.o msgqueue.o
13 obj-$(CONFIG_SCSI_POWERTECSCSI) += powertec.o fas216.o queue.o msgqueue.o
14 obj-$(CONFIG_SCSI_EESOXSCSI) += eesox.o fas216.o queue.o msgqueue.o
/linux-master/include/net/
H A Dnetdev_rx_queue.h10 /* This structure contains an instance of an RX queue. */
24 /* NAPI instance for the queue
31 * RX queue sysfs structures and functions.
35 ssize_t (*show)(struct netdev_rx_queue *queue, char *buf);
36 ssize_t (*store)(struct netdev_rx_queue *queue,
48 get_netdev_rx_queue_index(struct netdev_rx_queue *queue) argument
50 struct net_device *dev = queue->dev;
51 int index = queue - dev->_rx;
/linux-master/drivers/gpu/drm/msm/
H A Dmsm_submitqueue.c71 struct msm_gpu_submitqueue *queue = container_of(kref, local
74 idr_destroy(&queue->fence_idr);
76 msm_file_private_put(queue->ctx);
78 kfree(queue);
162 struct msm_gpu_submitqueue *queue; local
177 queue = kzalloc(sizeof(*queue), GFP_KERNEL);
179 if (!queue)
182 kref_init(&queue->ref);
183 queue
237 msm_submitqueue_query_faults(struct msm_gpu_submitqueue *queue, struct drm_msm_submitqueue_query *args) argument
260 struct msm_gpu_submitqueue *queue; local
[all...]
/linux-master/drivers/crypto/hisilicon/sec/
H A Dsec_drv.c227 static int sec_queue_map_io(struct sec_queue *queue) argument
229 struct device *dev = queue->dev_info->dev;
234 2 + queue->queue_id);
236 dev_err(dev, "Failed to get queue %u memory resource\n",
237 queue->queue_id);
240 queue->regs = ioremap(res->start, resource_size(res));
241 if (!queue->regs)
247 static void sec_queue_unmap_io(struct sec_queue *queue) argument
249 iounmap(queue->regs);
252 static int sec_queue_ar_pkgattr(struct sec_queue *queue, u3 argument
267 sec_queue_aw_pkgattr(struct sec_queue *queue, u32 aw_pkg) argument
519 sec_queue_ar_alloc(struct sec_queue *queue, u32 alloc) argument
536 sec_queue_aw_alloc(struct sec_queue *queue, u32 alloc) argument
553 sec_queue_reorder(struct sec_queue *queue, bool reorder) argument
566 sec_queue_depth(struct sec_queue *queue, u32 depth) argument
578 sec_queue_cmdbase_addr(struct sec_queue *queue, u64 addr) argument
584 sec_queue_outorder_addr(struct sec_queue *queue, u64 addr) argument
592 sec_queue_errbase_addr(struct sec_queue *queue, u64 addr) argument
600 sec_queue_irq_disable(struct sec_queue *queue) argument
605 sec_queue_irq_enable(struct sec_queue *queue) argument
610 sec_queue_abn_irq_disable(struct sec_queue *queue) argument
615 sec_queue_stop(struct sec_queue *queue) argument
622 sec_queue_start(struct sec_queue *queue) argument
651 sec_queue_free(struct sec_queue *queue) argument
681 struct sec_queue *queue = q; local
728 sec_queue_irq_init(struct sec_queue *queue) argument
745 sec_queue_irq_uninit(struct sec_queue *queue) argument
774 struct sec_queue *queue; local
799 struct sec_queue *queue = ERR_PTR(-ENODEV); local
821 sec_queue_stop_release(struct sec_queue *queue) argument
843 sec_queue_empty(struct sec_queue *queue) argument
858 sec_queue_send(struct sec_queue *queue, struct sec_bd_info *msg, void *ctx) argument
885 sec_queue_can_enqueue(struct sec_queue *queue, int num) argument
892 sec_queue_hw_init(struct sec_queue *queue) argument
1001 sec_queue_base_init(struct sec_dev_info *info, struct sec_queue *queue, int queue_id) argument
1078 sec_queue_res_cfg(struct sec_queue *queue) argument
1130 sec_queue_free_ring_pages(struct sec_queue *queue) argument
1142 sec_queue_config(struct sec_dev_info *info, struct sec_queue *queue, int queue_id) argument
1165 sec_queue_unconfig(struct sec_dev_info *info, struct sec_queue *queue) argument
[all...]

Completed in 275 milliseconds

1234567891011>>