/linux-master/drivers/nvme/host/ |
H A D | tcp.c | 105 struct nvme_tcp_queue *queue; member in struct:nvme_tcp_request 204 static int nvme_tcp_try_send(struct nvme_tcp_queue *queue); 211 static inline int nvme_tcp_queue_id(struct nvme_tcp_queue *queue) argument 213 return queue - queue->ctrl->queues; 224 static inline struct blk_mq_tags *nvme_tcp_tagset(struct nvme_tcp_queue *queue) argument 226 u32 queue_idx = nvme_tcp_queue_id(queue); 229 return queue->ctrl->admin_tag_set.tags[queue_idx]; 230 return queue->ctrl->tag_set.tags[queue_idx - 1]; 233 static inline u8 nvme_tcp_hdgst_len(struct nvme_tcp_queue *queue) argument 238 nvme_tcp_ddgst_len(struct nvme_tcp_queue *queue) argument 353 nvme_tcp_send_all(struct nvme_tcp_queue *queue) argument 363 nvme_tcp_queue_more(struct nvme_tcp_queue *queue) argument 372 struct nvme_tcp_queue *queue = req->queue; local 393 nvme_tcp_process_req_list(struct nvme_tcp_queue *queue) argument 405 nvme_tcp_fetch_request(struct nvme_tcp_queue *queue) argument 451 nvme_tcp_verify_hdgst(struct nvme_tcp_queue *queue, void *pdu, size_t pdu_len) argument 478 nvme_tcp_check_ddgst(struct nvme_tcp_queue *queue, void *pdu) argument 514 struct nvme_tcp_queue *queue = &ctrl->queues[queue_idx]; local 535 struct nvme_tcp_queue *queue = &ctrl->queues[hctx_idx + 1]; local 545 struct nvme_tcp_queue *queue = &ctrl->queues[0]; local 552 nvme_tcp_recv_state(struct nvme_tcp_queue *queue) argument 559 nvme_tcp_init_recv_ctx(struct nvme_tcp_queue *queue) argument 577 nvme_tcp_process_nvme_cqe(struct nvme_tcp_queue *queue, struct nvme_completion *cqe) argument 603 nvme_tcp_handle_c2h_data(struct nvme_tcp_queue *queue, struct nvme_tcp_data_pdu *pdu) argument 637 nvme_tcp_handle_comp(struct nvme_tcp_queue *queue, struct nvme_tcp_rsp_pdu *pdu) argument 662 struct nvme_tcp_queue *queue = req->queue; local 693 nvme_tcp_handle_r2t(struct nvme_tcp_queue *queue, struct nvme_tcp_r2t_pdu *pdu) argument 742 nvme_tcp_recv_pdu(struct nvme_tcp_queue *queue, struct sk_buff *skb, unsigned int *offset, size_t *len) argument 800 nvme_tcp_recv_data(struct nvme_tcp_queue *queue, struct sk_buff *skb, unsigned int *offset, size_t *len) argument 871 nvme_tcp_recv_ddgst(struct nvme_tcp_queue *queue, struct sk_buff *skb, unsigned int *offset, size_t *len) argument 919 struct nvme_tcp_queue *queue = desc->arg.data; local 954 struct nvme_tcp_queue *queue; local 968 struct nvme_tcp_queue *queue; local 981 struct nvme_tcp_queue *queue; local 1007 nvme_tcp_done_send_req(struct nvme_tcp_queue *queue) argument 1027 struct nvme_tcp_queue *queue = req->queue; local 1090 struct nvme_tcp_queue *queue = req->queue; local 1131 struct nvme_tcp_queue *queue = req->queue; local 1165 struct nvme_tcp_queue *queue = req->queue; local 1196 nvme_tcp_try_send(struct nvme_tcp_queue *queue) argument 1246 nvme_tcp_try_recv(struct nvme_tcp_queue *queue) argument 1264 struct nvme_tcp_queue *queue = local 1295 nvme_tcp_free_crypto(struct nvme_tcp_queue *queue) argument 1304 nvme_tcp_alloc_crypto(struct nvme_tcp_queue *queue) argument 1339 struct nvme_tcp_queue *queue = &ctrl->queues[0]; local 1356 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; local 1378 nvme_tcp_init_connection(struct nvme_tcp_queue *queue) argument 1506 nvme_tcp_admin_queue(struct nvme_tcp_queue *queue) argument 1511 nvme_tcp_default_queue(struct nvme_tcp_queue *queue) argument 1520 nvme_tcp_read_queue(struct nvme_tcp_queue *queue) argument 1531 nvme_tcp_poll_queue(struct nvme_tcp_queue *queue) argument 1544 nvme_tcp_set_queue_io_cpu(struct nvme_tcp_queue *queue) argument 1565 struct nvme_tcp_queue *queue = data; local 1592 nvme_tcp_start_tls(struct nvme_ctrl *nctrl, struct nvme_tcp_queue *queue, key_serial_t pskid) argument 1644 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; local 1796 nvme_tcp_restore_sock_ops(struct nvme_tcp_queue *queue) argument 1808 __nvme_tcp_stop_queue(struct nvme_tcp_queue *queue) argument 1818 struct nvme_tcp_queue *queue = &ctrl->queues[qid]; local 1829 nvme_tcp_setup_sock_ops(struct nvme_tcp_queue *queue) argument 1848 struct nvme_tcp_queue *queue = &ctrl->queues[idx]; local 2374 nvme_tcp_set_sg_inline(struct nvme_tcp_queue *queue, struct nvme_command *c, u32 data_len) argument 2398 struct nvme_tcp_queue *queue = &ctrl->queues[0]; local 2471 nvme_tcp_map_data(struct nvme_tcp_queue *queue, struct request *rq) argument 2496 struct nvme_tcp_queue *queue = req->queue; local 2547 struct nvme_tcp_queue *queue = hctx->driver_data; local 2557 struct nvme_tcp_queue *queue = hctx->driver_data; local 2586 struct nvme_tcp_queue *queue = hctx->driver_data; local 2602 struct nvme_tcp_queue *queue = &to_tcp_ctrl(ctrl)->queues[0]; local [all...] |
/linux-master/drivers/net/wireless/ralink/rt2x00/ |
H A D | rt2x00queue.c | 12 Abstract: rt2x00 queue specific routines. 25 struct data_queue *queue = entry->queue; local 26 struct rt2x00_dev *rt2x00dev = queue->rt2x00dev; 37 frame_size = queue->data_size + queue->desc_size + queue->winfo_size; 95 struct device *dev = entry->queue->rt2x00dev->dev; 112 struct device *dev = entry->queue->rt2x00dev->dev; 488 struct rt2x00_dev *rt2x00dev = entry->queue 529 struct data_queue *queue = entry->queue; local 540 rt2x00queue_kick_tx_queue(struct data_queue *queue, struct txentry_desc *txdesc) argument 599 rt2x00queue_write_tx_frame(struct data_queue *queue, struct sk_buff *skb, struct ieee80211_sta *sta, bool local) argument 784 rt2x00queue_for_each_entry(struct data_queue *queue, enum queue_index start, enum queue_index end, void *data, bool (*fn)(struct queue_entry *entry, void *data)) argument 839 rt2x00queue_get_entry(struct data_queue *queue, enum queue_index index) argument 863 struct data_queue *queue = entry->queue; local 890 rt2x00queue_pause_queue_nocheck(struct data_queue *queue) argument 907 rt2x00queue_pause_queue(struct data_queue *queue) argument 918 rt2x00queue_unpause_queue(struct data_queue *queue) argument 949 rt2x00queue_start_queue(struct data_queue *queue) argument 969 rt2x00queue_stop_queue(struct data_queue *queue) argument 986 rt2x00queue_flush_queue(struct data_queue *queue, bool drop) argument 1025 struct data_queue *queue; local 1040 struct data_queue *queue; local 1059 struct data_queue *queue; local 1068 rt2x00queue_reset(struct data_queue *queue) argument 1086 struct data_queue *queue; local 1097 rt2x00queue_alloc_entries(struct data_queue *queue) argument 1134 rt2x00queue_free_skbs(struct data_queue *queue) argument 1146 rt2x00queue_alloc_rxskbs(struct data_queue *queue) argument 1163 struct data_queue *queue; local 1202 struct data_queue *queue; local 1212 rt2x00queue_init(struct rt2x00_dev *rt2x00dev, struct data_queue *queue, enum data_queue_qid qid) argument 1233 struct data_queue *queue; local [all...] |
H A D | rt2x00mmio.c | 51 struct data_queue *queue = rt2x00dev->rx; local 58 entry = rt2x00queue_get_entry(queue, Q_INDEX); 69 skbdesc->desc_len = entry->queue->desc_size; 88 void rt2x00mmio_flush_queue(struct data_queue *queue, bool drop) argument 92 for (i = 0; !rt2x00queue_empty(queue) && i < 10; i++) 101 struct data_queue *queue) 112 queue->limit * queue->desc_size, &dma, 118 * Initialize all queue entries to contain valid addresses. 120 for (i = 0; i < queue 100 rt2x00mmio_alloc_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument 129 rt2x00mmio_free_queue_dma(struct rt2x00_dev *rt2x00dev, struct data_queue *queue) argument 144 struct data_queue *queue; local 180 struct data_queue *queue; local [all...] |
/linux-master/drivers/net/wireless/broadcom/b43legacy/ |
H A D | pio.h | 43 struct b43legacy_pioqueue *queue; member in struct:b43legacy_pio_txpacket 49 (packet)->queue->tx_packets_cache)) 74 /* Packets on the txrunning queue are completely 84 u16 b43legacy_pio_read(struct b43legacy_pioqueue *queue, argument 87 return b43legacy_read16(queue->dev, queue->mmio_base + offset); 91 void b43legacy_pio_write(struct b43legacy_pioqueue *queue, argument 94 b43legacy_write16(queue->dev, queue->mmio_base + offset, value); 105 void b43legacy_pio_rx(struct b43legacy_pioqueue *queue); 137 b43legacy_pio_rx(struct b43legacy_pioqueue *queue) argument 141 b43legacy_pio_tx_suspend(struct b43legacy_pioqueue *queue) argument 145 b43legacy_pio_tx_resume(struct b43legacy_pioqueue *queue) argument [all...] |
/linux-master/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.h | 196 static inline void *hw_qeit_calc(struct hw_queue *queue, u64 q_offset) argument 200 if (q_offset >= queue->queue_length) 201 q_offset -= queue->queue_length; 202 current_page = (queue->queue_pages)[q_offset >> EHEA_PAGESHIFT]; 206 static inline void *hw_qeit_get(struct hw_queue *queue) argument 208 return hw_qeit_calc(queue, queue->current_q_offset); 211 static inline void hw_qeit_inc(struct hw_queue *queue) argument 213 queue->current_q_offset += queue 221 hw_qeit_get_inc(struct hw_queue *queue) argument 228 hw_qeit_get_inc_valid(struct hw_queue *queue) argument 245 hw_qeit_get_valid(struct hw_queue *queue) argument 261 hw_qeit_reset(struct hw_queue *queue) argument 267 hw_qeit_eq_get_inc(struct hw_queue *queue) argument 281 hw_eqit_eq_get_inc_valid(struct hw_queue *queue) argument 295 struct hw_queue *queue; local 310 struct hw_queue *queue = &my_qp->hw_squeue; local 327 struct hw_queue *queue = &qp->hw_rqueue1; local [all...] |
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_queue.c | 86 WARN(1, "Invalid queue type"); 116 pvr_context_put(fence->queue->ctx); 125 switch (fence->queue->type) { 139 WARN(1, "Invalid queue type"); 148 switch (fence->queue->type) { 162 WARN(1, "Invalid queue type"); 254 * @queue: The queue this fence belongs to. 259 * pvr_queue_fence::queue field too. 263 struct pvr_queue *queue, 262 pvr_queue_fence_init(struct dma_fence *f, struct pvr_queue *queue, const struct dma_fence_ops *fence_ops, struct pvr_queue_fence_ctx *fence_ctx) argument 287 pvr_queue_cccb_fence_init(struct dma_fence *fence, struct pvr_queue *queue) argument 305 pvr_queue_job_fence_init(struct dma_fence *fence, struct pvr_queue *queue) argument 388 pvr_queue_get_job_cccb_fence(struct pvr_queue *queue, struct pvr_job *job) argument 441 pvr_queue_get_job_kccb_fence(struct pvr_queue *queue, struct pvr_job *job) argument 461 pvr_queue_get_paired_frag_job_dep(struct pvr_queue *queue, struct pvr_job *job) argument 499 struct pvr_queue *queue = container_of(s_entity, struct pvr_queue, entity); local 560 pvr_queue_update_active_state_locked(struct pvr_queue *queue) argument 592 pvr_queue_update_active_state(struct pvr_queue *queue) argument 603 struct pvr_queue *queue = container_of(job->base.sched, struct pvr_queue, scheduler); local 745 struct pvr_queue *queue = container_of(job->base.sched, local 756 pvr_queue_stop(struct pvr_queue *queue, struct pvr_job *bad_job) argument 761 pvr_queue_start(struct pvr_queue *queue) argument 802 struct pvr_queue *queue = container_of(sched, struct pvr_queue, scheduler); local 902 pvr_queue_signal_done_fences(struct pvr_queue *queue) argument 932 pvr_queue_check_job_waiting_for_cccb_space(struct pvr_queue *queue) argument 985 pvr_queue_process(struct pvr_queue *queue) argument 994 get_dm_type(struct pvr_queue *queue) argument 1018 init_fw_context(struct pvr_queue *queue, void *fw_ctx_map) argument 1049 pvr_queue_cleanup_fw_context(struct pvr_queue *queue) argument 1075 struct pvr_queue *queue; local 1151 struct pvr_queue *queue = container_of(job->base.sched, struct pvr_queue, scheduler); local 1163 struct pvr_queue *queue = priv; local 1214 struct pvr_queue *queue; local 1327 struct pvr_queue *queue; local 1339 struct pvr_queue *queue; local 1358 pvr_queue_kill(struct pvr_queue *queue) argument 1372 pvr_queue_destroy(struct pvr_queue *queue) argument [all...] |
/linux-master/drivers/net/ethernet/engleder/ |
H A D | tsnep_xdp.c | 24 struct tsnep_queue *queue; local 31 queue = &adapter->queue[queue_id]; 32 if (queue->rx->queue_index != queue_id || 33 queue->tx->queue_index != queue_id) { 35 "XSK support only for TX/RX queue pairs\n"); 48 retval = tsnep_enable_xsk(queue, pool); 61 struct tsnep_queue *queue; local 71 queue = &adapter->queue[queue_i [all...] |
/linux-master/drivers/net/xen-netback/ |
H A D | netback.c | 58 /* The time that packets can stay on the guest Rx internal queue 106 static void xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx, 109 static void make_tx_response(struct xenvif_queue *queue, 114 static void xenvif_idx_unmap(struct xenvif_queue *queue, u16 pending_idx); 116 static inline int tx_work_todo(struct xenvif_queue *queue); 118 static inline unsigned long idx_to_pfn(struct xenvif_queue *queue, argument 121 return page_to_pfn(queue->mmap_pages[idx]); 124 static inline unsigned long idx_to_kaddr(struct xenvif_queue *queue, argument 127 return (unsigned long)pfn_to_kaddr(idx_to_pfn(queue, idx)); 160 void xenvif_kick_thread(struct xenvif_queue *queue) argument 165 xenvif_napi_schedule_or_enable_events(struct xenvif_queue *queue) argument 179 tx_add_credit(struct xenvif_queue *queue) argument 200 struct xenvif_queue *queue = from_timer(queue, t, credit_timeout); local 205 xenvif_tx_err(struct xenvif_queue *queue, struct xen_netif_tx_request *txp, unsigned int extra_count, RING_IDX end) argument 230 xenvif_count_requests(struct xenvif_queue *queue, struct xen_netif_tx_request *first, unsigned int extra_count, struct xen_netif_tx_request *txp, int work_to_do) argument 339 xenvif_tx_create_map_op(struct xenvif_queue *queue, u16 pending_idx, struct xen_netif_tx_request *txp, unsigned int extra_count, struct gnttab_map_grant_ref *mop) argument 374 xenvif_get_requests(struct xenvif_queue *queue, struct sk_buff *skb, struct xen_netif_tx_request *first, struct xen_netif_tx_request *txfrags, unsigned *copy_ops, unsigned *map_ops, unsigned int frag_overflow, struct sk_buff *nskb, unsigned int extra_count, unsigned int data_len) argument 522 xenvif_grant_handle_set(struct xenvif_queue *queue, u16 pending_idx, grant_handle_t handle) argument 536 xenvif_grant_handle_reset(struct xenvif_queue *queue, u16 pending_idx) argument 549 xenvif_tx_check_gop(struct xenvif_queue *queue, struct sk_buff *skb, struct gnttab_map_grant_ref **gopp_map, struct gnttab_copy **gopp_copy) argument 686 xenvif_fill_frags(struct xenvif_queue *queue, struct sk_buff *skb) argument 724 xenvif_get_extras(struct xenvif_queue *queue, struct xen_netif_extra_info *extras, unsigned int *extra_count, int work_to_do) argument 787 checksum_setup(struct xenvif_queue *queue, struct sk_buff *skb) argument 809 tx_credit_exceeded(struct xenvif_queue *queue, unsigned size) argument 914 xenvif_tx_build_gops(struct xenvif_queue *queue, int budget, unsigned *copy_ops, unsigned *map_ops) argument 1110 xenvif_handle_frag_list(struct xenvif_queue *queue, struct sk_buff *skb) argument 1170 xenvif_tx_submit(struct xenvif_queue *queue) argument 1287 struct xenvif_queue *queue = ubuf_to_queue(ubuf); local 1315 xenvif_tx_dealloc_action(struct xenvif_queue *queue) argument 1383 xenvif_tx_action(struct xenvif_queue *queue, int budget) argument 1418 _make_tx_response(struct xenvif_queue *queue, const struct xen_netif_tx_request *txp, unsigned int extra_count, s8 status) argument 1436 push_tx_responses(struct xenvif_queue *queue) argument 1445 xenvif_idx_release(struct xenvif_queue *queue, u16 pending_idx, s8 status) argument 1471 make_tx_response(struct xenvif_queue *queue, const struct xen_netif_tx_request *txp, unsigned int extra_count, s8 status) argument 1486 xenvif_idx_unmap(struct xenvif_queue *queue, u16 pending_idx) argument 1511 tx_work_todo(struct xenvif_queue *queue) argument 1519 tx_dealloc_work_todo(struct xenvif_queue *queue) argument 1524 xenvif_unmap_frontend_data_rings(struct xenvif_queue *queue) argument 1534 xenvif_map_frontend_data_rings(struct xenvif_queue *queue, grant_ref_t tx_ring_ref, grant_ref_t rx_ring_ref) argument 1581 xenvif_dealloc_kthread_should_stop(struct xenvif_queue *queue) argument 1592 struct xenvif_queue *queue = data; local [all...] |
H A D | interface.c | 44 /* Number of bytes allowed on the internal guest Rx queue. */ 52 void xenvif_skb_zerocopy_prepare(struct xenvif_queue *queue, argument 56 atomic_inc(&queue->inflight_packets); 59 void xenvif_skb_zerocopy_complete(struct xenvif_queue *queue) argument 61 atomic_dec(&queue->inflight_packets); 67 wake_up(&queue->dealloc_wq); 77 static bool xenvif_handle_tx_interrupt(struct xenvif_queue *queue) argument 81 rc = RING_HAS_UNCONSUMED_REQUESTS(&queue->tx); 83 napi_schedule(&queue->napi); 89 struct xenvif_queue *queue local 105 struct xenvif_queue *queue = local 132 xenvif_handle_rx_interrupt(struct xenvif_queue *queue) argument 144 struct xenvif_queue *queue = dev_id; local 160 struct xenvif_queue *queue = dev_id; local 208 struct xenvif_queue *queue = NULL; local 274 struct xenvif_queue *queue = NULL; local 306 struct xenvif_queue *queue = NULL; local 322 struct xenvif_queue *queue = NULL; local 561 xenvif_init_queue(struct xenvif_queue *queue) argument 671 xenvif_disconnect_queue(struct xenvif_queue *queue) argument 703 xenvif_connect_data(struct xenvif_queue *queue, unsigned long tx_ring_ref, unsigned long rx_ring_ref, unsigned int tx_evtchn, unsigned int rx_evtchn) argument 804 struct xenvif_queue *queue = NULL; local 838 xenvif_deinit_queue(struct xenvif_queue *queue) argument [all...] |
/linux-master/drivers/net/ |
H A D | xen-netfront.c | 94 /* IRQ name is queue name with "-tx" or "-rx" appended */ 166 /* Multi-queue support */ 218 static struct sk_buff *xennet_get_rx_skb(struct netfront_queue *queue, argument 222 struct sk_buff *skb = queue->rx_skbs[i]; 223 queue->rx_skbs[i] = NULL; 227 static grant_ref_t xennet_get_rx_ref(struct netfront_queue *queue, argument 231 grant_ref_t ref = queue->grant_rx_ref[i]; 232 queue->grant_rx_ref[i] = INVALID_GRANT_REF; 248 struct netfront_queue *queue = from_timer(queue, local 252 netfront_tx_slot_available(struct netfront_queue *queue) argument 258 xennet_maybe_wake_tx(struct netfront_queue *queue) argument 270 xennet_alloc_one_rx_buffer(struct netfront_queue *queue) argument 298 xennet_alloc_rx_buffers(struct netfront_queue *queue) argument 365 struct netfront_queue *queue = NULL; local 389 xennet_tx_buf_gc(struct netfront_queue *queue) argument 465 struct netfront_queue *queue; member in struct:xennet_gnttab_make_txreq 482 struct netfront_queue *queue = info->queue; local 603 xennet_mark_tx_pending(struct netfront_queue *queue) argument 612 xennet_xdp_xmit_one(struct net_device *dev, struct netfront_queue *queue, struct xdp_frame *xdpf) argument 650 struct netfront_queue *queue = NULL; local 719 struct netfront_queue *queue = NULL; local 872 struct netfront_queue *queue; local 886 struct netfront_queue *queue = &info->queues[i]; local 903 xennet_set_rx_rsp_cons(struct netfront_queue *queue, RING_IDX val) argument 913 xennet_move_rx_slot(struct netfront_queue *queue, struct sk_buff *skb, grant_ref_t ref) argument 926 xennet_get_extras(struct netfront_queue *queue, struct xen_netif_extra_info *extras, RING_IDX rp) argument 968 xennet_run_xdp(struct netfront_queue *queue, struct page *pdata, struct xen_netif_rx_response *rx, struct bpf_prog *prog, struct xdp_buff *xdp, bool *need_xdp_flush) argument 1015 xennet_get_responses(struct netfront_queue *queue, struct netfront_rx_info *rinfo, RING_IDX rp, struct sk_buff_head *list, bool *need_xdp_flush) argument 1158 xennet_fill_frags(struct netfront_queue *queue, struct sk_buff *skb, struct sk_buff_head *list) argument 1221 handle_incoming_queue(struct netfront_queue *queue, struct sk_buff_head *rxq) argument 1259 struct netfront_queue *queue = container_of(napi, struct netfront_queue, napi); local 1417 xennet_release_tx_bufs(struct netfront_queue *queue) argument 1439 xennet_release_rx_bufs(struct netfront_queue *queue) argument 1508 xennet_handle_tx(struct netfront_queue *queue, unsigned int *eoi) argument 1533 xennet_handle_rx(struct netfront_queue *queue, unsigned int *eoi) argument 1817 struct netfront_queue *queue = &info->queues[i]; local 1899 setup_netfront_single(struct netfront_queue *queue) argument 1925 setup_netfront_split(struct netfront_queue *queue) argument 1969 setup_netfront(struct xenbus_device *dev, struct netfront_queue *queue, unsigned int feature_split_evtchn) argument 2020 xennet_init_queue(struct netfront_queue *queue) argument 2076 write_queue_xenstore_keys(struct netfront_queue *queue, struct xenbus_transaction *xbt, int write_hierarchical) argument 2160 xennet_create_page_pool(struct netfront_queue *queue) argument 2215 struct netfront_queue *queue = &info->queues[i]; local 2260 struct netfront_queue *queue = NULL; local 2424 struct netfront_queue *queue = NULL; local [all...] |
/linux-master/block/ |
H A D | blk-mq-cpumap.c | 3 * CPU <-> hardware queue mapping helpers 21 unsigned int queue, cpu; local 30 for (queue = 0; queue < qmap->nr_queues; queue++) { 31 for_each_cpu(cpu, &masks[queue]) 32 qmap->mq_map[cpu] = qmap->queue_offset + queue; 39 * blk_mq_hw_queue_to_node - Look up the memory node for a hardware queue index 40 * @qmap: CPU to hardware queue map. 41 * @index: hardware queue inde [all...] |
/linux-master/drivers/md/dm-vdo/ |
H A D | funnel-queue.h | 13 * A funnel queue is a simple (almost) lock-free queue that accepts entries from multiple threads 19 * mechanism to ensure that only one thread is consuming from the queue. If more than one thread 20 * attempts to consume from the queue, the resulting behavior is undefined. Clients must not 21 * directly access or manipulate the internals of the queue, which are only exposed for the purpose 25 * the queue entries, and pointers to those structures are used exclusively by the queue. No macros 26 * are defined to template the queue, so the offset of the funnel_queue_entry in the records placed 27 * in the queue must all be the same so the client can derive their structure pointer from the 31 * soon as they are returned since this queue i 82 vdo_funnel_queue_put(struct funnel_queue *queue, struct funnel_queue_entry *entry) argument [all...] |
H A D | funnel-workqueue.h | 33 void vdo_enqueue_work_queue(struct vdo_work_queue *queue, struct vdo_completion *completion); 35 void vdo_finish_work_queue(struct vdo_work_queue *queue); 37 void vdo_free_work_queue(struct vdo_work_queue *queue); 39 void vdo_dump_work_queue(struct vdo_work_queue *queue); 46 struct vdo_thread *vdo_get_work_queue_owner(struct vdo_work_queue *queue); 48 bool __must_check vdo_work_queue_type_is(struct vdo_work_queue *queue,
|
/linux-master/drivers/net/wireless/ath/ath5k/ |
H A D | qcu.c | 35 * basically we have 10 queues to play with. Each queue has a matching 36 * QCU that controls when the queue will get triggered and multiple QCUs 39 * and DCUs allowing us to have different DFS settings for each queue. 41 * When a frame goes into a TX queue, QCU decides when it'll trigger a 43 * it's buffer or -if it's a beacon queue- if it's time to fire up the queue 58 * ath5k_hw_num_tx_pending() - Get number of pending frames for a given queue 60 * @queue: One of enum ath5k_tx_queue_id 63 ath5k_hw_num_tx_pending(struct ath5k_hw *ah, unsigned int queue) argument 66 AR5K_ASSERT_ENTRY(queue, a 94 ath5k_hw_release_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument 138 ath5k_hw_get_tx_queueprops(struct ath5k_hw *ah, int queue, struct ath5k_txq_info *queue_info) argument 154 ath5k_hw_set_tx_queueprops(struct ath5k_hw *ah, int queue, const struct ath5k_txq_info *qinfo) argument 206 unsigned int queue; local 281 ath5k_hw_set_tx_retry_limits(struct ath5k_hw *ah, unsigned int queue) argument 324 ath5k_hw_reset_tx_queue(struct ath5k_hw *ah, unsigned int queue) argument [all...] |
/linux-master/drivers/soc/ixp4xx/ |
H A D | ixp4xx-qmgr.c | 29 void qmgr_put_entry(unsigned int queue, u32 val) argument 32 BUG_ON(!qmgr_queue_descs[queue]); /* not yet requested */ 35 qmgr_queue_descs[queue], queue, val); 37 __raw_writel(val, &qmgr_regs->acc[queue][0]); 40 u32 qmgr_get_entry(unsigned int queue) argument 43 val = __raw_readl(&qmgr_regs->acc[queue][0]); 45 BUG_ON(!qmgr_queue_descs[queue]); /* not yet requested */ 48 qmgr_queue_descs[queue], queue, va 53 __qmgr_get_stat1(unsigned int queue) argument 59 __qmgr_get_stat2(unsigned int queue) argument 72 qmgr_stat_empty(unsigned int queue) argument 84 qmgr_stat_below_low_watermark(unsigned int queue) argument 98 qmgr_stat_full(unsigned int queue) argument 112 qmgr_stat_overflow(unsigned int queue) argument 117 qmgr_set_irq(unsigned int queue, int src, void (*handler)(void *pdev), void *pdev) argument 205 qmgr_enable_irq(unsigned int queue) argument 217 qmgr_disable_irq(unsigned int queue) argument 239 qmgr_request_queue(unsigned int queue, unsigned int len , unsigned int nearly_empty_watermark, unsigned int nearly_full_watermark, const char *desc_format, const char* name) argument 329 qmgr_release_queue(unsigned int queue) argument [all...] |
/linux-master/net/sunrpc/ |
H A D | sched.c | 91 * queue->lock and bh_disabled in order to avoid races within 95 __rpc_disable_timer(struct rpc_wait_queue *queue, struct rpc_task *task) argument 101 if (list_empty(&queue->timer_list.list)) 102 cancel_delayed_work(&queue->timer_list.dwork); 106 rpc_set_queue_timer(struct rpc_wait_queue *queue, unsigned long expires) argument 109 queue->timer_list.expires = expires; 114 mod_delayed_work(rpciod_workqueue, &queue->timer_list.dwork, expires); 121 __rpc_add_timer(struct rpc_wait_queue *queue, struct rpc_task *task, argument 125 if (list_empty(&queue->timer_list.list) || time_before(timeout, queue 130 rpc_set_waitqueue_priority(struct rpc_wait_queue *queue, int priority) argument 138 rpc_reset_waitqueue_priority(struct rpc_wait_queue *queue) argument 193 __rpc_add_wait_queue_priority(struct rpc_wait_queue *queue, struct rpc_task *task, unsigned char queue_priority) argument 205 __rpc_add_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task, unsigned char queue_priority) argument 233 __rpc_remove_wait_queue(struct rpc_wait_queue *queue, struct rpc_task *task) argument 243 __rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname, unsigned char nr_queues) argument 259 rpc_init_priority_wait_queue(struct rpc_wait_queue *queue, const char *qname) argument 265 rpc_init_wait_queue(struct rpc_wait_queue *queue, const char *qname) argument 271 rpc_destroy_wait_queue(struct rpc_wait_queue *queue) argument 506 __rpc_do_wake_up_task_on_wq(struct workqueue_struct *wq, struct rpc_wait_queue *queue, struct rpc_task *task) argument 527 rpc_wake_up_task_on_wq_queue_action_locked(struct workqueue_struct *wq, struct rpc_wait_queue *queue, struct rpc_task *task, bool (*action)(struct rpc_task *, void *), void *data) argument 546 rpc_wake_up_task_queue_locked(struct rpc_wait_queue *queue, struct rpc_task *task) argument 556 rpc_wake_up_queued_task(struct rpc_wait_queue *queue, struct rpc_task *task) argument 573 rpc_wake_up_task_queue_set_status_locked(struct rpc_wait_queue *queue, struct rpc_task *task, int status) argument 590 rpc_wake_up_queued_task_set_status(struct rpc_wait_queue *queue, struct rpc_task *task, int status) argument 603 __rpc_find_next_queued_priority(struct rpc_wait_queue *queue) argument 650 __rpc_find_next_queued(struct rpc_wait_queue *queue) argument 662 rpc_wake_up_first_on_wq(struct workqueue_struct *wq, struct rpc_wait_queue *queue, bool (*func)(struct rpc_task *, void *), void *data) argument 681 rpc_wake_up_first(struct rpc_wait_queue *queue, bool (*func)(struct rpc_task *, void *), void *data) argument 696 rpc_wake_up_next(struct rpc_wait_queue *queue) argument 707 rpc_wake_up_locked(struct rpc_wait_queue *queue) argument 725 rpc_wake_up(struct rpc_wait_queue *queue) argument 738 rpc_wake_up_status_locked(struct rpc_wait_queue *queue, int status) argument 757 rpc_wake_up_status(struct rpc_wait_queue *queue, int status) argument 767 struct rpc_wait_queue *queue = container_of(work, local 857 struct rpc_wait_queue *queue; local 874 struct rpc_wait_queue *queue; local 911 struct rpc_wait_queue *queue; local [all...] |
/linux-master/arch/powerpc/boot/dts/fsl/ |
H A D | pq3-etsec2-grp2-0.dtsi | 36 queue-group@b4000 {
|
H A D | pq3-etsec2-grp2-1.dtsi | 36 queue-group@b5000 {
|
H A D | pq3-etsec2-grp2-2.dtsi | 36 queue-group@b6000 {
|
/linux-master/drivers/staging/rtl8723bs/include/ |
H A D | cmd_osdep.h | 15 int _rtw_enqueue_cmd(struct __queue *queue, struct cmd_obj *obj); 16 extern struct cmd_obj *_rtw_dequeue_cmd(struct __queue *queue);
|
/linux-master/drivers/net/wireless/st/cw1200/ |
H A D | queue.h | 3 * O(1) TX queue with built-in allocator for ST-Ericsson CW1200 drivers 33 struct list_head queue; member in struct:cw1200_queue 39 spinlock_t lock; /* Protect queue entry */ 68 int cw1200_queue_init(struct cw1200_queue *queue, 73 int cw1200_queue_clear(struct cw1200_queue *queue); 75 void cw1200_queue_deinit(struct cw1200_queue *queue); 77 size_t cw1200_queue_get_num_queued(struct cw1200_queue *queue, 79 int cw1200_queue_put(struct cw1200_queue *queue, 82 int cw1200_queue_get(struct cw1200_queue *queue, 87 int cw1200_queue_requeue(struct cw1200_queue *queue, u3 [all...] |
/linux-master/drivers/scsi/arm/ |
H A D | Makefile | 8 obj-$(CONFIG_SCSI_ACORNSCSI_3) += acornscsi_mod.o queue.o msgqueue.o 9 obj-$(CONFIG_SCSI_ARXESCSI) += arxescsi.o fas216.o queue.o msgqueue.o 11 obj-$(CONFIG_SCSI_CUMANA_2) += cumana_2.o fas216.o queue.o msgqueue.o 13 obj-$(CONFIG_SCSI_POWERTECSCSI) += powertec.o fas216.o queue.o msgqueue.o 14 obj-$(CONFIG_SCSI_EESOXSCSI) += eesox.o fas216.o queue.o msgqueue.o
|
/linux-master/include/net/ |
H A D | netdev_rx_queue.h | 10 /* This structure contains an instance of an RX queue. */ 24 /* NAPI instance for the queue 31 * RX queue sysfs structures and functions. 35 ssize_t (*show)(struct netdev_rx_queue *queue, char *buf); 36 ssize_t (*store)(struct netdev_rx_queue *queue, 48 get_netdev_rx_queue_index(struct netdev_rx_queue *queue) argument 50 struct net_device *dev = queue->dev; 51 int index = queue - dev->_rx;
|
/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_submitqueue.c | 71 struct msm_gpu_submitqueue *queue = container_of(kref, local 74 idr_destroy(&queue->fence_idr); 76 msm_file_private_put(queue->ctx); 78 kfree(queue); 162 struct msm_gpu_submitqueue *queue; local 177 queue = kzalloc(sizeof(*queue), GFP_KERNEL); 179 if (!queue) 182 kref_init(&queue->ref); 183 queue 237 msm_submitqueue_query_faults(struct msm_gpu_submitqueue *queue, struct drm_msm_submitqueue_query *args) argument 260 struct msm_gpu_submitqueue *queue; local [all...] |
/linux-master/drivers/crypto/hisilicon/sec/ |
H A D | sec_drv.c | 227 static int sec_queue_map_io(struct sec_queue *queue) argument 229 struct device *dev = queue->dev_info->dev; 234 2 + queue->queue_id); 236 dev_err(dev, "Failed to get queue %u memory resource\n", 237 queue->queue_id); 240 queue->regs = ioremap(res->start, resource_size(res)); 241 if (!queue->regs) 247 static void sec_queue_unmap_io(struct sec_queue *queue) argument 249 iounmap(queue->regs); 252 static int sec_queue_ar_pkgattr(struct sec_queue *queue, u3 argument 267 sec_queue_aw_pkgattr(struct sec_queue *queue, u32 aw_pkg) argument 519 sec_queue_ar_alloc(struct sec_queue *queue, u32 alloc) argument 536 sec_queue_aw_alloc(struct sec_queue *queue, u32 alloc) argument 553 sec_queue_reorder(struct sec_queue *queue, bool reorder) argument 566 sec_queue_depth(struct sec_queue *queue, u32 depth) argument 578 sec_queue_cmdbase_addr(struct sec_queue *queue, u64 addr) argument 584 sec_queue_outorder_addr(struct sec_queue *queue, u64 addr) argument 592 sec_queue_errbase_addr(struct sec_queue *queue, u64 addr) argument 600 sec_queue_irq_disable(struct sec_queue *queue) argument 605 sec_queue_irq_enable(struct sec_queue *queue) argument 610 sec_queue_abn_irq_disable(struct sec_queue *queue) argument 615 sec_queue_stop(struct sec_queue *queue) argument 622 sec_queue_start(struct sec_queue *queue) argument 651 sec_queue_free(struct sec_queue *queue) argument 681 struct sec_queue *queue = q; local 728 sec_queue_irq_init(struct sec_queue *queue) argument 745 sec_queue_irq_uninit(struct sec_queue *queue) argument 774 struct sec_queue *queue; local 799 struct sec_queue *queue = ERR_PTR(-ENODEV); local 821 sec_queue_stop_release(struct sec_queue *queue) argument 843 sec_queue_empty(struct sec_queue *queue) argument 858 sec_queue_send(struct sec_queue *queue, struct sec_bd_info *msg, void *ctx) argument 885 sec_queue_can_enqueue(struct sec_queue *queue, int num) argument 892 sec_queue_hw_init(struct sec_queue *queue) argument 1001 sec_queue_base_init(struct sec_dev_info *info, struct sec_queue *queue, int queue_id) argument 1078 sec_queue_res_cfg(struct sec_queue *queue) argument 1130 sec_queue_free_ring_pages(struct sec_queue *queue) argument 1142 sec_queue_config(struct sec_dev_info *info, struct sec_queue *queue, int queue_id) argument 1165 sec_queue_unconfig(struct sec_dev_info *info, struct sec_queue *queue) argument [all...] |