Searched refs:queue (Results 226 - 250 of 1368) sorted by relevance

1234567891011>>

/linux-master/drivers/media/platform/ti/omap3isp/
H A Dispvideo.h84 spinlock_t lock; /* Pipeline state and queue flags */
125 * @irqlist: List head for insertion into IRQ queue
137 /* Set if DMA queue becomes empty when ISP_PIPELINE_STREAM_CONTINUOUS */
139 /* Set when queuing buffer to an empty DMA queue */
148 * @queue: Resume streaming when a buffer is queued. Called on VIDIOC_QBUF
152 int(*queue)(struct isp_video *video, struct isp_buffer *buffer); member in struct:isp_video_operations
177 /* Video buffers queue */
178 struct vb2_queue *queue; member in struct:isp_video
179 struct mutex queue_lock; /* protects the queue */
192 struct vb2_queue queue; member in struct:isp_video_fh
[all...]
/linux-master/drivers/staging/media/omap4iss/
H A Diss_video.h78 spinlock_t lock; /* Pipeline state and queue flags */
128 /* Set if DMA queue becomes empty when ISS_PIPELINE_STREAM_CONTINUOUS */
130 /* Set when queuing buffer to an empty DMA queue */
139 * @queue: Resume streaming when a buffer is queued. Called on VIDIOC_QBUF
143 int (*queue)(struct iss_video *video, struct iss_buffer *buffer); member in struct:iss_video_operations
168 /* Video buffers queue */
169 struct vb2_queue *queue; member in struct:iss_video
182 struct vb2_queue queue; member in struct:iss_video_fh
189 container_of(q, struct iss_video_fh, queue)
/linux-master/drivers/usb/gadget/function/
H A Duvc_v4l2.c245 if (fmt->type != video->queue.queue.type)
390 if (b->type != video->queue.queue.type)
393 return uvcg_alloc_buffers(&video->queue, b);
403 return uvcg_query_buffer(&video->queue, b);
414 ret = uvcg_queue_buffer(&video->queue, b);
431 return uvcg_dequeue_buffer(&video->queue, b, file->f_flags & O_NONBLOCK);
442 if (type != video->queue.queue
[all...]
/linux-master/tools/testing/selftests/netfilter/
H A DMakefile18 TEST_GEN_FILES = nf-queue connect_close audit_logread sctp_collision \
/linux-master/kernel/bpf/
H A Dcpumap.c43 * which queue in bpf_cpu_map_entry contains packets.
66 struct ptr_ring *queue; member in struct:bpf_cpu_map_entry
120 /* The tear-down procedure should have made sure that queue is
121 * empty. See __cpu_map_entry_replace() and work-queue
273 * per CPU stored packets are flushed to this queue. Wait honoring
274 * kthread_stop signal until queue is empty.
276 while (!kthread_should_stop() || !__ptr_ring_empty(rcpu->queue)) {
286 if (__ptr_ring_empty(rcpu->queue)) {
289 if (__ptr_ring_empty(rcpu->queue)) {
304 * consume side valid as no-resize allowed of queue
[all...]
/linux-master/drivers/net/ethernet/ti/icssg/
H A Dicssg_config.c78 * @queue: Queue number
79 * @pd_addr_start: Packet descriptor queue reserved memory
81 * @special: Indicates whether this queue is a special queue or not
84 int queue; member in struct:map
90 /* Hardware queue map for ICSSG */
148 int queue = 0, i, j; local
153 queue = ICSSG_NUM_TX_QUEUES;
156 regmap_write(miig_rt, ICSSG_QUEUE_RESET_OFFSET, queue);
157 queue
[all...]
/linux-master/drivers/misc/vmw_vmci/
H A Dvmci_queue_pair.c62 * In more detail. When a VMCI queue pair is first created, it will be in the
71 * vmci_qp_broker_set_page_store to specify the UVAs of the queue pair at
76 * - VMCIQPB_CREATED_MEM: this state is the result when the queue pair
77 * is created by a VMX using the queue pair device backend that
78 * sets the UVAs of the queue pair immediately and stores the
82 * Once the queue pair is in one of the created states (with the exception of
84 * queue pair. Again we have two new states possible:
89 * - from VMCIQPB_CREATED_NO_MEM when a new-style VMX allocates a queue
90 * pair, and attaches to a queue pair previously created by the host side.
92 * - from VMCIQPB_CREATED_MEM when the host side attaches to a queue pai
250 struct vmci_queue *queue = q; local
274 struct vmci_queue *queue; local
331 qp_memcpy_to_queue_iter(struct vmci_queue *queue, u64 queue_offset, struct iov_iter *from, size_t size) argument
379 qp_memcpy_from_queue_iter(struct iov_iter *to, const struct vmci_queue *queue, u64 queue_offset, size_t size) argument
528 struct vmci_queue *queue; local
566 qp_host_free_queue(struct vmci_queue *queue, u64 queue_size) argument
610 qp_acquire_queue_mutex(struct vmci_queue *queue) argument
621 qp_release_queue_mutex(struct vmci_queue *queue) argument
[all...]
/linux-master/arch/mips/cavium-octeon/executive/
H A DMakefile13 obj-y += cvmx-pko.o cvmx-spi.o cvmx-cmd-queue.o \
/linux-master/drivers/net/ethernet/aquantia/atlantic/hw_atl2/
H A Dhw_atl2_llh.h16 u32 queue);
36 u32 queue);
41 /* set tx random TC-queue mapping enable bit */
/linux-master/drivers/net/ethernet/freescale/
H A Ducc_geth_ethtool.c10 * Can only get/set settings of the first queue.
216 int queue = 0; local
223 ring->rx_pending = ug_info->bdRingLenRx[queue];
224 ring->rx_mini_pending = ug_info->bdRingLenRx[queue];
225 ring->rx_jumbo_pending = ug_info->bdRingLenRx[queue];
226 ring->tx_pending = ug_info->bdRingLenTx[queue];
237 int queue = 0, ret = 0; local
258 ug_info->bdRingLenRx[queue] = ring->rx_pending;
259 ug_info->bdRingLenTx[queue] = ring->tx_pending;
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_device_queue_manager_cik.c39 struct queue *q,
135 struct queue *q,
H A Dkfd_device_queue_manager_vi.c39 struct queue *q,
141 struct queue *q,
/linux-master/drivers/dma/
H A Dfsl-qdma.c149 * @cfg8b_w1: Compound descriptor command queue origin produced
175 u8 queue; member in struct:fsl_pre_status
185 struct fsl_qdma_queue *queue; member in struct:fsl_qdma_chan
224 struct fsl_qdma_queue *queue; member in struct:fsl_qdma_engine
312 struct fsl_qdma_queue *fsl_queue = fsl_chan->queue;
378 /* Status notification is enqueued to status queue. */
400 static int fsl_qdma_pre_request_enqueue_desc(struct fsl_qdma_queue *queue) argument
405 for (i = 0; i < queue->n_cq + FSL_COMMAND_QUEUE_OVERFLLOW; i++) {
410 dma_pool_alloc(queue->comp_pool, GFP_KERNEL,
416 dma_pool_alloc(queue
461 struct fsl_qdma_queue *queue = fsl_chan->queue; local
[all...]
/linux-master/drivers/net/wwan/t7xx/
H A Dt7xx_port_proxy.h109 int t7xx_port_proxy_recv_skb(struct cldma_queue *queue, struct sk_buff *skb);
110 int t7xx_port_proxy_recv_skb_from_dedicated_queue(struct cldma_queue *queue, struct sk_buff *skb);
/linux-master/drivers/md/dm-vdo/
H A DMakefile19 funnel-queue.o \
42 wait-queue.o \
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_gfx.c44 int pipe, int queue)
51 bit += queue;
57 int *mec, int *pipe, int *queue)
59 *queue = bit % adev->gfx.mec.num_queue_per_pipe;
68 int xcc_id, int mec, int pipe, int queue)
70 return test_bit(amdgpu_gfx_mec_queue_to_bit(adev, mec, pipe, queue),
75 int me, int pipe, int queue)
82 bit += queue;
88 int *me, int *pipe, int *queue)
90 *queue
43 amdgpu_gfx_mec_queue_to_bit(struct amdgpu_device *adev, int mec, int pipe, int queue) argument
56 amdgpu_queue_mask_bit_to_mec_queue(struct amdgpu_device *adev, int bit, int *mec, int *pipe, int *queue) argument
67 amdgpu_gfx_is_mec_queue_enabled(struct amdgpu_device *adev, int xcc_id, int mec, int pipe, int queue) argument
74 amdgpu_gfx_me_queue_to_bit(struct amdgpu_device *adev, int me, int pipe, int queue) argument
87 amdgpu_gfx_bit_to_me_queue(struct amdgpu_device *adev, int bit, int *me, int *pipe, int *queue) argument
97 amdgpu_gfx_is_me_queue_enabled(struct amdgpu_device *adev, int me, int pipe, int queue) argument
176 int queue = ring->queue; local
210 int i, j, queue, pipe; local
246 int i, queue, pipe; local
276 int mec, pipe, queue; local
591 int mec, pipe, queue; local
[all...]
/linux-master/drivers/usb/gadget/udc/
H A Dgoku_udc.c280 INIT_LIST_HEAD(&req->queue);
293 WARN_ON(!list_empty(&req->queue));
305 list_del_init(&req->queue);
324 /* don't modify queue heads during completion callback */
491 if (dbuff && !list_empty(&ep->queue)) {
492 req = list_entry(ep->queue.next,
493 struct goku_request, queue);
525 if (unlikely(list_empty (&ep->queue)))
527 req = list_entry(ep->queue.next, struct goku_request, queue);
[all...]
/linux-master/drivers/media/pci/cx23885/
H A Dcx23885-vbi.c199 list_add_tail(&buf->queue, &q->active);
207 queue);
209 list_add_tail(&buf->queue, &q->active);
222 struct cx23885_buffer, queue);
238 struct cx23885_buffer, queue);
240 list_del(&buf->queue);
/linux-master/drivers/media/pci/mgb4/
H A Dmgb4_vout.c66 * the queue. See video_source_store() in mgb4_sysfs_out.c for
151 dev_dbg(dev, "frame queue error (%d)\n", (int)addr);
241 if (vb2_is_busy(&voutdev->queue))
362 dev_dbg(dev, "frame queue error (%d)\n", (int)addr);
487 /* Frame queue */
519 voutdev->queue.type = V4L2_BUF_TYPE_VIDEO_OUTPUT;
520 voutdev->queue.io_modes = VB2_MMAP | VB2_DMABUF | VB2_WRITE;
521 voutdev->queue.buf_struct_size = sizeof(struct mgb4_frame_buffer);
522 voutdev->queue.ops = &queue_ops;
523 voutdev->queue
[all...]
/linux-master/drivers/platform/surface/aggregator/
H A Dssh_request_layer.c104 spin_lock(&rtl->queue.lock);
107 spin_unlock(&rtl->queue.lock);
113 spin_unlock(&rtl->queue.lock);
121 spin_lock(&rtl->queue.lock);
122 empty = list_empty(&rtl->queue.head);
123 spin_unlock(&rtl->queue.lock);
212 spin_lock(&rtl->queue.lock);
215 list_for_each_entry_safe(p, n, &rtl->queue.head, node) {
224 /* Remove from queue and mark as transmitting. */
236 spin_unlock(&rtl->queue
[all...]
/linux-master/drivers/gpu/drm/v3d/
H A Dv3d_sched.c145 v3d->queue[V3D_BIN].start_ns = file->start_ns[V3D_BIN];
200 v3d->queue[V3D_RENDER].start_ns = file->start_ns[V3D_RENDER];
236 v3d->queue[V3D_TFU].start_ns = file->start_ns[V3D_TFU];
283 v3d->queue[V3D_CSD].start_ns = file->start_ns[V3D_CSD];
544 v3d->queue[V3D_CPU].start_ns = file->start_ns[V3D_CPU];
555 v3d->queue[V3D_CPU].enabled_ns += runtime;
558 v3d->queue[V3D_CPU].jobs_sent++;
561 v3d->queue[V3D_CPU].start_ns = 0;
575 v3d->queue[V3D_CACHE_CLEAN].start_ns = file->start_ns[V3D_CACHE_CLEAN];
582 v3d->queue[V3D_CACHE_CLEA
[all...]
/linux-master/drivers/misc/
H A Dibmvmc.c112 /* routines for managing a command/response queue */
145 struct crq_queue *queue = &adapter->queue; local
155 queue->msg_token,
156 queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
157 free_page((unsigned long)queue->msgs);
175 struct crq_queue *queue = &adapter->queue; local
181 /* Clean out the queue */
205 crq_queue_next_crq(struct crq_queue *queue) argument
2116 struct crq_queue *queue = &adapter->queue; local
[all...]
/linux-master/drivers/net/wireless/realtek/rtw88/
H A Dpci.c34 enum rtw_tx_queue_type queue)
36 switch (queue) {
101 skb_queue_walk_safe(&tx_ring->queue, skb, tmp) {
102 __skb_unlink(skb, &tx_ring->queue);
197 skb_queue_head_init(&tx_ring->queue);
539 enum rtw_tx_queue_type queue; local
542 for (queue = 0; queue < RTK_MAX_TX_QUEUE_NUM; queue++) {
543 tx_ring = &rtwpci->tx_rings[queue];
33 rtw_pci_get_tx_qsel(struct sk_buff *skb, enum rtw_tx_queue_type queue) argument
605 enum rtw_tx_queue_type queue; local
770 rtw_pci_tx_kick_off_queue(struct rtw_dev *rtwdev, enum rtw_tx_queue_type queue) argument
790 enum rtw_tx_queue_type queue; local
797 rtw_pci_tx_write_data(struct rtw_dev *rtwdev, struct rtw_tx_pkt_info *pkt_info, struct sk_buff *skb, enum rtw_tx_queue_type queue) argument
918 enum rtw_tx_queue_type queue = rtw_tx_queue_mapping(skb); local
[all...]
/linux-master/drivers/net/ethernet/stmicro/stmmac/
H A Ddwxgmac2_core.c83 u32 queue)
88 value = readl(ioaddr + XGMAC_RXQ_CTRL0) & ~XGMAC_RXQEN(queue);
90 value |= 0x1 << XGMAC_RXQEN_SHIFT(queue);
92 value |= 0x2 << XGMAC_RXQEN_SHIFT(queue);
97 u32 queue)
117 /* First assign new priorities to a queue, then
120 if (queue < 4) {
121 ctrl2 |= (prio << XGMAC_PSRQ_SHIFT(queue)) &
122 XGMAC_PSRQ(queue);
127 queue
82 dwxgmac2_rx_queue_enable(struct mac_device_info *hw, u8 mode, u32 queue) argument
96 dwxgmac2_rx_queue_prio(struct mac_device_info *hw, u32 prio, u32 queue) argument
137 dwxgmac2_tx_queue_prio(struct mac_device_info *hw, u32 prio, u32 queue) argument
154 dwxgmac2_rx_queue_routing(struct mac_device_info *hw, u8 packet, u32 queue) argument
244 dwxgmac2_set_mtl_tx_queue_weight(struct stmmac_priv *priv, struct mac_device_info *hw, u32 weight, u32 queue) argument
253 dwxgmac2_map_mtl_to_dma(struct mac_device_info *hw, u32 queue, u32 chan) argument
270 dwxgmac2_config_cbs(struct stmmac_priv *priv, struct mac_device_info *hw, u32 send_slope, u32 idle_slope, u32 high_credit, u32 low_credit, u32 queue) argument
1579 dwxlgmac2_rx_queue_enable(struct mac_device_info *hw, u8 mode, u32 queue) argument
[all...]
/linux-master/drivers/net/ethernet/marvell/
H A Dmvneta.c141 * registers. For a given CPU if the bit associated to a queue is not
149 /* bits 0..7 = TXQ SENT, one bit per queue.
150 * bits 8..15 = RXQ OCCUP, one bit per queue.
151 * bits 16..23 = RXQ FREE, one bit per queue.
661 /* Number of this TX queue, in the range 0-7 */
709 /* rx queue number, in the range 0-7 */
744 * the first one to be used. Therefore, let's just allocate one queue.
944 /* Set rx queue offset */
1254 int queue; local
1259 for (queue
1375 mvneta_set_ucast_table(struct mvneta_port *pp, int queue) argument
1392 mvneta_set_special_mcast_table(struct mvneta_port *pp, int queue) argument
1410 mvneta_set_other_mcast_table(struct mvneta_port *pp, int queue) argument
1477 int queue; local
1597 int queue; local
1633 mvneta_set_ucast_addr(struct mvneta_port *pp, u8 last_nibble, int queue) argument
1663 mvneta_mac_addr_set(struct mvneta_port *pp, const unsigned char *addr, int queue) argument
1862 int queue = fls(cause) - 1; local
3071 mvneta_set_special_mcast_addr(struct mvneta_port *pp, unsigned char last_byte, int queue) argument
3106 mvneta_set_other_mcast_addr(struct mvneta_port *pp, unsigned char crc8, int queue) argument
3139 mvneta_mcast_addr_set(struct mvneta_port *pp, unsigned char *p_addr, int queue) argument
3421 int queue; local
3643 int queue; local
3652 int queue; local
3662 int queue; local
3681 int queue; local
4680 int queue; local
5190 mvneta_setup_queue_rates(struct mvneta_port *pp, int queue, u64 min_rate, u64 max_rate) argument
5343 int queue; local
5765 int queue; local
5812 int err, queue; local
[all...]

Completed in 333 milliseconds

1234567891011>>