/linux-master/drivers/media/platform/ti/omap3isp/ |
H A D | ispvideo.h | 84 spinlock_t lock; /* Pipeline state and queue flags */ 125 * @irqlist: List head for insertion into IRQ queue 137 /* Set if DMA queue becomes empty when ISP_PIPELINE_STREAM_CONTINUOUS */ 139 /* Set when queuing buffer to an empty DMA queue */ 148 * @queue: Resume streaming when a buffer is queued. Called on VIDIOC_QBUF 152 int(*queue)(struct isp_video *video, struct isp_buffer *buffer); member in struct:isp_video_operations 177 /* Video buffers queue */ 178 struct vb2_queue *queue; member in struct:isp_video 179 struct mutex queue_lock; /* protects the queue */ 192 struct vb2_queue queue; member in struct:isp_video_fh [all...] |
/linux-master/drivers/staging/media/omap4iss/ |
H A D | iss_video.h | 78 spinlock_t lock; /* Pipeline state and queue flags */ 128 /* Set if DMA queue becomes empty when ISS_PIPELINE_STREAM_CONTINUOUS */ 130 /* Set when queuing buffer to an empty DMA queue */ 139 * @queue: Resume streaming when a buffer is queued. Called on VIDIOC_QBUF 143 int (*queue)(struct iss_video *video, struct iss_buffer *buffer); member in struct:iss_video_operations 168 /* Video buffers queue */ 169 struct vb2_queue *queue; member in struct:iss_video 182 struct vb2_queue queue; member in struct:iss_video_fh 189 container_of(q, struct iss_video_fh, queue)
|
/linux-master/drivers/usb/gadget/function/ |
H A D | uvc_v4l2.c | 245 if (fmt->type != video->queue.queue.type) 390 if (b->type != video->queue.queue.type) 393 return uvcg_alloc_buffers(&video->queue, b); 403 return uvcg_query_buffer(&video->queue, b); 414 ret = uvcg_queue_buffer(&video->queue, b); 431 return uvcg_dequeue_buffer(&video->queue, b, file->f_flags & O_NONBLOCK); 442 if (type != video->queue.queue [all...] |
/linux-master/tools/testing/selftests/netfilter/ |
H A D | Makefile | 18 TEST_GEN_FILES = nf-queue connect_close audit_logread sctp_collision \
|
/linux-master/kernel/bpf/ |
H A D | cpumap.c | 43 * which queue in bpf_cpu_map_entry contains packets. 66 struct ptr_ring *queue; member in struct:bpf_cpu_map_entry 120 /* The tear-down procedure should have made sure that queue is 121 * empty. See __cpu_map_entry_replace() and work-queue 273 * per CPU stored packets are flushed to this queue. Wait honoring 274 * kthread_stop signal until queue is empty. 276 while (!kthread_should_stop() || !__ptr_ring_empty(rcpu->queue)) { 286 if (__ptr_ring_empty(rcpu->queue)) { 289 if (__ptr_ring_empty(rcpu->queue)) { 304 * consume side valid as no-resize allowed of queue [all...] |
/linux-master/drivers/net/ethernet/ti/icssg/ |
H A D | icssg_config.c | 78 * @queue: Queue number 79 * @pd_addr_start: Packet descriptor queue reserved memory 81 * @special: Indicates whether this queue is a special queue or not 84 int queue; member in struct:map 90 /* Hardware queue map for ICSSG */ 148 int queue = 0, i, j; local 153 queue = ICSSG_NUM_TX_QUEUES; 156 regmap_write(miig_rt, ICSSG_QUEUE_RESET_OFFSET, queue); 157 queue [all...] |
/linux-master/drivers/misc/vmw_vmci/ |
H A D | vmci_queue_pair.c | 62 * In more detail. When a VMCI queue pair is first created, it will be in the 71 * vmci_qp_broker_set_page_store to specify the UVAs of the queue pair at 76 * - VMCIQPB_CREATED_MEM: this state is the result when the queue pair 77 * is created by a VMX using the queue pair device backend that 78 * sets the UVAs of the queue pair immediately and stores the 82 * Once the queue pair is in one of the created states (with the exception of 84 * queue pair. Again we have two new states possible: 89 * - from VMCIQPB_CREATED_NO_MEM when a new-style VMX allocates a queue 90 * pair, and attaches to a queue pair previously created by the host side. 92 * - from VMCIQPB_CREATED_MEM when the host side attaches to a queue pai 250 struct vmci_queue *queue = q; local 274 struct vmci_queue *queue; local 331 qp_memcpy_to_queue_iter(struct vmci_queue *queue, u64 queue_offset, struct iov_iter *from, size_t size) argument 379 qp_memcpy_from_queue_iter(struct iov_iter *to, const struct vmci_queue *queue, u64 queue_offset, size_t size) argument 528 struct vmci_queue *queue; local 566 qp_host_free_queue(struct vmci_queue *queue, u64 queue_size) argument 610 qp_acquire_queue_mutex(struct vmci_queue *queue) argument 621 qp_release_queue_mutex(struct vmci_queue *queue) argument [all...] |
/linux-master/arch/mips/cavium-octeon/executive/ |
H A D | Makefile | 13 obj-y += cvmx-pko.o cvmx-spi.o cvmx-cmd-queue.o \
|
/linux-master/drivers/net/ethernet/aquantia/atlantic/hw_atl2/ |
H A D | hw_atl2_llh.h | 16 u32 queue); 36 u32 queue); 41 /* set tx random TC-queue mapping enable bit */
|
/linux-master/drivers/net/ethernet/freescale/ |
H A D | ucc_geth_ethtool.c | 10 * Can only get/set settings of the first queue. 216 int queue = 0; local 223 ring->rx_pending = ug_info->bdRingLenRx[queue]; 224 ring->rx_mini_pending = ug_info->bdRingLenRx[queue]; 225 ring->rx_jumbo_pending = ug_info->bdRingLenRx[queue]; 226 ring->tx_pending = ug_info->bdRingLenTx[queue]; 237 int queue = 0, ret = 0; local 258 ug_info->bdRingLenRx[queue] = ring->rx_pending; 259 ug_info->bdRingLenTx[queue] = ring->tx_pending;
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_device_queue_manager_cik.c | 39 struct queue *q, 135 struct queue *q,
|
H A D | kfd_device_queue_manager_vi.c | 39 struct queue *q, 141 struct queue *q,
|
/linux-master/drivers/dma/ |
H A D | fsl-qdma.c | 149 * @cfg8b_w1: Compound descriptor command queue origin produced 175 u8 queue; member in struct:fsl_pre_status 185 struct fsl_qdma_queue *queue; member in struct:fsl_qdma_chan 224 struct fsl_qdma_queue *queue; member in struct:fsl_qdma_engine 312 struct fsl_qdma_queue *fsl_queue = fsl_chan->queue; 378 /* Status notification is enqueued to status queue. */ 400 static int fsl_qdma_pre_request_enqueue_desc(struct fsl_qdma_queue *queue) argument 405 for (i = 0; i < queue->n_cq + FSL_COMMAND_QUEUE_OVERFLLOW; i++) { 410 dma_pool_alloc(queue->comp_pool, GFP_KERNEL, 416 dma_pool_alloc(queue 461 struct fsl_qdma_queue *queue = fsl_chan->queue; local [all...] |
/linux-master/drivers/net/wwan/t7xx/ |
H A D | t7xx_port_proxy.h | 109 int t7xx_port_proxy_recv_skb(struct cldma_queue *queue, struct sk_buff *skb); 110 int t7xx_port_proxy_recv_skb_from_dedicated_queue(struct cldma_queue *queue, struct sk_buff *skb);
|
/linux-master/drivers/md/dm-vdo/ |
H A D | Makefile | 19 funnel-queue.o \ 42 wait-queue.o \
|
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_gfx.c | 44 int pipe, int queue) 51 bit += queue; 57 int *mec, int *pipe, int *queue) 59 *queue = bit % adev->gfx.mec.num_queue_per_pipe; 68 int xcc_id, int mec, int pipe, int queue) 70 return test_bit(amdgpu_gfx_mec_queue_to_bit(adev, mec, pipe, queue), 75 int me, int pipe, int queue) 82 bit += queue; 88 int *me, int *pipe, int *queue) 90 *queue 43 amdgpu_gfx_mec_queue_to_bit(struct amdgpu_device *adev, int mec, int pipe, int queue) argument 56 amdgpu_queue_mask_bit_to_mec_queue(struct amdgpu_device *adev, int bit, int *mec, int *pipe, int *queue) argument 67 amdgpu_gfx_is_mec_queue_enabled(struct amdgpu_device *adev, int xcc_id, int mec, int pipe, int queue) argument 74 amdgpu_gfx_me_queue_to_bit(struct amdgpu_device *adev, int me, int pipe, int queue) argument 87 amdgpu_gfx_bit_to_me_queue(struct amdgpu_device *adev, int bit, int *me, int *pipe, int *queue) argument 97 amdgpu_gfx_is_me_queue_enabled(struct amdgpu_device *adev, int me, int pipe, int queue) argument 176 int queue = ring->queue; local 210 int i, j, queue, pipe; local 246 int i, queue, pipe; local 276 int mec, pipe, queue; local 591 int mec, pipe, queue; local [all...] |
/linux-master/drivers/usb/gadget/udc/ |
H A D | goku_udc.c | 280 INIT_LIST_HEAD(&req->queue); 293 WARN_ON(!list_empty(&req->queue)); 305 list_del_init(&req->queue); 324 /* don't modify queue heads during completion callback */ 491 if (dbuff && !list_empty(&ep->queue)) { 492 req = list_entry(ep->queue.next, 493 struct goku_request, queue); 525 if (unlikely(list_empty (&ep->queue))) 527 req = list_entry(ep->queue.next, struct goku_request, queue); [all...] |
/linux-master/drivers/media/pci/cx23885/ |
H A D | cx23885-vbi.c | 199 list_add_tail(&buf->queue, &q->active); 207 queue); 209 list_add_tail(&buf->queue, &q->active); 222 struct cx23885_buffer, queue); 238 struct cx23885_buffer, queue); 240 list_del(&buf->queue);
|
/linux-master/drivers/media/pci/mgb4/ |
H A D | mgb4_vout.c | 66 * the queue. See video_source_store() in mgb4_sysfs_out.c for 151 dev_dbg(dev, "frame queue error (%d)\n", (int)addr); 241 if (vb2_is_busy(&voutdev->queue)) 362 dev_dbg(dev, "frame queue error (%d)\n", (int)addr); 487 /* Frame queue */ 519 voutdev->queue.type = V4L2_BUF_TYPE_VIDEO_OUTPUT; 520 voutdev->queue.io_modes = VB2_MMAP | VB2_DMABUF | VB2_WRITE; 521 voutdev->queue.buf_struct_size = sizeof(struct mgb4_frame_buffer); 522 voutdev->queue.ops = &queue_ops; 523 voutdev->queue [all...] |
/linux-master/drivers/platform/surface/aggregator/ |
H A D | ssh_request_layer.c | 104 spin_lock(&rtl->queue.lock); 107 spin_unlock(&rtl->queue.lock); 113 spin_unlock(&rtl->queue.lock); 121 spin_lock(&rtl->queue.lock); 122 empty = list_empty(&rtl->queue.head); 123 spin_unlock(&rtl->queue.lock); 212 spin_lock(&rtl->queue.lock); 215 list_for_each_entry_safe(p, n, &rtl->queue.head, node) { 224 /* Remove from queue and mark as transmitting. */ 236 spin_unlock(&rtl->queue [all...] |
/linux-master/drivers/gpu/drm/v3d/ |
H A D | v3d_sched.c | 145 v3d->queue[V3D_BIN].start_ns = file->start_ns[V3D_BIN]; 200 v3d->queue[V3D_RENDER].start_ns = file->start_ns[V3D_RENDER]; 236 v3d->queue[V3D_TFU].start_ns = file->start_ns[V3D_TFU]; 283 v3d->queue[V3D_CSD].start_ns = file->start_ns[V3D_CSD]; 544 v3d->queue[V3D_CPU].start_ns = file->start_ns[V3D_CPU]; 555 v3d->queue[V3D_CPU].enabled_ns += runtime; 558 v3d->queue[V3D_CPU].jobs_sent++; 561 v3d->queue[V3D_CPU].start_ns = 0; 575 v3d->queue[V3D_CACHE_CLEAN].start_ns = file->start_ns[V3D_CACHE_CLEAN]; 582 v3d->queue[V3D_CACHE_CLEA [all...] |
/linux-master/drivers/misc/ |
H A D | ibmvmc.c | 112 /* routines for managing a command/response queue */ 145 struct crq_queue *queue = &adapter->queue; local 155 queue->msg_token, 156 queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL); 157 free_page((unsigned long)queue->msgs); 175 struct crq_queue *queue = &adapter->queue; local 181 /* Clean out the queue */ 205 crq_queue_next_crq(struct crq_queue *queue) argument 2116 struct crq_queue *queue = &adapter->queue; local [all...] |
/linux-master/drivers/net/wireless/realtek/rtw88/ |
H A D | pci.c | 34 enum rtw_tx_queue_type queue) 36 switch (queue) { 101 skb_queue_walk_safe(&tx_ring->queue, skb, tmp) { 102 __skb_unlink(skb, &tx_ring->queue); 197 skb_queue_head_init(&tx_ring->queue); 539 enum rtw_tx_queue_type queue; local 542 for (queue = 0; queue < RTK_MAX_TX_QUEUE_NUM; queue++) { 543 tx_ring = &rtwpci->tx_rings[queue]; 33 rtw_pci_get_tx_qsel(struct sk_buff *skb, enum rtw_tx_queue_type queue) argument 605 enum rtw_tx_queue_type queue; local 770 rtw_pci_tx_kick_off_queue(struct rtw_dev *rtwdev, enum rtw_tx_queue_type queue) argument 790 enum rtw_tx_queue_type queue; local 797 rtw_pci_tx_write_data(struct rtw_dev *rtwdev, struct rtw_tx_pkt_info *pkt_info, struct sk_buff *skb, enum rtw_tx_queue_type queue) argument 918 enum rtw_tx_queue_type queue = rtw_tx_queue_mapping(skb); local [all...] |
/linux-master/drivers/net/ethernet/stmicro/stmmac/ |
H A D | dwxgmac2_core.c | 83 u32 queue) 88 value = readl(ioaddr + XGMAC_RXQ_CTRL0) & ~XGMAC_RXQEN(queue); 90 value |= 0x1 << XGMAC_RXQEN_SHIFT(queue); 92 value |= 0x2 << XGMAC_RXQEN_SHIFT(queue); 97 u32 queue) 117 /* First assign new priorities to a queue, then 120 if (queue < 4) { 121 ctrl2 |= (prio << XGMAC_PSRQ_SHIFT(queue)) & 122 XGMAC_PSRQ(queue); 127 queue 82 dwxgmac2_rx_queue_enable(struct mac_device_info *hw, u8 mode, u32 queue) argument 96 dwxgmac2_rx_queue_prio(struct mac_device_info *hw, u32 prio, u32 queue) argument 137 dwxgmac2_tx_queue_prio(struct mac_device_info *hw, u32 prio, u32 queue) argument 154 dwxgmac2_rx_queue_routing(struct mac_device_info *hw, u8 packet, u32 queue) argument 244 dwxgmac2_set_mtl_tx_queue_weight(struct stmmac_priv *priv, struct mac_device_info *hw, u32 weight, u32 queue) argument 253 dwxgmac2_map_mtl_to_dma(struct mac_device_info *hw, u32 queue, u32 chan) argument 270 dwxgmac2_config_cbs(struct stmmac_priv *priv, struct mac_device_info *hw, u32 send_slope, u32 idle_slope, u32 high_credit, u32 low_credit, u32 queue) argument 1579 dwxlgmac2_rx_queue_enable(struct mac_device_info *hw, u8 mode, u32 queue) argument [all...] |
/linux-master/drivers/net/ethernet/marvell/ |
H A D | mvneta.c | 141 * registers. For a given CPU if the bit associated to a queue is not 149 /* bits 0..7 = TXQ SENT, one bit per queue. 150 * bits 8..15 = RXQ OCCUP, one bit per queue. 151 * bits 16..23 = RXQ FREE, one bit per queue. 661 /* Number of this TX queue, in the range 0-7 */ 709 /* rx queue number, in the range 0-7 */ 744 * the first one to be used. Therefore, let's just allocate one queue. 944 /* Set rx queue offset */ 1254 int queue; local 1259 for (queue 1375 mvneta_set_ucast_table(struct mvneta_port *pp, int queue) argument 1392 mvneta_set_special_mcast_table(struct mvneta_port *pp, int queue) argument 1410 mvneta_set_other_mcast_table(struct mvneta_port *pp, int queue) argument 1477 int queue; local 1597 int queue; local 1633 mvneta_set_ucast_addr(struct mvneta_port *pp, u8 last_nibble, int queue) argument 1663 mvneta_mac_addr_set(struct mvneta_port *pp, const unsigned char *addr, int queue) argument 1862 int queue = fls(cause) - 1; local 3071 mvneta_set_special_mcast_addr(struct mvneta_port *pp, unsigned char last_byte, int queue) argument 3106 mvneta_set_other_mcast_addr(struct mvneta_port *pp, unsigned char crc8, int queue) argument 3139 mvneta_mcast_addr_set(struct mvneta_port *pp, unsigned char *p_addr, int queue) argument 3421 int queue; local 3643 int queue; local 3652 int queue; local 3662 int queue; local 3681 int queue; local 4680 int queue; local 5190 mvneta_setup_queue_rates(struct mvneta_port *pp, int queue, u64 min_rate, u64 max_rate) argument 5343 int queue; local 5765 int queue; local 5812 int err, queue; local [all...] |