/linux-master/drivers/net/wireless/mediatek/mt76/mt7603/ |
H A D | beacon.c | 122 if (dev->mphy.q_tx[MT_TXQ_BEACON]->queued > 0)
|
/linux-master/drivers/net/ethernet/mediatek/ |
H A D | mtk_wed_wo.h | 226 int queued; member in struct:mtk_wed_wo_queue
|
/linux-master/drivers/media/pci/zoran/ |
H A D | zoran.h | 245 unsigned long jpg_que_head; /* Index where to put next buffer which is queued */ 252 unsigned long jpg_queued_num; /* count of frames queued since grab/play started */ 279 unsigned int queued; member in struct:zoran
|
/linux-master/net/caif/ |
H A D | caif_socket.c | 126 bool queued = false; local 149 queued = !sock_flag(sk, SOCK_DEAD); 150 if (queued) 154 if (queued) 902 * Ensure that packets are not queued after this point in time.
|
/linux-master/drivers/net/wireless/mediatek/mt76/mt7996/ |
H A D | debugfs.c | 574 u32 ctrl, head, tail, queued; local 586 queued = mt76_get_field(dev, MT_FL_Q3_CTRL, 590 seq_printf(s, "queued:0x%03x head:0x%03x tail:0x%03x\n", 591 queued, head, tail); 617 seq_printf(s, "\tSTA %pM wcid %d: AC%d%d queued:%d\n", 709 seq_puts(file, " queue | hw-queued | head | tail |\n"); 717 queue_map[i].queue, q->queued, q->head,
|
/linux-master/kernel/sched/ |
H A D | idle.c | 506 static void task_tick_idle(struct rq *rq, struct task_struct *curr, int queued) argument
|
H A D | core.c | 954 * it's already queued (either by us or someone else) and will get the 1005 * queued for wakeup. 1757 * Lock the task and the rq where the task is (or was) queued. 2293 int running, queued, match; local 2332 queued = task_on_rq_queued(p); 2337 * still queued so it will wait. 2340 queued = 1; 2371 if (unlikely(queued)) { 2516 * move_queued_task - move a queued task to new rq. 2757 bool queued, runnin local 7152 int prio, oldprio, queued, running, queue_flag = local 7276 bool queued, running; local 7709 int retval, oldprio, newprio, queued, running; local 9392 bool queued, running; local 10485 int queued, running, queue_flags = local [all...] |
/linux-master/fs/ocfs2/cluster/ |
H A D | heartbeat.c | 814 int queued = 0; local 832 queued = 1; 837 if (queued) 898 int queued = 0; local 992 queued = 1; 1044 queued = 1; 1060 if (queued)
|
/linux-master/block/ |
H A D | blk-mq.c | 2002 static void blk_mq_commit_rqs(struct blk_mq_hw_ctx *hctx, int queued, argument 2005 if (hctx->queue->mq_ops->commit_rqs && queued) { 2006 trace_block_unplug(hctx->queue, queued, !from_schedule); 2020 int queued; local 2031 queued = 0; 2048 * once the request is queued to lld, no need to cover the 2056 queued++; 2085 blk_mq_commit_rqs(hctx, queued, false); 2170 * BLK_MQ_CPU_WORK_BATCH queued items. 2687 int queued local 2824 int queued = 0; local [all...] |
H A D | bfq-iosched.c | 478 if (bfqd->queued != 0) { 686 * queue / cgroup already has many requests allocated and queued, this does not 1866 * processes. So let also stably-merged queued enjoy weight 2214 bfqq->queued[rq_is_sync(rq)]++; 2216 * Updating of 'bfqd->queued' is protected by 'bfqd->lock', however, it 2219 WRITE_ONCE(bfqd->queued, bfqd->queued + 1); 2394 bfqq->queued[sync]--; 2396 * Updating of 'bfqd->queued' is protected by 'bfqd->lock', however, it 2399 WRITE_ONCE(bfqd->queued, bfq [all...] |
H A D | mq-deadline.c | 289 /* Number of requests queued for a given priority level. */ 529 * For a zoned block device, if we only have writes queued and none of 670 uint32_t queued; local 676 queued = dd_queued(dd, prio); 679 WARN_ONCE(queued != 0, 912 * dispatching requests if all the queued requests are write requests directed 1237 {"queued", 0400, dd_queued_show},
|
/linux-master/drivers/net/wireless/mediatek/mt76/mt7915/ |
H A D | debugfs.c | 801 u32 ctrl, head, tail, queued; local 813 queued = mt76_get_field(dev, MT_FL_Q3_CTRL, 817 seq_printf(s, "queued:0x%03x head:0x%03x tail:0x%03x\n", 818 queued, head, tail); 844 seq_printf(s, "\tSTA %pM wcid %d: AC%d%d queued:%d\n", 936 seq_puts(file, " queue | hw-queued | head | tail |\n"); 944 queue_map[i].queue, q->queued, q->head,
|
/linux-master/fs/nfsd/ |
H A D | nfs4callback.c | 1242 * task was queued, we need to resubmit it after setting up 1547 bool queued; local 1550 queued = nfsd4_queue_cb(cb); 1551 if (!queued) 1553 return queued;
|
/linux-master/drivers/dma/qcom/ |
H A D | hidma_ll.c | 160 tre->queued = 0; 174 * Multiple TREs may be queued and waiting in the pending queue. 223 tre->queued = 0; 276 * get queued to the SW for processing. 539 tre->queued = 1;
|
/linux-master/drivers/net/wireless/mediatek/mt76/ |
H A D | tx.c | 443 q->queued + MT_TXQ_FREE_THR >= q->ndesc; 529 q->queued + 2 * MT_TXQ_FREE_THR >= q->ndesc) { 789 q->queued--;
|
/linux-master/tools/testing/selftests/net/mptcp/ |
H A D | mptcp_connect.c | 1218 int queued = 1; local 1227 if (ioctl(fd, SIOCOUTQ, &queued) < 0) 1230 if (!queued)
|
/linux-master/arch/powerpc/platforms/powernv/ |
H A D | opal.c | 310 * If the registration succeeded, replay any queued messages that came 334 bool queued = false; local 343 queued = true; 347 if (queued)
|
/linux-master/drivers/tty/hvc/ |
H A D | hvc_iucv.c | 461 int queued; local 470 queued = hvc_iucv_queue(priv, buf, count); 473 return queued; 547 sync_wait = !list_empty(&priv->tty_outqueue); /* anything queued ? */
|
/linux-master/drivers/net/wireless/ath/carl9170/ |
H A D | tx.c | 1434 goto queued; 1442 goto queued; 1447 queued:
|
/linux-master/drivers/net/wireless/st/cw1200/ |
H A D | wsm.c | 1584 int queued; local 1589 queued = cw1200_queue_get_num_queued(&priv->tx_queue[i], 1591 if (!queued) 1593 *total += queued;
|
/linux-master/net/netfilter/ |
H A D | nfnetlink_queue.c | 74 * Following fields are dirtied for each queued packet, 958 unsigned int queued; local 995 queued = 0; 1002 queued++; 1007 if (queued) { 1008 if (err) /* some segments are already queued */
|
/linux-master/include/linux/ |
H A D | iommu.h | 330 * @queued: Indicates that the flush will be queued 336 * them. @queued is set to indicate when ->iotlb_flush_all() will be called 344 bool queued; member in struct:iommu_iotlb_gather 613 * @iotlb_sync: Flush all queued ranges from the hardware TLBs and empty flush 933 return gather && gather->queued;
|
/linux-master/fs/ |
H A D | aio.c | 1877 bool queued; member in struct:aio_poll_table 1888 if (unlikely(pt->queued)) { 1893 pt->queued = true; 1925 apt.queued = false; 1934 if (likely(apt.queued)) { 2077 * the number of iocbs queued. May return -EINVAL if the aio_context
|
/linux-master/drivers/crypto/inside-secure/ |
H A D | safexcel_cipher.c | 693 int queued, i, ret = 0; local 734 queued = totlen_src; 799 if (queued < len) 800 len = queued; 803 !(queued - len), 816 queued -= len; 817 if (!queued)
|
/linux-master/drivers/iommu/ |
H A D | dma-iommu.c | 265 /* The IOVAs will be torn down separately, so just free our queued pages */ 817 else if (gather && gather->queued) 839 iotlb_gather.queued = READ_ONCE(cookie->fq_domain); 844 if (!iotlb_gather.queued)
|