/linux-master/drivers/gpu/drm/msm/ |
H A D | msm_rd.c | 380 pid_nr(submit->pid), submit->seqno); 383 pid_nr(submit->pid), submit->seqno);
|
H A D | msm_gpu.c | 317 if (submit->seqno == fence) { 650 int index = submit->seqno % MSM_GPU_SUBMIT_STATS_COUNT; 760 submit->seqno = submit->hw_fence->seqno; 784 gpu->cur_ctx_seqno = submit->queue->ctx->seqno;
|
H A D | msm_gpu.h | 199 * The ctx->seqno value of the last context to submit rendering, 201 * that support per-context pgtables). Tracked by seqno rather 361 * @seqno: unique per process seqno 369 int seqno; member in struct:msm_file_private 498 * seqno, protected by submitqueue lock 534 u32 seqno; member in struct:msm_gpu_state::__anon746
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_sched_job.c | 149 xe_assert(job_to_xe(job), cf->base.seqno == fences[j]->seqno); 246 * Can safely check just LRC[0] seqno as that is last seqno written when
|
/linux-master/drivers/net/wireless/ralink/rt2x00/ |
H A D | rt2x00queue.c | 183 u16 seqno; local 193 * seqno on retransmitted data (non-QOS) and management frames. 194 * To workaround the problem let's generate seqno in software. 196 * hence hardware has to assign seqno for them. 219 seqno = atomic_add_return(0x10, &intf->seqno); 221 seqno = atomic_read(&intf->seqno); 224 hdr->seq_ctrl |= cpu_to_le16(seqno);
|
/linux-master/net/batman-adv/ |
H A D | routing.c | 1174 u32 seqno; local 1211 seqno = ntohl(bcast_packet->seqno); 1214 seqno)) 1217 seq_diff = seqno - orig_node->last_bcast_seqno; 1229 orig_node->last_bcast_seqno = seqno;
|
/linux-master/drivers/net/wireless/ath/ath9k/ |
H A D | xmit.c | 154 static void ath_send_bar(struct ath_atx_tid *tid, u16 seqno) argument 160 seqno << IEEE80211_SEQ_SEQ_SHIFT); 322 u16 seqno = bf->bf_state.seqno; local 328 index = ATH_BA_INDEX(tid->seq_start, seqno); 345 u16 seqno = bf->bf_state.seqno; local 351 index = ATH_BA_INDEX(tid->seq_start, seqno); 476 ba_index = ATH_BA_INDEX(seq_st, bf->bf_state.seqno); 573 u16 seqno local 949 u16 seqno; local 2227 u16 seqno; local [all...] |
/linux-master/drivers/net/ethernet/sfc/ |
H A D | mcdi.c | 156 u32 xflags, seqno; local 162 ++mcdi->seqno; 163 seqno = mcdi->seqno & SEQ_MASK; 177 MCDI_HEADER_SEQ, seqno, 189 MCDI_HEADER_SEQ, seqno, 323 } else if ((respseq ^ mcdi->seqno) & SEQ_MASK) { 326 respseq, mcdi->seqno); 514 * the seqno check in efx_mcdi_ev_cpl() will fail 516 ++mcdi->seqno; 562 efx_mcdi_ev_cpl(struct efx_nic *efx, unsigned int seqno, unsigned int datalen, unsigned int mcdi_err) argument [all...] |
/linux-master/drivers/net/wireless/ath/ath6kl/ |
H A D | htc_pipe.c | 166 packet->info.tx.seqno = ep->seqno; 167 ep->seqno++; 192 packet->info.tx.seqno = ep->seqno; 195 ep->seqno++; 244 htc_hdr->ctrl[1] = (u8) packet->info.tx.seqno;
|
/linux-master/net/tipc/ |
H A D | msg.c | 820 * @seqno: sequence number of buffer to add 823 bool __tipc_skb_queue_sorted(struct sk_buff_head *list, u16 seqno, argument 828 if (skb_queue_empty(list) || less(seqno, buf_seqno(skb_peek(list)))) { 833 if (more(seqno, buf_seqno(skb_peek_tail(list)))) { 839 if (more(seqno, buf_seqno(_skb))) 841 if (seqno == buf_seqno(_skb))
|
H A D | crypto.c | 147 * @seqno: the key seqno (cluster scope) 164 atomic64_t seqno ____cacheline_aligned; 606 atomic64_set(&tmp->seqno, 0); 660 atomic64_set(&aead->seqno, 0); 798 memcpy(iv + 4, (u8 *)&ehdr->seqno, 8); 923 memcpy(iv + 4, (u8 *)&ehdr->seqno, 8); 1037 u64 seqno; local 1045 /* Obtain a seqno first: 1046 * Use the key seqno ( [all...] |
/linux-master/drivers/gpu/drm/msm/adreno/ |
H A D | a3xx_gpu.c | 43 if (gpu->cur_ctx_seqno == submit->queue->ctx->seqno) 56 OUT_RING(ring, submit->seqno); 73 OUT_RING(ring, submit->seqno);
|
/linux-master/drivers/net/ethernet/google/gve/ |
H A D | gve_rx.c | 66 rx->desc.seqno = 1; 942 return (GVE_SEQNO(flags_seq) == rx->desc.seqno); 1013 while ((GVE_SEQNO(desc->flags_seq) == rx->desc.seqno) && 1023 rx->desc.seqno = gve_next_seqno(rx->desc.seqno); 1034 GVE_SEQNO(desc->flags_seq), rx->desc.seqno);
|
/linux-master/drivers/gpu/drm/virtio/ |
H A D | virtgpu_vq.c | 218 trace_virtio_gpu_cmd_response(vgdev->ctrlq.vq, resp, entry->seqno); 267 trace_virtio_gpu_cmd_response(vgdev->cursorq.vq, resp, entry->seqno); 360 vbuf->seqno = ++vgdev->ctrlq.seqno; 361 trace_virtio_gpu_cmd_queue(vq, virtio_gpu_vbuf_ctrl_hdr(vbuf), vbuf->seqno); 473 vbuf->seqno = ++vgdev->cursorq.seqno; 476 vbuf->seqno);
|
H A D | virtgpu_drv.h | 172 uint32_t seqno; member in struct:virtio_gpu_vbuffer 202 uint32_t seqno; member in struct:virtio_gpu_queue
|
/linux-master/drivers/gpu/drm/nouveau/ |
H A D | nouveau_fence.c | 140 if ((int)(seq - fence->base.seqno) < 0) 482 ret = (int)(fctx->read(chan) - fence->base.seqno) >= 0;
|
/linux-master/drivers/gpu/drm/qxl/ |
H A D | qxl_release.c | 105 release->base.seqno = ++qdev->release_seqno; 426 release->id | 0xf0000000, release->base.seqno);
|
/linux-master/drivers/gpio/ |
H A D | gpiolib-cdev.c | 424 * @req_seqno: the seqno for the current edge event in the sequence of 426 * @line_seqno: the seqno for the current edge event in the sequence of 533 * @seqno: the sequence number for edge events generated on all lines in 549 atomic_t seqno; member in struct:linereq 792 le.seqno = (lr->num_lines == 1) ? le.line_seqno : line->req_seqno; 827 &lr->seqno); 890 line->req_seqno = atomic_inc_return(&lr->seqno); 909 le.seqno = (lr->num_lines == 1) ? le.line_seqno : line->req_seqno; 929 line->req_seqno = atomic_inc_return(&lr->seqno); 1017 le.seqno [all...] |
/linux-master/drivers/media/pci/tw5864/ |
H A D | tw5864-core.c | 143 cur_frame->seqno = input->frame_seqno;
|
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_gt_requests.c | 257 f->seqno);
|
/linux-master/fs/xfs/libxfs/ |
H A D | xfs_alloc.h | 276 xfs_failaddr_t xfs_validate_ag_length(struct xfs_buf *bp, uint32_t seqno,
|
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_device.h | 264 /** @kccb.fence_ctx.seqno: Sequence number incremented 266 atomic_t seqno; member in struct:pvr_device::__anon696::__anon697
|
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_dcb.h | 124 __be32 seqno; member in struct:i40e_cee_ctrl_tlv
|
/linux-master/net/smc/ |
H A D | smc.h | 134 u16 seqno; /* connection seq # */ member in struct:smc_host_cdc_msg
|
/linux-master/include/uapi/drm/ |
H A D | vmwgfx_drm.h | 308 * fence seqno that when signaled, indicates that the command buffer has 352 * @seqno: Fence sequence number in fifo. A fence object with a lower 353 * seqno will signal the EXEC flag before a fence object with a higher 354 * seqno. This can be used by user-space to avoid kernel calls to determine 355 * whether a fence has signaled the EXEC flag. Note that @seqno will 357 * @passed_seqno: The highest seqno number processed by the hardware 359 * to determine whether a fence seqno might be stale. 376 __u32 seqno; member in struct:drm_vmw_fence_rep
|