/linux-master/fs/smb/server/ |
H A D | smb2pdu.c | 1883 * failure to make it harder to send enough random connection requests 7027 command_list = &conn->requests;
|
H A D | connection.c | 77 INIT_LIST_HEAD(&conn->requests); 118 requests_queue = &conn->requests; 265 * ksmbd_conn_handler_loop() - session thread to listen on new smb requests
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | intel_guc_submission.c | 72 * requests of that context until the G2H is returned. Because this interaction 84 * fence is used to stall all requests associated with this guc_id until the 116 * schedule disable is in flight, etc... Also protects list of inflight requests 708 * Corner case where requests were sitting in the priority list or a 921 * submitting all the requests generated in parallel. 970 * We need to coalesce all multi-lrc requests in 972 * guaranteed that all of these requests will be 1693 flush_work(&guc->ct.requests.worker); 1736 * All pending requests for this context will be zapped, and any 1780 &ce->guc_state.requests, [all...] |
/linux-master/drivers/gpu/drm/i915/gt/ |
H A D | intel_gt.c | 529 struct i915_request *requests[I915_NUM_ENGINES] = {}; local 576 requests[id] = i915_request_get(rq); 593 for (id = 0; id < ARRAY_SIZE(requests); id++) { 597 rq = requests[id]; 628 for (id = 0; id < ARRAY_SIZE(requests); id++) { 632 rq = requests[id]; 678 /* If the device is asleep, we have no requests outstanding */ 828 * all in-flight requests so that we can quickly unbind the active
|
H A D | intel_engine_cs.c | 712 /* Free the requests! dma-resv keeps fences around for an eternity */ 1570 GEM_BUG_ON(!list_empty(&engine->sched_engine->requests)); 1782 * pending MI_FORCE_WAKEUP requests that the CS has initiated to complete. The 1919 * Return true if there are no requests pending, nothing left to be submitted 1935 /* ELSP is empty, but there are ready requests? E.g. after reset */ 2345 void intel_engine_dump_active_requests(struct list_head *requests, argument 2353 list_for_each_entry(rq, requests, sched.link) { 2378 * The GPU is still running so requests are still executing and any 2520 list_for_each_entry_from_reverse(request, &tl->requests, link) { 2531 list_for_each_entry(request, &engine->sched_engine->requests, [all...] |
H A D | intel_execlists_submission.c | 90 * After processing, if any requests were retired and the queue is not empty 91 * then a new execution list can be submitted. The two requests at the front of 93 * an execution list, if subsequent requests have the same ID as the first then 94 * the two requests must be combined. This is done simply by discarding requests 95 * at the head of the queue until either only one requests is left (in which case 96 * we use a NULL second context) or the first two requests have unique IDs. 98 * By always executing the first two requests in the queue the driver ensures 218 list_for_each_entry_from_reverse(rq, &tl->requests, link) { 270 * will not matter (i.e. all requests t [all...] |
H A D | selftest_context.c | 54 if (list_empty(&tl->requests)) 57 rq = list_last_entry(&tl->requests, typeof(*rq), link); 212 * with no more outstanding requests).
|
/linux-master/drivers/usb/dwc2/ |
H A D | hcd.c | 754 * register to flush requests for this channel. In the latter case, sets a 762 * Some time may elapse before the core flushes any posted requests for this 1385 chan->requests++; 1473 chan->requests++; 1493 * Return: 1 if a new request is queued, 0 if no more requests are required 1519 * handler to requeue a NAK'd request. Instead, IN requests 1521 * transfer completes, the extra requests for the channel will 1533 chan->requests++; 1551 chan->requests++; 1672 /* Flush out any channel requests i [all...] |
/linux-master/include/linux/ |
H A D | kvm_host.h | 168 * Architecture-independent vcpu->requests bit members 188 BUILD_BUG_ON((unsigned)(nr) >= (sizeof_field(struct kvm_vcpu, requests) * 8) - KVM_REQUEST_ARCH_BASE); \ 340 u64 requests; member in struct:kvm_vcpu 554 * The memory barrier ensures a previous write to vcpu->requests cannot 2151 set_bit(req & KVM_REQUEST_MASK, (void *)&vcpu->requests); 2158 * vcpu->requests. The vCPU won't clear the request, so it will stay 2169 return READ_ONCE(vcpu->requests); 2174 return test_bit(req & KVM_REQUEST_MASK, (void *)&vcpu->requests); 2179 clear_bit(req & KVM_REQUEST_MASK, (void *)&vcpu->requests);
|
/linux-master/drivers/perf/ |
H A D | qcom_l2_pmu.c | 693 L2CACHE_EVENT_ATTR(total-requests, L2_EVENT_TOTAL_REQUESTS),
|
/linux-master/drivers/md/dm-vdo/indexer/ |
H A D | indexer.h | 33 * There are five types of requests: 143 /* The number of threads used to process index requests */ 151 * requests processed since the index was opened. 185 /* The number of delete requests that deleted an existing entry */ 187 /* The number of delete requests that did nothing */ 193 /* The total number of requests processed */ 194 u64 requests; member in struct:uds_index_stats 214 /* Zone message requests are used to communicate between index zones. */ 278 /* If true, continue this request before processing newer requests */
|
H A D | index-session.c | 23 * requests currently in progress so that they can be drained when suspending or closing the index. 38 /* The session has loaded an index, which can handle requests. */ 46 /* The session's index is closing and draining requests. */ 48 /* The session is being destroyed and is draining requests. */ 162 count_once(&session_stats->requests); 413 * Suspend index operation, draining all current index requests and preventing new index requests 706 stats->requests = READ_ONCE(session_stats->requests);
|
H A D | index-session.h | 23 * will not accept any further requests and can only be closed. Closing the index will clear the 28 /* Post requests that found an entry */ 30 /* Post requests found in the open chapter */ 32 /* Post requests found in the dense index */ 34 /* Post requests found in the sparse index */ 36 /* Post requests that did not find an entry */ 38 /* Update requests that found an entry */ 40 /* Update requests that did not find an entry */ 42 /* Delete requests that found an entry */ 44 /* Delete requests tha 51 u64 requests; variable [all...] |
/linux-master/drivers/iommu/ |
H A D | virtio-iommu.c | 43 struct list_head requests; member in struct:viommu_dev 149 * __viommu_sync_req - Complete all in-flight requests 151 * Wait for all added requests to complete. When this function returns, all 152 * requests that were in-flight at the time of the call have completed. 165 while (!list_empty(&viommu->requests)) { 195 dev_dbg(viommu->dev, "could not sync requests (%d)\n", ret); 208 * Otherwise don't kick the queue nor wait for requests to complete. 256 list_add_tail(&req->list, &viommu->requests); 298 dev_dbg(viommu->dev, "could not sync requests (%d)\n", ret); 427 * viommu_replay_mappings - re-send MAP requests [all...] |
/linux-master/drivers/gpu/drm/i915/ |
H A D | i915_request.c | 379 &i915_request_timeline(rq)->requests)); 395 * We only loosely track inflight requests across preemption, 427 tmp = list_first_entry(&tl->requests, typeof(*tmp), link); 455 * requests, we know that only the currently executing request 616 * requests that we have unsubmitted from HW, but left running 828 list_for_each_entry_safe(rq, rn, &tl->requests, link) 849 if (list_empty(&tl->requests)) 853 rq = list_first_entry(&tl->requests, typeof(*rq), link); 862 rq = list_last_entry(&tl->requests, typeof(*rq), link); 865 /* Retire our old requests i [all...] |
H A D | i915_gpu_error.c | 1757 error_uc->guc.last_fence = uc->guc.ct.requests.last_fence;
|
/linux-master/drivers/gpu/drm/i915/gem/selftests/ |
H A D | i915_gem_mman.c | 634 cancel_delayed_work_sync(&to_gt(i915)->requests.retire_work);
|
/linux-master/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_execbuffer.c | 260 /** our requests to build */ 261 struct i915_request *requests[MAX_ENGINE_INSTANCE + 1]; member in struct:i915_execbuffer 1974 * Using two helper loops for the order of which requests / batches are created 1994 if (eb->requests[i]) 1995 return eb->requests[i]; 2044 struct i915_request *rq = eb->requests[j]; 2148 if (!eb->requests[j]) 2151 err = _i915_vma_move_to_active(vma, eb->requests[j], 2155 &eb->requests[j]->fence, 2188 if (!eb->requests[ [all...] |
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | cwsr_trap_handler_gfx10.asm | 1225 // Otherwise retain PRIV=1 for subsequent context save requests.
|
/linux-master/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_dcb_nl.c | 91 pfc->requests[i] = pf->stats.priority_xoff_tx[i];
|
/linux-master/drivers/net/ethernet/hisilicon/hns3/hns3pf/ |
H A D | hclge_dcb.c | 320 hclge_pfc_tx_stats_get(hdev, pfc->requests);
|
/linux-master/drivers/mailbox/ |
H A D | bcm-flexrm-mailbox.c | 266 struct brcm_message *requests[RING_MAX_REQ_COUNT]; member in struct:flexrm_ring 991 /* If no requests possible then save data pointer and goto done. */ 998 ring->requests[reqid] = msg; 1003 ring->requests[reqid] = NULL; 1065 ring->requests[reqid] = NULL; 1123 msg = ring->requests[reqid]; 1132 ring->requests[reqid] = NULL; 1396 /* Abort all in-flight requests */ 1398 msg = ring->requests[reqid]; 1403 ring->requests[reqi [all...] |
/linux-master/drivers/block/ |
H A D | xen-blkfront.c | 112 * Id of the sibling if we ever need 2 requests when handling a 133 * Maximum number of segments in indirect requests, the actual value used by 141 "Maximum amount of segments in indirect requests (default is 32)"); 231 struct list_head requests; member in struct:blkfront_info 651 * previous requests. This is OK as long as 734 /* Check if we have enough persistent grants to allocate a requests */ 815 /* Link the 2 requests together */ 1244 * Clear persistent grants present in requests already 1304 /* Prevent new requests being issued until we fix things up. */ 2032 /* Kick any other new requests queue [all...] |
/linux-master/block/ |
H A D | bfq-iosched.c | 37 * processes issuing sequential requests (to boost the throughput), 49 * the I/O requests in a bfq_queue come from an interactive or a soft 164 /* Expiration time of async (0) and sync (1) requests, in ns. */ 179 /* Default maximum budget values, in sectors and number of requests. */ 215 * first requests from each cooperator. After that, there is very 225 /* hw_tag detection: parallel requests threshold and min samples needed. */ 471 * Scheduler run of queue, if there are requests pending and no one in the 502 unsigned int wrap = 0; /* bit mask: requests behind the disk head? */ 642 * gets more requests than high prio queue from lower 656 "too many requests [all...] |
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt_dcb.c | 624 pfc->requests[i] = le64_to_cpu(*(stats + tx_off));
|