Searched refs:requests (Results 1 - 25 of 92) sorted by last modified time

1234

/linux-master/fs/smb/server/
H A Dsmb2pdu.c1883 * failure to make it harder to send enough random connection requests
7027 command_list = &conn->requests;
H A Dconnection.c77 INIT_LIST_HEAD(&conn->requests);
118 requests_queue = &conn->requests;
265 * ksmbd_conn_handler_loop() - session thread to listen on new smb requests
/linux-master/drivers/gpu/drm/i915/gt/uc/
H A Dintel_guc_submission.c72 * requests of that context until the G2H is returned. Because this interaction
84 * fence is used to stall all requests associated with this guc_id until the
116 * schedule disable is in flight, etc... Also protects list of inflight requests
708 * Corner case where requests were sitting in the priority list or a
921 * submitting all the requests generated in parallel.
970 * We need to coalesce all multi-lrc requests in
972 * guaranteed that all of these requests will be
1693 flush_work(&guc->ct.requests.worker);
1736 * All pending requests for this context will be zapped, and any
1780 &ce->guc_state.requests,
[all...]
/linux-master/drivers/gpu/drm/i915/gt/
H A Dintel_gt.c529 struct i915_request *requests[I915_NUM_ENGINES] = {}; local
576 requests[id] = i915_request_get(rq);
593 for (id = 0; id < ARRAY_SIZE(requests); id++) {
597 rq = requests[id];
628 for (id = 0; id < ARRAY_SIZE(requests); id++) {
632 rq = requests[id];
678 /* If the device is asleep, we have no requests outstanding */
828 * all in-flight requests so that we can quickly unbind the active
H A Dintel_engine_cs.c712 /* Free the requests! dma-resv keeps fences around for an eternity */
1570 GEM_BUG_ON(!list_empty(&engine->sched_engine->requests));
1782 * pending MI_FORCE_WAKEUP requests that the CS has initiated to complete. The
1919 * Return true if there are no requests pending, nothing left to be submitted
1935 /* ELSP is empty, but there are ready requests? E.g. after reset */
2345 void intel_engine_dump_active_requests(struct list_head *requests, argument
2353 list_for_each_entry(rq, requests, sched.link) {
2378 * The GPU is still running so requests are still executing and any
2520 list_for_each_entry_from_reverse(request, &tl->requests, link) {
2531 list_for_each_entry(request, &engine->sched_engine->requests,
[all...]
H A Dintel_execlists_submission.c90 * After processing, if any requests were retired and the queue is not empty
91 * then a new execution list can be submitted. The two requests at the front of
93 * an execution list, if subsequent requests have the same ID as the first then
94 * the two requests must be combined. This is done simply by discarding requests
95 * at the head of the queue until either only one requests is left (in which case
96 * we use a NULL second context) or the first two requests have unique IDs.
98 * By always executing the first two requests in the queue the driver ensures
218 list_for_each_entry_from_reverse(rq, &tl->requests, link) {
270 * will not matter (i.e. all requests t
[all...]
H A Dselftest_context.c54 if (list_empty(&tl->requests))
57 rq = list_last_entry(&tl->requests, typeof(*rq), link);
212 * with no more outstanding requests).
/linux-master/drivers/usb/dwc2/
H A Dhcd.c754 * register to flush requests for this channel. In the latter case, sets a
762 * Some time may elapse before the core flushes any posted requests for this
1385 chan->requests++;
1473 chan->requests++;
1493 * Return: 1 if a new request is queued, 0 if no more requests are required
1519 * handler to requeue a NAK'd request. Instead, IN requests
1521 * transfer completes, the extra requests for the channel will
1533 chan->requests++;
1551 chan->requests++;
1672 /* Flush out any channel requests i
[all...]
/linux-master/include/linux/
H A Dkvm_host.h168 * Architecture-independent vcpu->requests bit members
188 BUILD_BUG_ON((unsigned)(nr) >= (sizeof_field(struct kvm_vcpu, requests) * 8) - KVM_REQUEST_ARCH_BASE); \
340 u64 requests; member in struct:kvm_vcpu
554 * The memory barrier ensures a previous write to vcpu->requests cannot
2151 set_bit(req & KVM_REQUEST_MASK, (void *)&vcpu->requests);
2158 * vcpu->requests. The vCPU won't clear the request, so it will stay
2169 return READ_ONCE(vcpu->requests);
2174 return test_bit(req & KVM_REQUEST_MASK, (void *)&vcpu->requests);
2179 clear_bit(req & KVM_REQUEST_MASK, (void *)&vcpu->requests);
/linux-master/drivers/perf/
H A Dqcom_l2_pmu.c693 L2CACHE_EVENT_ATTR(total-requests, L2_EVENT_TOTAL_REQUESTS),
/linux-master/drivers/md/dm-vdo/indexer/
H A Dindexer.h33 * There are five types of requests:
143 /* The number of threads used to process index requests */
151 * requests processed since the index was opened.
185 /* The number of delete requests that deleted an existing entry */
187 /* The number of delete requests that did nothing */
193 /* The total number of requests processed */
194 u64 requests; member in struct:uds_index_stats
214 /* Zone message requests are used to communicate between index zones. */
278 /* If true, continue this request before processing newer requests */
H A Dindex-session.c23 * requests currently in progress so that they can be drained when suspending or closing the index.
38 /* The session has loaded an index, which can handle requests. */
46 /* The session's index is closing and draining requests. */
48 /* The session is being destroyed and is draining requests. */
162 count_once(&session_stats->requests);
413 * Suspend index operation, draining all current index requests and preventing new index requests
706 stats->requests = READ_ONCE(session_stats->requests);
H A Dindex-session.h23 * will not accept any further requests and can only be closed. Closing the index will clear the
28 /* Post requests that found an entry */
30 /* Post requests found in the open chapter */
32 /* Post requests found in the dense index */
34 /* Post requests found in the sparse index */
36 /* Post requests that did not find an entry */
38 /* Update requests that found an entry */
40 /* Update requests that did not find an entry */
42 /* Delete requests that found an entry */
44 /* Delete requests tha
51 u64 requests; variable
[all...]
/linux-master/drivers/iommu/
H A Dvirtio-iommu.c43 struct list_head requests; member in struct:viommu_dev
149 * __viommu_sync_req - Complete all in-flight requests
151 * Wait for all added requests to complete. When this function returns, all
152 * requests that were in-flight at the time of the call have completed.
165 while (!list_empty(&viommu->requests)) {
195 dev_dbg(viommu->dev, "could not sync requests (%d)\n", ret);
208 * Otherwise don't kick the queue nor wait for requests to complete.
256 list_add_tail(&req->list, &viommu->requests);
298 dev_dbg(viommu->dev, "could not sync requests (%d)\n", ret);
427 * viommu_replay_mappings - re-send MAP requests
[all...]
/linux-master/drivers/gpu/drm/i915/
H A Di915_request.c379 &i915_request_timeline(rq)->requests));
395 * We only loosely track inflight requests across preemption,
427 tmp = list_first_entry(&tl->requests, typeof(*tmp), link);
455 * requests, we know that only the currently executing request
616 * requests that we have unsubmitted from HW, but left running
828 list_for_each_entry_safe(rq, rn, &tl->requests, link)
849 if (list_empty(&tl->requests))
853 rq = list_first_entry(&tl->requests, typeof(*rq), link);
862 rq = list_last_entry(&tl->requests, typeof(*rq), link);
865 /* Retire our old requests i
[all...]
H A Di915_gpu_error.c1757 error_uc->guc.last_fence = uc->guc.ct.requests.last_fence;
/linux-master/drivers/gpu/drm/i915/gem/selftests/
H A Di915_gem_mman.c634 cancel_delayed_work_sync(&to_gt(i915)->requests.retire_work);
/linux-master/drivers/gpu/drm/i915/gem/
H A Di915_gem_execbuffer.c260 /** our requests to build */
261 struct i915_request *requests[MAX_ENGINE_INSTANCE + 1]; member in struct:i915_execbuffer
1974 * Using two helper loops for the order of which requests / batches are created
1994 if (eb->requests[i])
1995 return eb->requests[i];
2044 struct i915_request *rq = eb->requests[j];
2148 if (!eb->requests[j])
2151 err = _i915_vma_move_to_active(vma, eb->requests[j],
2155 &eb->requests[j]->fence,
2188 if (!eb->requests[
[all...]
/linux-master/drivers/gpu/drm/amd/amdkfd/
H A Dcwsr_trap_handler_gfx10.asm1225 // Otherwise retain PRIV=1 for subsequent context save requests.
/linux-master/drivers/net/ethernet/intel/i40e/
H A Di40e_dcb_nl.c91 pfc->requests[i] = pf->stats.priority_xoff_tx[i];
/linux-master/drivers/net/ethernet/hisilicon/hns3/hns3pf/
H A Dhclge_dcb.c320 hclge_pfc_tx_stats_get(hdev, pfc->requests);
/linux-master/drivers/mailbox/
H A Dbcm-flexrm-mailbox.c266 struct brcm_message *requests[RING_MAX_REQ_COUNT]; member in struct:flexrm_ring
991 /* If no requests possible then save data pointer and goto done. */
998 ring->requests[reqid] = msg;
1003 ring->requests[reqid] = NULL;
1065 ring->requests[reqid] = NULL;
1123 msg = ring->requests[reqid];
1132 ring->requests[reqid] = NULL;
1396 /* Abort all in-flight requests */
1398 msg = ring->requests[reqid];
1403 ring->requests[reqi
[all...]
/linux-master/drivers/block/
H A Dxen-blkfront.c112 * Id of the sibling if we ever need 2 requests when handling a
133 * Maximum number of segments in indirect requests, the actual value used by
141 "Maximum amount of segments in indirect requests (default is 32)");
231 struct list_head requests; member in struct:blkfront_info
651 * previous requests. This is OK as long as
734 /* Check if we have enough persistent grants to allocate a requests */
815 /* Link the 2 requests together */
1244 * Clear persistent grants present in requests already
1304 /* Prevent new requests being issued until we fix things up. */
2032 /* Kick any other new requests queue
[all...]
/linux-master/block/
H A Dbfq-iosched.c37 * processes issuing sequential requests (to boost the throughput),
49 * the I/O requests in a bfq_queue come from an interactive or a soft
164 /* Expiration time of async (0) and sync (1) requests, in ns. */
179 /* Default maximum budget values, in sectors and number of requests. */
215 * first requests from each cooperator. After that, there is very
225 /* hw_tag detection: parallel requests threshold and min samples needed. */
471 * Scheduler run of queue, if there are requests pending and no one in the
502 unsigned int wrap = 0; /* bit mask: requests behind the disk head? */
642 * gets more requests than high prio queue from lower
656 "too many requests
[all...]
/linux-master/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt_dcb.c624 pfc->requests[i] = le64_to_cpu(*(stats + tx_off));

Completed in 518 milliseconds

1234