Searched refs:last (Results 1 - 25 of 1032) sorted by last modified time

1234567891011>>

/linux-master/sound/pci/emu10k1/
H A Demuproc.c444 voice->last,
H A Demumixer.c1521 if (!mix->epcm->voices[0]->last) {
1588 if (!mix->epcm->voices[0]->last) {
1653 if (!mix->epcm->voices[0]->last) {
/linux-master/net/ipv4/
H A Dudp_offload.c347 /* last packet can be partial gso_size, account for that in checksum */
441 if (NAPI_GRO_CB(p)->last == p)
444 NAPI_GRO_CB(p)->last->next = skb;
448 NAPI_GRO_CB(p)->last = skb;
H A Dudp.c46 * Alan Cox : Cache last socket
249 unsigned short first, last; local
262 last = first + udptable->mask + 1;
285 } while (++first != last);
1701 struct sk_buff *last; local
1718 err, &last);
1740 err, &last);
3182 /* Resume from the last iterated socket at the
/linux-master/net/core/
H A Dgro.c131 lp = NAPI_GRO_CB(p)->last;
211 if (NAPI_GRO_CB(p)->last == p)
214 NAPI_GRO_CB(p)->last->next = skb;
215 NAPI_GRO_CB(p)->last = skb;
526 /* Must be called before setting NAPI_GRO_CB(skb)->{age|last} */
529 NAPI_GRO_CB(skb)->last = skb;
H A Dfilter.c2929 u64 last = start + len; local
2947 if (start >= offset + l || last >= msg->sg.size)
/linux-master/include/sound/
H A Demu10k1.h229 /* I2S0 can phase track the last S/PDIF input */
1480 unsigned char last; member in struct:snd_emu10k1_voice
/linux-master/include/net/
H A Dgro.h26 struct sk_buff *last; member in struct:napi_gro_cb::__anon89::__anon91
/linux-master/drivers/spi/
H A Dspi.c1028 bool last = false; local
1033 last = true;
1036 return last;
1047 * isn't actually changing from the last time this was called.
3494 * Release the last reference on the controller if its driver
3671 /* Clear cs_change and delay for all but the last */
3740 * We set up xferp to the last entry we have inserted,
3777 * but note that xfer is advanced to the last transfer inserted
3818 * but note that xfer is advanced to the last transfer inserted
4217 /* Don't change cs_change on the last entr
[all...]
/linux-master/drivers/nvme/host/
H A Dtcp.c376 bool sync, bool last)
395 if (last && nvme_tcp_queue_has_pending(queue))
634 "queue %d tag %#x SUCCESS set but not last PDU\n",
1045 bool last = nvme_tcp_pdu_last_send(req, len); local
1049 if (last && !queue->data_digest && !nvme_tcp_queue_more(queue))
1068 * update the request iterator except for the last payload send
1075 /* fully successful last send in current PDU */
1076 if (last && ret == len) {
1905 int first, int last)
1909 for (i = first; i < last;
375 nvme_tcp_queue_request(struct nvme_tcp_request *req, bool sync, bool last) argument
1904 nvme_tcp_start_io_queues(struct nvme_ctrl *ctrl, int first, int last) argument
[all...]
H A Dpci.c898 nvme_write_sq_db(nvmeq, bd->last);
1462 * nvme_poll(). This is the last chance for the driver to see a natural
/linux-master/drivers/gpu/drm/xe/
H A Dxe_vm.c51 * Check if the userptr vma has been invalidated since last successful
2381 struct list_head *ops_list, bool last)
2521 XE_WARN_ON(!last_op && last && !list_empty(ops_list));
2527 if (last) {
2706 bool last = op->flags & XE_VMA_OP_LAST; local
2708 if (last) {
2719 if (last)
2378 vm_bind_ioctl_ops_parse(struct xe_vm *vm, struct xe_exec_queue *q, struct drm_gpuva_ops *ops, struct xe_sync_entry *syncs, u32 num_syncs, struct list_head *ops_list, bool last) argument
/linux-master/drivers/gpu/drm/xe/compat-i915-headers/
H A Di915_drv.h103 #define IS_GRAPHICS_VER(xe, first, last) \
105 (xe)->info.graphics_verx100 <= (last*100 + 99))
112 #define IS_DISPLAY_STEP(xe, first, last) ({u8 __step = (xe)->info.step.display; first <= __step && __step <= last; })
113 #define IS_GRAPHICS_STEP(xe, first, last) ({u8 __step = (xe)->info.step.graphics; first <= __step && __step <= last; })
130 #define IS_DG1_GRAPHICS_STEP(xe, first, last) (IS_DG1(xe) && IS_GRAPHICS_STEP(xe, first, last))
131 #define IS_DG2_GRAPHICS_STEP(xe, variant, first, last) \
133 IS_GRAPHICS_STEP(xe, first, last))
[all...]
/linux-master/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_object.h69 uint64_t last; member in struct:amdgpu_bo_va_mapping
/linux-master/drivers/firewire/
H A Dohci.c103 struct descriptor *last);
142 struct descriptor *last; member in struct:context
145 * The last descriptor block in the DMA program. It contains the branch
761 * We search for the buffer that contains the last AR packet DMA data written
767 unsigned int i, next_i, last = ctx->last_buffer_index; local
773 /* A buffer that is not yet completely filled must be the last one. */
774 while (i != last && res_count == 0) {
792 if (MAX_AR_PACKET_SIZE > PAGE_SIZE && i != last) {
989 * all packets up to the buffer end here. If the last packet
1099 struct descriptor *d, *last; local
1337 struct descriptor *d, *last; local
1472 handle_at_packet(struct context *context, struct descriptor *d, struct descriptor *last) argument
2825 handle_ir_packet_per_buffer(struct context *context, struct descriptor *d, struct descriptor *last) argument
2860 handle_ir_buffer_fill(struct context *context, struct descriptor *d, struct descriptor *last) argument
2948 handle_it_packet(struct context *context, struct descriptor *d, struct descriptor *last) argument
3261 struct descriptor *d, *last, *pd; local
[all...]
/linux-master/net/sunrpc/
H A Dxprtsock.c1804 unsigned short last; local
1834 last = port;
1836 if (port > last)
/linux-master/kernel/sched/
H A Dfair.c952 struct rb_node *last = rb_last(&cfs_rq->tasks_timeline.rb_root); local
954 if (!last)
957 return __node_2_se(last);
2686 * Get the fraction of time the task has been running since the last
2852 /* Decay existing window, copy faults since last scan */
3425 * It is possible to reach the end of the VMA list but the last few
4099 * differential update where we store the last value we propagated. This in
4617 * util_avg. Given that period_contrib might have moved since the last
4927 * already ~1% close to its last activation value.
5397 /* return excess runtime on last dequeu
[all...]
/linux-master/kernel/
H A Dcpu.c59 * @last: For multi-instance rollback, remember how far we got
77 struct hlist_node *last; member in struct:cpuhp_cpu_state
705 st->last = NULL;
734 * If we have st->last we need to undo partial multi_instance of this
737 if (!st->last) {
1084 st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last);
1092 st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last);
1138 st->last = NULL;
1150 if ((ret = st->result) && st->last) {
1161 st->node = st->last
[all...]
/linux-master/drivers/dma/
H A Dpl330.c290 * The PL330 core does not modify this and uses the last
380 /* Index of the last enqueued request */
382 /* Index of the last submitted request or -1 if the DMA is stopped */
538 bool last; member in struct:dma_pl330_desc
2086 /* Try to submit a req imm. next to the last completed cookie */
2468 if (desc->last)
2503 * We returned the last one of the circular list of descriptor(s)
2504 * from prep_xxx, so the argument to submit corresponds to the last
2509 struct dma_pl330_desc *desc, *last = to_desc(tx); local
2517 while (!list_empty(&last
[all...]
/linux-master/net/wireless/
H A Dnl80211.c50 NL80211_MCGRP_TESTMODE /* keep last - ifdef! */
3670 * Fragments (apart from the last one) are required to
7561 /* must be last in here for error handling */
9481 * All scan plans but the last one must specify
9491 * The last scan plan must not specify the number of
12817 s32 last, low, high; local
12846 last = cqm_config->last_rssi_event_value;
12852 if (last < cqm_config->rssi_thresholds[i])
/linux-master/net/tls/
H A Dtls_strp.c283 struct sk_buff *nskb, *first, *last; local
304 last = first->prev;
305 last->next = nskb;
/linux-master/net/ax25/
H A Daf_ax25.c1665 struct sk_buff *skb, *last; local
1684 skb = __skb_try_recv_datagram(sk, sk_queue, flags, &off, &err, &last);
1692 &timeo, last)) {
1694 &err, &last);
/linux-master/include/linux/
H A Dmm.h1008 return vmi->mas.last + 1;
1037 vmi->mas.last = vma->vm_end - 1;
1500 * last reference.
1520 * the last references.
3249 unsigned long start, unsigned long last);
3251 unsigned long start, unsigned long last);
3253 #define vma_interval_tree_foreach(vma, root, start, last) \
3254 for (vma = vma_interval_tree_iter_first(root, start, last); \
3255 vma; vma = vma_interval_tree_iter_next(vma, start, last))
3263 unsigned long start, unsigned long last);
[all...]
/linux-master/fs/netfs/
H A Dbuffered_write.c569 pgoff_t last = (start + len - 1) / PAGE_SIZE, next; local
574 _debug("kill %lx (to %lx)", index, last);
594 } while (index = next, index <= last);
607 pgoff_t last = (start + len - 1) / PAGE_SIZE, next; local
627 } while (index = next, index <= last);
643 pgoff_t last; local
652 last = (wreq->start + wreq->len - 1) / PAGE_SIZE;
653 xas_for_each(&xas, folio, last) {
656 wreq->len, wreq->start, folio->index, last);
/linux-master/drivers/net/ethernet/broadcom/bnxt/
H A Dbnxt.c777 int j, last; local
799 last = tx_buf->nr_frags;
801 for (j = 0; j < last; j++) {
1289 u16 last; local
1293 last = RING_CMP(*raw_cons);
1295 &cpr->cp_desc_ring[CP_RING(last)][CP_IDX(last)];
3219 int k, last; local
3253 last = tx_buf->nr_frags;
3255 for (k = 0; k < last;
8452 bnxt_hwrm_func_backing_store_cfg_v2(struct bnxt *bp, struct bnxt_ctx_mem_type *ctxm, bool last) argument
[all...]

Completed in 558 milliseconds

1234567891011>>