/linux-master/sound/pci/emu10k1/ |
H A D | emuproc.c | 444 voice->last,
|
H A D | emumixer.c | 1521 if (!mix->epcm->voices[0]->last) { 1588 if (!mix->epcm->voices[0]->last) { 1653 if (!mix->epcm->voices[0]->last) {
|
/linux-master/net/ipv4/ |
H A D | udp_offload.c | 347 /* last packet can be partial gso_size, account for that in checksum */ 441 if (NAPI_GRO_CB(p)->last == p) 444 NAPI_GRO_CB(p)->last->next = skb; 448 NAPI_GRO_CB(p)->last = skb;
|
H A D | udp.c | 46 * Alan Cox : Cache last socket 249 unsigned short first, last; local 262 last = first + udptable->mask + 1; 285 } while (++first != last); 1701 struct sk_buff *last; local 1718 err, &last); 1740 err, &last); 3182 /* Resume from the last iterated socket at the
|
/linux-master/net/core/ |
H A D | gro.c | 131 lp = NAPI_GRO_CB(p)->last; 211 if (NAPI_GRO_CB(p)->last == p) 214 NAPI_GRO_CB(p)->last->next = skb; 215 NAPI_GRO_CB(p)->last = skb; 526 /* Must be called before setting NAPI_GRO_CB(skb)->{age|last} */ 529 NAPI_GRO_CB(skb)->last = skb;
|
H A D | filter.c | 2929 u64 last = start + len; local 2947 if (start >= offset + l || last >= msg->sg.size)
|
/linux-master/include/sound/ |
H A D | emu10k1.h | 229 /* I2S0 can phase track the last S/PDIF input */ 1480 unsigned char last; member in struct:snd_emu10k1_voice
|
/linux-master/include/net/ |
H A D | gro.h | 26 struct sk_buff *last; member in struct:napi_gro_cb::__anon89::__anon91
|
/linux-master/drivers/spi/ |
H A D | spi.c | 1028 bool last = false; local 1033 last = true; 1036 return last; 1047 * isn't actually changing from the last time this was called. 3494 * Release the last reference on the controller if its driver 3671 /* Clear cs_change and delay for all but the last */ 3740 * We set up xferp to the last entry we have inserted, 3777 * but note that xfer is advanced to the last transfer inserted 3818 * but note that xfer is advanced to the last transfer inserted 4217 /* Don't change cs_change on the last entr [all...] |
/linux-master/drivers/nvme/host/ |
H A D | tcp.c | 376 bool sync, bool last) 395 if (last && nvme_tcp_queue_has_pending(queue)) 634 "queue %d tag %#x SUCCESS set but not last PDU\n", 1045 bool last = nvme_tcp_pdu_last_send(req, len); local 1049 if (last && !queue->data_digest && !nvme_tcp_queue_more(queue)) 1068 * update the request iterator except for the last payload send 1075 /* fully successful last send in current PDU */ 1076 if (last && ret == len) { 1905 int first, int last) 1909 for (i = first; i < last; 375 nvme_tcp_queue_request(struct nvme_tcp_request *req, bool sync, bool last) argument 1904 nvme_tcp_start_io_queues(struct nvme_ctrl *ctrl, int first, int last) argument [all...] |
H A D | pci.c | 898 nvme_write_sq_db(nvmeq, bd->last); 1462 * nvme_poll(). This is the last chance for the driver to see a natural
|
/linux-master/drivers/gpu/drm/xe/ |
H A D | xe_vm.c | 51 * Check if the userptr vma has been invalidated since last successful 2381 struct list_head *ops_list, bool last) 2521 XE_WARN_ON(!last_op && last && !list_empty(ops_list)); 2527 if (last) { 2706 bool last = op->flags & XE_VMA_OP_LAST; local 2708 if (last) { 2719 if (last) 2378 vm_bind_ioctl_ops_parse(struct xe_vm *vm, struct xe_exec_queue *q, struct drm_gpuva_ops *ops, struct xe_sync_entry *syncs, u32 num_syncs, struct list_head *ops_list, bool last) argument
|
/linux-master/drivers/gpu/drm/xe/compat-i915-headers/ |
H A D | i915_drv.h | 103 #define IS_GRAPHICS_VER(xe, first, last) \ 105 (xe)->info.graphics_verx100 <= (last*100 + 99)) 112 #define IS_DISPLAY_STEP(xe, first, last) ({u8 __step = (xe)->info.step.display; first <= __step && __step <= last; }) 113 #define IS_GRAPHICS_STEP(xe, first, last) ({u8 __step = (xe)->info.step.graphics; first <= __step && __step <= last; }) 130 #define IS_DG1_GRAPHICS_STEP(xe, first, last) (IS_DG1(xe) && IS_GRAPHICS_STEP(xe, first, last)) 131 #define IS_DG2_GRAPHICS_STEP(xe, variant, first, last) \ 133 IS_GRAPHICS_STEP(xe, first, last)) [all...] |
/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_object.h | 69 uint64_t last; member in struct:amdgpu_bo_va_mapping
|
/linux-master/drivers/firewire/ |
H A D | ohci.c | 103 struct descriptor *last); 142 struct descriptor *last; member in struct:context 145 * The last descriptor block in the DMA program. It contains the branch 761 * We search for the buffer that contains the last AR packet DMA data written 767 unsigned int i, next_i, last = ctx->last_buffer_index; local 773 /* A buffer that is not yet completely filled must be the last one. */ 774 while (i != last && res_count == 0) { 792 if (MAX_AR_PACKET_SIZE > PAGE_SIZE && i != last) { 989 * all packets up to the buffer end here. If the last packet 1099 struct descriptor *d, *last; local 1337 struct descriptor *d, *last; local 1472 handle_at_packet(struct context *context, struct descriptor *d, struct descriptor *last) argument 2825 handle_ir_packet_per_buffer(struct context *context, struct descriptor *d, struct descriptor *last) argument 2860 handle_ir_buffer_fill(struct context *context, struct descriptor *d, struct descriptor *last) argument 2948 handle_it_packet(struct context *context, struct descriptor *d, struct descriptor *last) argument 3261 struct descriptor *d, *last, *pd; local [all...] |
/linux-master/net/sunrpc/ |
H A D | xprtsock.c | 1804 unsigned short last; local 1834 last = port; 1836 if (port > last)
|
/linux-master/kernel/sched/ |
H A D | fair.c | 952 struct rb_node *last = rb_last(&cfs_rq->tasks_timeline.rb_root); local 954 if (!last) 957 return __node_2_se(last); 2686 * Get the fraction of time the task has been running since the last 2852 /* Decay existing window, copy faults since last scan */ 3425 * It is possible to reach the end of the VMA list but the last few 4099 * differential update where we store the last value we propagated. This in 4617 * util_avg. Given that period_contrib might have moved since the last 4927 * already ~1% close to its last activation value. 5397 /* return excess runtime on last dequeu [all...] |
/linux-master/kernel/ |
H A D | cpu.c | 59 * @last: For multi-instance rollback, remember how far we got 77 struct hlist_node *last; member in struct:cpuhp_cpu_state 705 st->last = NULL; 734 * If we have st->last we need to undo partial multi_instance of this 737 if (!st->last) { 1084 st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last); 1092 st->result = cpuhp_invoke_callback(cpu, state, bringup, st->node, &st->last); 1138 st->last = NULL; 1150 if ((ret = st->result) && st->last) { 1161 st->node = st->last [all...] |
/linux-master/drivers/dma/ |
H A D | pl330.c | 290 * The PL330 core does not modify this and uses the last 380 /* Index of the last enqueued request */ 382 /* Index of the last submitted request or -1 if the DMA is stopped */ 538 bool last; member in struct:dma_pl330_desc 2086 /* Try to submit a req imm. next to the last completed cookie */ 2468 if (desc->last) 2503 * We returned the last one of the circular list of descriptor(s) 2504 * from prep_xxx, so the argument to submit corresponds to the last 2509 struct dma_pl330_desc *desc, *last = to_desc(tx); local 2517 while (!list_empty(&last [all...] |
/linux-master/net/wireless/ |
H A D | nl80211.c | 50 NL80211_MCGRP_TESTMODE /* keep last - ifdef! */ 3670 * Fragments (apart from the last one) are required to 7561 /* must be last in here for error handling */ 9481 * All scan plans but the last one must specify 9491 * The last scan plan must not specify the number of 12817 s32 last, low, high; local 12846 last = cqm_config->last_rssi_event_value; 12852 if (last < cqm_config->rssi_thresholds[i])
|
/linux-master/net/tls/ |
H A D | tls_strp.c | 283 struct sk_buff *nskb, *first, *last; local 304 last = first->prev; 305 last->next = nskb;
|
/linux-master/net/ax25/ |
H A D | af_ax25.c | 1665 struct sk_buff *skb, *last; local 1684 skb = __skb_try_recv_datagram(sk, sk_queue, flags, &off, &err, &last); 1692 &timeo, last)) { 1694 &err, &last);
|
/linux-master/include/linux/ |
H A D | mm.h | 1008 return vmi->mas.last + 1; 1037 vmi->mas.last = vma->vm_end - 1; 1500 * last reference. 1520 * the last references. 3249 unsigned long start, unsigned long last); 3251 unsigned long start, unsigned long last); 3253 #define vma_interval_tree_foreach(vma, root, start, last) \ 3254 for (vma = vma_interval_tree_iter_first(root, start, last); \ 3255 vma; vma = vma_interval_tree_iter_next(vma, start, last)) 3263 unsigned long start, unsigned long last); [all...] |
/linux-master/fs/netfs/ |
H A D | buffered_write.c | 569 pgoff_t last = (start + len - 1) / PAGE_SIZE, next; local 574 _debug("kill %lx (to %lx)", index, last); 594 } while (index = next, index <= last); 607 pgoff_t last = (start + len - 1) / PAGE_SIZE, next; local 627 } while (index = next, index <= last); 643 pgoff_t last; local 652 last = (wreq->start + wreq->len - 1) / PAGE_SIZE; 653 xas_for_each(&xas, folio, last) { 656 wreq->len, wreq->start, folio->index, last);
|
/linux-master/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt.c | 777 int j, last; local 799 last = tx_buf->nr_frags; 801 for (j = 0; j < last; j++) { 1289 u16 last; local 1293 last = RING_CMP(*raw_cons); 1295 &cpr->cp_desc_ring[CP_RING(last)][CP_IDX(last)]; 3219 int k, last; local 3253 last = tx_buf->nr_frags; 3255 for (k = 0; k < last; 8452 bnxt_hwrm_func_backing_store_cfg_v2(struct bnxt *bp, struct bnxt_ctx_mem_type *ctxm, bool last) argument [all...] |