/linux-master/arch/arm/mach-omap1/ |
H A D | ams-delta-fiq-handler.S | 155 @ r10 already contains 0, reuse it 180 @ r10 already contains 0, reuse it
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | radix_pgtable.c | 952 struct page *reuse) 960 if (!reuse) { 984 get_page(reuse); 985 p = page_to_virt(reuse); 986 pr_debug("Tail page reuse vmemmap mapping\n"); 1136 struct page *reuse) 949 radix__vmemmap_pte_populate(pmd_t *pmdp, unsigned long addr, int node, struct vmem_altmap *altmap, struct page *reuse) argument 1134 radix__vmemmap_populate_address(unsigned long addr, int node, struct vmem_altmap *altmap, struct page *reuse) argument
|
/linux-master/arch/x86/crypto/ |
H A D | sha256-avx2-asm.S | 592 leaq K256+0*32(%rip), INP ## reuse INP as scratch reg
|
/linux-master/drivers/gpu/drm/amd/amdkfd/ |
H A D | cwsr_trap_handler_gfx8.asm | 647 //reuse s_restore_m0 as a temp register
|
/linux-master/drivers/infiniband/core/ |
H A D | cma.c | 3574 int rdma_set_reuseaddr(struct rdma_cm_id *id, int reuse) argument 3582 if ((reuse && id_priv->state != RDMA_CM_LISTEN) || 3584 id_priv->reuseaddr = reuse;
|
/linux-master/drivers/net/ethernet/engleder/ |
H A D | tsnep_main.c | 1162 static int tsnep_rx_alloc(struct tsnep_rx *rx, int count, bool reuse) argument 1174 /* reuse only if no other allocation was successful */ 1175 if (i == 0 && reuse) 1190 static int tsnep_rx_refill(struct tsnep_rx *rx, int count, bool reuse) argument 1194 desc_refilled = tsnep_rx_alloc(rx, count, reuse); 1219 static int tsnep_rx_alloc_zc(struct tsnep_rx *rx, int count, bool reuse) argument 1235 if (reuse) { 1260 static int tsnep_rx_refill_zc(struct tsnep_rx *rx, int count, bool reuse) argument 1264 desc_refilled = tsnep_rx_alloc_zc(rx, count, reuse); 1462 bool reuse local 1563 bool reuse = desc_available >= TSNEP_RING_RX_REUSE; local [all...] |
/linux-master/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_ll2.c | 172 bool reuse = false; local 193 reuse = true; 196 if (!reuse) 200 /* If need to reuse or there's no replacement buffer, repost this */
|
/linux-master/drivers/rpmsg/ |
H A D | qcom_glink_native.c | 62 * @reuse: To mark if the intent can be reused after first use 71 bool reuse; member in struct:glink_core_rx_intent 262 /* Free all non-reuse intents pending rx_done work */ 264 if (!intent->reuse) { 537 bool reuse; local 545 reuse = intent->reuse; 547 cmd.id = reuse ? GLINK_CMD_RX_DONE_W_REUSE : GLINK_CMD_RX_DONE; 552 if (!reuse) { 573 if (!intent->reuse) { 740 qcom_glink_handle_rx_done(struct qcom_glink *glink, u32 cid, uint32_t iid, bool reuse) argument [all...] |
/linux-master/drivers/staging/rtl8723bs/include/ |
H A D | rtw_recv.h | 294 u8 reuse; member in struct:recv_buf
|
/linux-master/fs/f2fs/ |
H A D | compress.c | 152 void f2fs_destroy_compress_ctx(struct compress_ctx *cc, bool reuse) argument 159 if (!reuse)
|
H A D | f2fs.h | 1362 ALLOC_MODE_REUSE, /* reuse segments as much as possible */ 4304 void f2fs_destroy_compress_ctx(struct compress_ctx *cc, bool reuse);
|
/linux-master/fs/nilfs2/ |
H A D | super.c | 555 goto reuse; /* already attached checkpoint */ 563 reuse:
|
/linux-master/include/linux/ |
H A D | filter.h | 1040 struct sock *bpf_run_sk_reuseport(struct sock_reuseport *reuse, struct sock *sk, 1046 bpf_run_sk_reuseport(struct sock_reuseport *reuse, struct sock *sk, argument
|
H A D | mm.h | 3874 struct vmem_altmap *altmap, struct page *reuse);
|
/linux-master/include/net/sctp/ |
H A D | structs.h | 224 reuse:1, member in struct:sctp_sock
|
/linux-master/include/net/ |
H A D | sock_reuseport.h | 21 * reuse->socks[] group. 49 struct sock_reuseport *reuse; local 53 reuse = rcu_dereference(sk->sk_reuseport_cb); 54 if (reuse && reuse->has_conns)
|
H A D | tcp.h | 540 struct sock_reuseport *reuse; local 542 reuse = rcu_dereference(sk->sk_reuseport_cb); 543 if (likely(reuse)) { 544 last_overflow = READ_ONCE(reuse->synq_overflow_ts); 547 WRITE_ONCE(reuse->synq_overflow_ts, now); 564 struct sock_reuseport *reuse; local 566 reuse = rcu_dereference(sk->sk_reuseport_cb); 567 if (likely(reuse)) { 568 last_overflow = READ_ONCE(reuse->synq_overflow_ts);
|
/linux-master/include/rdma/ |
H A D | rdma_cm.h | 325 * rdma_set_reuseaddr - Allow the reuse of local addresses when binding 328 * @reuse: Value indicating if the bound address is reusable. 332 int rdma_set_reuseaddr(struct rdma_cm_id *id, int reuse);
|
/linux-master/include/sound/ |
H A D | wavefront.h | 224 u8 reuse:1; member in struct:wf_patch
|
/linux-master/include/xen/interface/io/ |
H A D | pvcalls.h | 59 uint8_t reuse; member in struct:xen_pvcalls_request::__anon189::xen_pvcalls_release
|
/linux-master/kernel/bpf/ |
H A D | reuseport_array.c | 212 * it must also be a SO_REUSEPORT sk (i.e. reuse cannot be NULL). 230 * The "osk" and "reuse" are protected by reuseport_lock. 237 struct sock_reuseport *reuse; local 287 reuse = rcu_dereference_protected(nsk->sk_reuseport_cb, 289 err = reuseport_array_update_check(array, nsk, osk, reuse, map_flags);
|
/linux-master/lib/ |
H A D | maple_tree.c | 391 * a reuse of the last bit in the node type. This is possible by using bit 1 to 1183 * Stores the maple node back into @mas->alloc for reuse. Updates allocated and 1188 struct maple_alloc *reuse = (struct maple_alloc *)used; local 1195 reuse->request_count = 0; 1196 reuse->node_count = 0; 1198 head->slot[head->node_count++] = reuse; 1203 reuse->total = 1; 1205 reuse->slot[0] = head; 1206 reuse->node_count = 1; 1207 reuse 3018 struct maple_node reuse, *newnode, *parent, *new_left, *left, *node; local 3888 struct maple_node reuse, *newnode; local [all...] |
/linux-master/mm/ |
H A D | huge_memory.c | 1541 goto reuse; 1560 goto reuse; 1564 * See do_wp_page(): we can only reuse the folio exclusively if 1579 reuse:
|
H A D | hugetlb_vmemmap.c | 232 * How many struct page structs need to be reset. When we reuse the head 280 * @reuse: reuse address. 285 unsigned long reuse) 293 BUG_ON(start - reuse != PAGE_SIZE); 295 return vmemmap_remap_range(reuse, end, &walk); 300 * to the page which @reuse is mapped to, then free vmemmap 306 * @reuse: reuse address. 314 unsigned long reuse, 284 vmemmap_remap_split(unsigned long start, unsigned long end, unsigned long reuse) argument 313 vmemmap_remap_free(unsigned long start, unsigned long end, unsigned long reuse, struct list_head *vmemmap_pages, unsigned long flags) argument 415 vmemmap_remap_alloc(unsigned long start, unsigned long end, unsigned long reuse, unsigned long flags) argument [all...] |
H A D | sparse-vmemmap.c | 146 struct page *reuse) 153 if (!reuse) { 167 get_page(reuse); 168 p = page_to_virt(reuse); 247 struct page *reuse) 267 pte = vmemmap_pte_populate(pmd, addr, node, altmap, reuse); 278 struct page *reuse) 284 pte = vmemmap_populate_address(addr, node, altmap, reuse); 144 vmemmap_pte_populate(pmd_t *pmd, unsigned long addr, int node, struct vmem_altmap *altmap, struct page *reuse) argument 245 vmemmap_populate_address(unsigned long addr, int node, struct vmem_altmap *altmap, struct page *reuse) argument 275 vmemmap_populate_range(unsigned long start, unsigned long end, int node, struct vmem_altmap *altmap, struct page *reuse) argument
|