/linux-master/net/core/ |
H A D | sock_reuseport.c | 22 struct sock_reuseport *reuse, bool bind_inany); 26 struct sock_reuseport *reuse; local 32 reuse = rcu_dereference_protected(sk->sk_reuseport_cb, 34 if (likely(reuse)) 35 reuse->has_conns = 1; 40 static void __reuseport_get_incoming_cpu(struct sock_reuseport *reuse) argument 43 WRITE_ONCE(reuse->incoming_cpu, reuse->incoming_cpu + 1); 46 static void __reuseport_put_incoming_cpu(struct sock_reuseport *reuse) argument 49 WRITE_ONCE(reuse 52 reuseport_get_incoming_cpu(struct sock *sk, struct sock_reuseport *reuse) argument 58 reuseport_put_incoming_cpu(struct sock *sk, struct sock_reuseport *reuse) argument 66 struct sock_reuseport *reuse; local 104 reuseport_sock_index(struct sock *sk, const struct sock_reuseport *reuse, bool closed) argument 124 __reuseport_add_sock(struct sock *sk, struct sock_reuseport *reuse) argument 134 __reuseport_detach_sock(struct sock *sk, struct sock_reuseport *reuse) argument 149 __reuseport_add_closed_sock(struct sock *sk, struct sock_reuseport *reuse) argument 158 __reuseport_detach_closed_sock(struct sock *sk, struct sock_reuseport *reuse) argument 178 struct sock_reuseport *reuse = kzalloc(size, GFP_ATOMIC); local 191 struct sock_reuseport *reuse; local 247 reuseport_grow(struct sock_reuseport *reuse) argument 305 struct sock_reuseport *reuse; local 323 struct sock_reuseport *old_reuse, *reuse; local 369 reuseport_resurrect(struct sock *sk, struct sock_reuseport *old_reuse, struct sock_reuseport *reuse, bool bind_inany) argument 430 struct sock_reuseport *reuse; local 466 struct sock_reuseport *reuse; local 498 run_bpf_filter(struct sock_reuseport *reuse, u16 socks, struct bpf_prog *prog, struct sk_buff *skb, int hdr_len) argument 528 reuseport_select_sock_by_hash(struct sock_reuseport *reuse, u32 hash, u16 num_socks) argument 574 struct sock_reuseport *reuse; local 625 struct sock_reuseport *reuse; local 686 struct sock_reuseport *reuse; local 718 struct sock_reuseport *reuse; local [all...] |
/linux-master/include/net/ |
H A D | sock_reuseport.h | 21 * reuse->socks[] group. 49 struct sock_reuseport *reuse; local 53 reuse = rcu_dereference(sk->sk_reuseport_cb); 54 if (reuse && reuse->has_conns)
|
H A D | tcp.h | 540 struct sock_reuseport *reuse; local 542 reuse = rcu_dereference(sk->sk_reuseport_cb); 543 if (likely(reuse)) { 544 last_overflow = READ_ONCE(reuse->synq_overflow_ts); 547 WRITE_ONCE(reuse->synq_overflow_ts, now); 564 struct sock_reuseport *reuse; local 566 reuse = rcu_dereference(sk->sk_reuseport_cb); 567 if (likely(reuse)) { 568 last_overflow = READ_ONCE(reuse->synq_overflow_ts);
|
/linux-master/mm/ |
H A D | hugetlb_vmemmap.c | 232 * How many struct page structs need to be reset. When we reuse the head 280 * @reuse: reuse address. 285 unsigned long reuse) 293 BUG_ON(start - reuse != PAGE_SIZE); 295 return vmemmap_remap_range(reuse, end, &walk); 300 * to the page which @reuse is mapped to, then free vmemmap 306 * @reuse: reuse address. 314 unsigned long reuse, 284 vmemmap_remap_split(unsigned long start, unsigned long end, unsigned long reuse) argument 313 vmemmap_remap_free(unsigned long start, unsigned long end, unsigned long reuse, struct list_head *vmemmap_pages, unsigned long flags) argument 415 vmemmap_remap_alloc(unsigned long start, unsigned long end, unsigned long reuse, unsigned long flags) argument [all...] |
H A D | sparse-vmemmap.c | 146 struct page *reuse) 153 if (!reuse) { 167 get_page(reuse); 168 p = page_to_virt(reuse); 247 struct page *reuse) 267 pte = vmemmap_pte_populate(pmd, addr, node, altmap, reuse); 278 struct page *reuse) 284 pte = vmemmap_populate_address(addr, node, altmap, reuse); 144 vmemmap_pte_populate(pmd_t *pmd, unsigned long addr, int node, struct vmem_altmap *altmap, struct page *reuse) argument 245 vmemmap_populate_address(unsigned long addr, int node, struct vmem_altmap *altmap, struct page *reuse) argument 275 vmemmap_populate_range(unsigned long start, unsigned long end, int node, struct vmem_altmap *altmap, struct page *reuse) argument
|
/linux-master/tools/testing/selftests/net/ |
H A D | bind_bhash.c | 33 int sock_fd, reuse = 1, err; local 52 err = setsockopt(sock_fd, SOL_SOCKET, opt, &reuse, sizeof(reuse));
|
H A D | bind_wildcard.c | 742 int i, int reuse) 754 if (i < 2 && reuse) { 755 ret = setsockopt(self->fd[i], SOL_SOCKET, reuse, &(int){1}, sizeof(int)); 763 if (reuse) { 739 bind_socket(struct __test_metadata *_metadata, FIXTURE_DATA(bind_wildcard) *self, const FIXTURE_VARIANT(bind_wildcard) *variant, int i, int reuse) argument
|
/linux-master/tools/testing/selftests/bpf/progs/ |
H A D | test_sockmap_listen.c | 121 int prog_reuseport(struct sk_reuseport_md *reuse) argument 128 err = bpf_sk_select_reuseport(reuse, &sock_map, &zero, 0); 130 err = bpf_sk_select_reuseport(reuse, &sock_hash, &zero, 0);
|
H A D | test_tcpbpf_kern.c | 46 struct bpf_sock_ops *reuse = skops; local 56 "%[reuse] = *(u32 *)(%[reuse] +96)" 57 : [reuse] "+r"(reuse)
|
/linux-master/include/xen/interface/io/ |
H A D | pvcalls.h | 59 uint8_t reuse; member in struct:xen_pvcalls_request::__anon189::xen_pvcalls_release
|
/linux-master/kernel/bpf/ |
H A D | reuseport_array.c | 212 * it must also be a SO_REUSEPORT sk (i.e. reuse cannot be NULL). 230 * The "osk" and "reuse" are protected by reuseport_lock. 237 struct sock_reuseport *reuse; local 287 reuse = rcu_dereference_protected(nsk->sk_reuseport_cb, 289 err = reuseport_array_update_check(array, nsk, osk, reuse, map_flags);
|
/linux-master/drivers/rpmsg/ |
H A D | qcom_glink_native.c | 62 * @reuse: To mark if the intent can be reused after first use 71 bool reuse; member in struct:glink_core_rx_intent 262 /* Free all non-reuse intents pending rx_done work */ 264 if (!intent->reuse) { 537 bool reuse; local 545 reuse = intent->reuse; 547 cmd.id = reuse ? GLINK_CMD_RX_DONE_W_REUSE : GLINK_CMD_RX_DONE; 552 if (!reuse) { 573 if (!intent->reuse) { 740 qcom_glink_handle_rx_done(struct qcom_glink *glink, u32 cid, uint32_t iid, bool reuse) argument [all...] |
/linux-master/include/rdma/ |
H A D | rdma_cm.h | 325 * rdma_set_reuseaddr - Allow the reuse of local addresses when binding 328 * @reuse: Value indicating if the bound address is reusable. 332 int rdma_set_reuseaddr(struct rdma_cm_id *id, int reuse);
|
/linux-master/drivers/net/ethernet/engleder/ |
H A D | tsnep_main.c | 1162 static int tsnep_rx_alloc(struct tsnep_rx *rx, int count, bool reuse) argument 1174 /* reuse only if no other allocation was successful */ 1175 if (i == 0 && reuse) 1190 static int tsnep_rx_refill(struct tsnep_rx *rx, int count, bool reuse) argument 1194 desc_refilled = tsnep_rx_alloc(rx, count, reuse); 1219 static int tsnep_rx_alloc_zc(struct tsnep_rx *rx, int count, bool reuse) argument 1235 if (reuse) { 1260 static int tsnep_rx_refill_zc(struct tsnep_rx *rx, int count, bool reuse) argument 1264 desc_refilled = tsnep_rx_alloc_zc(rx, count, reuse); 1462 bool reuse local 1563 bool reuse = desc_available >= TSNEP_RING_RX_REUSE; local [all...] |
/linux-master/arch/arm/mach-omap1/ |
H A D | ams-delta-fiq-handler.S | 155 @ r10 already contains 0, reuse it 180 @ r10 already contains 0, reuse it
|
/linux-master/arch/powerpc/mm/book3s64/ |
H A D | radix_pgtable.c | 952 struct page *reuse) 960 if (!reuse) { 984 get_page(reuse); 985 p = page_to_virt(reuse); 986 pr_debug("Tail page reuse vmemmap mapping\n"); 1136 struct page *reuse) 949 radix__vmemmap_pte_populate(pmd_t *pmdp, unsigned long addr, int node, struct vmem_altmap *altmap, struct page *reuse) argument 1134 radix__vmemmap_populate_address(unsigned long addr, int node, struct vmem_altmap *altmap, struct page *reuse) argument
|
/linux-master/net/ipv4/ |
H A D | inet_connection_sock.c | 464 bool reuse = sk->sk_reuse && sk->sk_state != TCP_LISTEN; local 467 tb->fastreuse = reuse; 481 if (!reuse) 518 bool reuse = sk->sk_reuse && sk->sk_state != TCP_LISTEN; local 560 (tb->fastreuse > 0 && reuse) ||
|
/linux-master/drivers/staging/rtl8723bs/include/ |
H A D | rtw_recv.h | 294 u8 reuse; member in struct:recv_buf
|
/linux-master/net/smc/ |
H A D | smc_stats.h | 191 SMC_STAT_RMB(smc, reuse, is_smcd, is_rx)
|
/linux-master/include/sound/ |
H A D | wavefront.h | 224 u8 reuse:1; member in struct:wf_patch
|
/linux-master/net/netfilter/ipset/ |
H A D | ip_set_hash_gen.h | 849 bool deleted = false, forceadd = false, reuse = false; local 896 deleted = reuse = true; 914 reuse = true; 917 if (reuse || forceadd) {
|
/linux-master/lib/ |
H A D | maple_tree.c | 391 * a reuse of the last bit in the node type. This is possible by using bit 1 to 1183 * Stores the maple node back into @mas->alloc for reuse. Updates allocated and 1188 struct maple_alloc *reuse = (struct maple_alloc *)used; local 1195 reuse->request_count = 0; 1196 reuse->node_count = 0; 1198 head->slot[head->node_count++] = reuse; 1203 reuse->total = 1; 1205 reuse->slot[0] = head; 1206 reuse->node_count = 1; 1207 reuse 3018 struct maple_node reuse, *newnode, *parent, *new_left, *left, *node; local 3888 struct maple_node reuse, *newnode; local [all...] |
/linux-master/net/sctp/ |
H A D | socket.c | 4263 sctp_sk(sk)->reuse = !!*val; 7738 val = sctp_sk(sk)->reuse; 8338 bool reuse = (sk->sk_reuse || sp->reuse); local 8415 if ((pp->fastreuse && reuse && 8437 (reuse && (sk2->sk_reuse || sp2->reuse) && 8465 if (reuse && sk->sk_state != SCTP_SS_LISTENING) 8478 (!reuse || sk->sk_state == SCTP_SS_LISTENING)) 8610 if (sk->sk_reuse || sctp_sk(sk)->reuse) [all...] |
/linux-master/include/linux/ |
H A D | filter.h | 1039 struct sock *bpf_run_sk_reuseport(struct sock_reuseport *reuse, struct sock *sk, 1045 bpf_run_sk_reuseport(struct sock_reuseport *reuse, struct sock *sk, argument
|
/linux-master/fs/nilfs2/ |
H A D | super.c | 555 goto reuse; /* already attached checkpoint */ 563 reuse:
|