/linux-master/net/rds/ |
H A D | tcp.c | 546 struct ctl_table *tbl; local 555 tbl = rds_tcp_sysctl_table; 557 tbl = kmemdup(rds_tcp_sysctl_table, 559 if (!tbl) { 563 rtn->ctl_table = tbl; 565 tbl[RDS_TCP_SNDBUF].data = &rtn->sndbuf_size; 566 tbl[RDS_TCP_RCVBUF].data = &rtn->rcvbuf_size; 567 rtn->rds_tcp_sysctl = register_net_sysctl_sz(net, "net/rds/tcp", tbl, 601 kfree(tbl);
|
/linux-master/net/openvswitch/ |
H A D | flow_netlink.c | 1773 const struct ovs_len_tbl *tbl) 1780 if (tbl[nla_type(nla)].len == OVS_ATTR_NESTED) 1781 nlattr_set(nla, val, tbl[nla_type(nla)].next ? : tbl); 1772 nlattr_set(struct nlattr *attr, u8 val, const struct ovs_len_tbl *tbl) argument
|
/linux-master/net/netlink/ |
H A D | af_netlink.c | 549 struct netlink_table *tbl = &nl_table[sk->sk_protocol]; local 554 listeners = nl_deref_protected(tbl->listeners); 558 for (i = 0; i < NLGRPLONGS(tbl->groups); i++) { 560 sk_for_each_bound(sk, &tbl->mc_list) { 2098 struct netlink_table *tbl = &nl_table[sk->sk_protocol]; local 2103 if (NLGRPSZ(tbl->groups) < NLGRPSZ(groups)) { 2107 old = nl_deref_protected(tbl->listeners); 2108 memcpy(new->masks, old->masks, NLGRPSZ(tbl->groups)); 2109 rcu_assign_pointer(tbl->listeners, new); 2113 tbl 2144 struct netlink_table *tbl = &nl_table[ksk->sk_protocol]; local [all...] |
/linux-master/net/netfilter/ipvs/ |
H A D | ip_vs_lblcr.c | 330 ip_vs_lblcr_hash(struct ip_vs_lblcr_table *tbl, struct ip_vs_lblcr_entry *en) argument 334 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); 335 atomic_inc(&tbl->entries); 341 ip_vs_lblcr_get(int af, struct ip_vs_lblcr_table *tbl, argument 347 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) 360 ip_vs_lblcr_new(struct ip_vs_lblcr_table *tbl, const union nf_inet_addr *daddr, argument 365 en = ip_vs_lblcr_get(af, tbl, daddr); 381 ip_vs_lblcr_hash(tbl, en); 396 struct ip_vs_lblcr_table *tbl = svc->sched_data; local 402 tbl 422 struct ip_vs_lblcr_table *tbl = svc->sched_data; local 459 struct ip_vs_lblcr_table *tbl = from_timer(tbl, t, periodic_timer); local 508 struct ip_vs_lblcr_table *tbl; local 546 struct ip_vs_lblcr_table *tbl = svc->sched_data; local 647 struct ip_vs_lblcr_table *tbl = svc->sched_data; local [all...] |
H A D | ip_vs_lblc.c | 9 * *lock(tbl->lock) ==> *lock(&tbl->lock) 10 * Wensong Zhang : fixed the uninitialized tbl->lock bug 167 ip_vs_lblc_hash(struct ip_vs_lblc_table *tbl, struct ip_vs_lblc_entry *en) argument 171 hlist_add_head_rcu(&en->list, &tbl->bucket[hash]); 172 atomic_inc(&tbl->entries); 178 ip_vs_lblc_get(int af, struct ip_vs_lblc_table *tbl, argument 184 hlist_for_each_entry_rcu(en, &tbl->bucket[hash], list) 197 ip_vs_lblc_new(struct ip_vs_lblc_table *tbl, const union nf_inet_addr *daddr, argument 202 en = ip_vs_lblc_get(af, tbl, dadd 230 struct ip_vs_lblc_table *tbl = svc->sched_data; local 257 struct ip_vs_lblc_table *tbl = svc->sched_data; local 295 struct ip_vs_lblc_table *tbl = from_timer(tbl, t, periodic_timer); local 345 struct ip_vs_lblc_table *tbl; local 383 struct ip_vs_lblc_table *tbl = svc->sched_data; local 483 struct ip_vs_lblc_table *tbl = svc->sched_data; local [all...] |
H A D | ip_vs_ctl.c | 4269 struct ctl_table *tbl; local 4284 tbl = kmemdup(vs_vars, sizeof(vs_vars), GFP_KERNEL); 4285 if (tbl == NULL) 4288 tbl = vs_vars; 4291 if (tbl[idx].proc_handler == proc_do_defense_mode) 4292 tbl[idx].extra2 = ipvs; 4296 tbl[idx++].data = &ipvs->sysctl_amemthresh; 4298 tbl[idx++].data = &ipvs->sysctl_am_droprate; 4299 tbl[idx++].data = &ipvs->sysctl_drop_entry; 4300 tbl[id [all...] |
/linux-master/net/mptcp/ |
H A D | ctrl.c | 121 struct ctl_table tbl = { local 129 ret = proc_dostring(&tbl, write, buffer, lenp, ppos); 140 struct ctl_table tbl = { .maxlen = MPTCP_SCHED_BUF_MAX, }; local 143 tbl.data = kmalloc(tbl.maxlen, GFP_USER); 144 if (!tbl.data) 147 mptcp_get_available_schedulers(tbl.data, MPTCP_SCHED_BUF_MAX); 148 ret = proc_dostring(&tbl, write, buffer, lenp, ppos); 149 kfree(tbl.data);
|
/linux-master/net/ipv6/ |
H A D | ndisc.c | 123 .tbl = &nd_tbl,
|
H A D | ip6_fib.c | 2578 lockdep_is_held(&iter->tbl->tb6_lock)); 2592 iter->w.root = &iter->tbl->tb6_root; 2601 static struct fib6_table *ipv6_route_seq_next_table(struct fib6_table *tbl, argument 2607 if (tbl) { 2608 h = (tbl->tb6_id & (FIB6_TABLE_HASHSZ - 1)) + 1; 2609 node = rcu_dereference(hlist_next_rcu(&tbl->tb6_hlist)); 2652 spin_lock_bh(&iter->tbl->tb6_lock); 2654 spin_unlock_bh(&iter->tbl->tb6_lock); 2663 iter->tbl = ipv6_route_seq_next_table(iter->tbl, ne [all...] |
/linux-master/net/ipv4/ |
H A D | sysctl_net_ipv4.c | 217 struct ctl_table tbl = { local 225 ret = proc_dostring(&tbl, write, buffer, lenp, ppos); 235 struct ctl_table tbl = { .maxlen = TCP_CA_BUF_MAX, }; local 238 tbl.data = kmalloc(tbl.maxlen, GFP_USER); 239 if (!tbl.data) 241 tcp_get_available_congestion_control(tbl.data, TCP_CA_BUF_MAX); 242 ret = proc_dostring(&tbl, write, buffer, lenp, ppos); 243 kfree(tbl.data); 251 struct ctl_table tbl local 292 struct ctl_table tbl = { .maxlen = ((TCP_FASTOPEN_KEY_LENGTH * local 374 struct ctl_table tbl = { .maxlen = TCP_ULP_BUF_MAX, }; local 394 struct ctl_table tbl; local 417 struct ctl_table tbl; local [all...] |
H A D | route.c | 3536 struct ctl_table *tbl; local 3539 tbl = ipv4_route_netns_table; 3543 tbl = kmemdup(tbl, sizeof(ipv4_route_netns_table), GFP_KERNEL); 3544 if (!tbl) 3549 if (tbl[0].procname != ipv4_route_flush_procname) 3557 tbl[i].data += (void *)net - (void *)&init_net; 3559 tbl[0].extra1 = net; 3562 tbl, table_size); 3568 if (tbl ! 3576 const struct ctl_table *tbl; local [all...] |
H A D | ipmr.c | 2735 struct mr_table *tbl; local 2739 tbl = NULL; 2740 ret = rtm_to_ipmr_mfcc(net, nlh, &mfcc, &mrtsock, &tbl, extack); 2746 return ipmr_mfc_add(net, tbl, &mfcc, mrtsock, parent); 2748 return ipmr_mfc_delete(tbl, &mfcc, parent);
|
H A D | fib_semantics.c | 1157 struct fib_table *tbl = NULL; local 1170 tbl = fib_get_table(net, table); 1172 if (tbl) 1173 err = fib_table_lookup(tbl, &fl4, &res, 1181 if (!tbl || err) {
|
H A D | devinet.c | 2675 struct ctl_table *tbl; local 2689 tbl = kmemdup(ctl_forward_entry, sizeof(ctl_forward_entry), GFP_KERNEL); 2690 if (!tbl) 2693 tbl[0].data = &all->data[IPV4_DEVCONF_FORWARDING - 1]; 2694 tbl[0].extra1 = all; 2695 tbl[0].extra2 = net; 2733 forw_hdr = register_net_sysctl_sz(net, "net/ipv4", tbl, 2750 kfree(tbl); 2763 const struct ctl_table *tbl; local 2765 tbl [all...] |
H A D | arp.c | 163 .tbl = &arp_tbl, 1204 struct neigh_table *tbl = &arp_tbl; local 1216 write_lock_bh(&tbl->lock); 1218 neigh_remove_one(neigh, tbl); 1219 write_unlock_bh(&tbl->lock);
|
/linux-master/net/core/ |
H A D | sysctl_net_core.c | 284 struct ctl_table tbl = { local 292 ret = proc_dostring(&tbl, write, buffer, lenp, ppos); 718 struct ctl_table *tbl; local 720 tbl = netns_core_table; 723 tbl = kmemdup(tbl, sizeof(netns_core_table), GFP_KERNEL); 724 if (tbl == NULL) 728 tbl[i].data += (char *)net - (char *)&init_net; 731 net->core.sysctl_hdr = register_net_sysctl_sz(net, "net/core", tbl, table_size); 738 if (tbl ! 746 const struct ctl_table *tbl; local [all...] |
H A D | neighbour.c | 57 static int pneigh_ifdown_and_unlock(struct neigh_table *tbl, 65 Neighbour hash table buckets are protected with rwlock tbl->lock. 123 atomic_dec(&n->tbl->gc_entries); 133 write_lock_bh(&n->tbl->lock); 147 atomic_dec(&n->tbl->gc_entries); 150 list_add_tail(&n->gc_list, &n->tbl->gc_list); 151 atomic_inc(&n->tbl->gc_entries); 155 write_unlock_bh(&n->tbl->lock); 162 write_lock_bh(&n->tbl->lock); 173 list_add_tail(&n->managed_list, &n->tbl 208 neigh_del(struct neighbour *n, struct neighbour __rcu **np, struct neigh_table *tbl) argument 229 neigh_remove_one(struct neighbour *ndel, struct neigh_table *tbl) argument 252 neigh_forced_gc(struct neigh_table *tbl) argument 380 neigh_flush_dev(struct neigh_table *tbl, struct net_device *dev, bool skip_perm) argument 434 neigh_changeaddr(struct neigh_table *tbl, struct net_device *dev) argument 442 __neigh_ifdown(struct neigh_table *tbl, struct net_device *dev, bool skip_perm) argument 455 neigh_carrier_down(struct neigh_table *tbl, struct net_device *dev) argument 462 neigh_ifdown(struct neigh_table *tbl, struct net_device *dev) argument 469 neigh_alloc(struct neigh_table *tbl, struct net_device *dev, u32 flags, bool exempt_from_gc) argument 577 neigh_hash_grow(struct neigh_table *tbl, unsigned long new_shift) argument 618 neigh_lookup(struct neigh_table *tbl, const void *pkey, struct net_device *dev) argument 639 ___neigh_create(struct neigh_table *tbl, const void *pkey, struct net_device *dev, u32 flags, bool exempt_from_gc, bool want_ref) argument 734 __neigh_create(struct neigh_table *tbl, const void *pkey, struct net_device *dev, bool want_ref) argument 769 __pneigh_lookup(struct neigh_table *tbl, struct net *net, const void *pkey, struct net_device *dev) argument 780 pneigh_lookup(struct neigh_table *tbl, struct net *net, const void *pkey, struct net_device *dev, int creat) argument 824 pneigh_delete(struct neigh_table *tbl, struct net *net, const void *pkey, struct net_device *dev) argument 849 pneigh_ifdown_and_unlock(struct neigh_table *tbl, struct net_device *dev) argument 950 struct neigh_table *tbl = container_of(work, struct neigh_table, gc_work.work); local 1512 neigh_event_ns(struct neigh_table *tbl, u8 *lladdr, void *saddr, struct net_device *dev) argument 1611 struct neigh_table *tbl = container_of(work, struct neigh_table, local 1625 struct neigh_table *tbl = from_timer(tbl, t, proxy_timer); local 1670 pneigh_enqueue(struct neigh_table *tbl, struct neigh_parms *p, struct sk_buff *skb) argument 1697 lookup_neigh_parms(struct neigh_table *tbl, struct net *net, int ifindex) argument 1711 neigh_parms_alloc(struct net_device *dev, struct neigh_table *tbl) argument 1754 neigh_parms_release(struct neigh_table *tbl, struct neigh_parms *parms) argument 1776 neigh_table_init(int index, struct neigh_table *tbl) argument 1839 neigh_table_clear(int index, struct neigh_table *tbl) argument 1871 struct neigh_table *tbl = NULL; local 1908 struct neigh_table *tbl; local 1975 struct neigh_table *tbl; local 2169 neightbl_fill_info(struct sk_buff *skb, struct neigh_table *tbl, u32 pid, u32 seq, int type, int flags) argument 2259 neightbl_fill_param_info(struct sk_buff *skb, struct neigh_table *tbl, struct neigh_parms *parms, u32 pid, u32 seq, int type, unsigned int flags) argument 2323 struct neigh_table *tbl; local 2516 struct neigh_table *tbl; local 2632 pneigh_fill_info(struct sk_buff *skb, struct pneigh_entry *pn, u32 pid, u32 seq, int type, unsigned int flags, struct neigh_table *tbl) argument 2712 neigh_dump_table(struct neigh_table *tbl, struct sk_buff *skb, struct netlink_callback *cb, struct neigh_dump_filter *filter) argument 2754 pneigh_dump_table(struct neigh_table *tbl, struct sk_buff *skb, struct netlink_callback *cb, struct neigh_dump_filter *filter) argument 2862 struct neigh_table *tbl; local 2906 neigh_valid_get_req(const struct nlmsghdr *nlh, struct neigh_table **tbl, void **dst, int *dev_idx, u8 *ndm_flags, struct netlink_ext_ack *extack) argument 3006 pneigh_get_reply(struct net *net, struct pneigh_entry *neigh, u32 pid, u32 seq, struct neigh_table *tbl) argument 3032 struct neigh_table *tbl = NULL; local 3088 neigh_for_each(struct neigh_table *tbl, void (*cb)(struct neighbour *, void *), void *cookie) argument 3111 __neigh_for_each_release(struct neigh_table *tbl, int (*cb)(struct neighbour *)) argument 3151 struct neigh_table *tbl; local 3299 struct neigh_table *tbl = state->tbl; local 3322 struct neigh_table *tbl = state->tbl; local 3422 struct neigh_table *tbl = state->tbl; local 3433 struct neigh_table *tbl = pde_data(file_inode(seq->file)); local 3450 struct neigh_table *tbl = pde_data(file_inode(seq->file)); local 3470 struct neigh_table *tbl = pde_data(file_inode(seq->file)); local 3816 struct neigh_table *tbl = p->tbl; local [all...] |
/linux-master/net/bridge/ |
H A D | br_vlan_tunnel.c | 37 static struct net_bridge_vlan *br_vlan_tunnel_lookup(struct rhashtable *tbl, argument 40 return rhashtable_lookup_fast(tbl, &tunnel_id,
|
/linux-master/net/atm/ |
H A D | clip.c | 294 if (neigh->tbl->family != AF_INET)
|
/linux-master/include/net/ |
H A D | udp.h | 303 struct udp_table *tbl, struct sk_buff *skb); 313 int dif, int sdif, struct udp_table *tbl,
|
H A D | ip6_fib.h | 551 struct fib6_table *tbl; member in struct:ipv6_route_iter
|
/linux-master/include/linux/ |
H A D | rhashtable.h | 119 static inline unsigned int rht_bucket_index(const struct bucket_table *tbl, argument 122 return hash & (tbl->size - 1); 156 struct rhashtable *ht, const struct bucket_table *tbl, 159 unsigned int hash = rht_key_get_hash(ht, key, params, tbl->hash_rnd); 161 return rht_bucket_index(tbl, hash); 165 struct rhashtable *ht, const struct bucket_table *tbl, 171 rht_bucket_index(tbl, params.obj_hashfn(ptr, params.key_len ?: 173 tbl->hash_rnd)) : 174 rht_key_hashfn(ht, tbl, ptr + params.key_offset, params); 180 * @tbl 155 rht_key_hashfn( struct rhashtable *ht, const struct bucket_table *tbl, const void *key, const struct rhashtable_params params) argument 164 rht_head_hashfn( struct rhashtable *ht, const struct bucket_table *tbl, const struct rhash_head *he, const struct rhashtable_params params) argument 182 rht_grow_above_75(const struct rhashtable *ht, const struct bucket_table *tbl) argument 195 rht_shrink_below_30(const struct rhashtable *ht, const struct bucket_table *tbl) argument 208 rht_grow_above_100(const struct rhashtable *ht, const struct bucket_table *tbl) argument 220 rht_grow_above_max(const struct rhashtable *ht, const struct bucket_table *tbl) argument 235 lockdep_rht_bucket_is_held(const struct bucket_table *tbl, u32 hash) argument 286 rht_bucket( const struct bucket_table *tbl, unsigned int hash) argument 293 rht_bucket_var( struct bucket_table *tbl, unsigned int hash) argument 300 rht_bucket_insert( struct rhashtable *ht, struct bucket_table *tbl, unsigned int hash) argument 326 rht_lock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt) argument 337 rht_lock_nested(struct bucket_table *tbl, struct rhash_lock_head __rcu **bucket, unsigned int subclass) argument 349 rht_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, unsigned long flags) argument 379 rht_ptr( struct rhash_lock_head __rcu *const *bkt, struct bucket_table *tbl, unsigned int hash) argument 401 rht_assign_unlock(struct bucket_table *tbl, struct rhash_lock_head __rcu **bkt, struct rhash_head *obj, unsigned long flags) argument 598 struct bucket_table *tbl; local 715 struct bucket_table *tbl; local 995 __rhashtable_remove_fast_one( struct rhashtable *ht, struct bucket_table *tbl, struct rhash_head *obj, const struct rhashtable_params params, bool rhlist) argument 1081 struct bucket_table *tbl; local 1148 __rhashtable_replace_fast( struct rhashtable *ht, struct bucket_table *tbl, struct rhash_head *obj_old, struct rhash_head *obj_new, const struct rhashtable_params params) argument 1216 struct bucket_table *tbl; local [all...] |
H A D | fs_parser.h | 82 extern int lookup_constant(const struct constant_table tbl[], const char *name, int not_found); 85 extern bool validate_constant_table(const struct constant_table *tbl, size_t tbl_size, 90 static inline bool validate_constant_table(const struct constant_table *tbl, size_t tbl_size, argument
|
/linux-master/drivers/ptp/ |
H A D | ptp_ocp.c | 1130 const struct ocp_selector *tbl[2]; member in struct:ocp_sma_op 1162 ptp_ocp_select_name_from_val(const struct ocp_selector *tbl, int val) argument 1166 for (i = 0; tbl[i].name; i++) 1167 if (tbl[i].value == val) 1168 return tbl[i].name; 1173 ptp_ocp_select_val_from_name(const struct ocp_selector *tbl, const char *name) argument 1178 for (i = 0; tbl[i].name; i++) { 1179 select = tbl[i].name; 1181 return tbl[i].value; 1187 ptp_ocp_select_table_show(const struct ocp_selector *tbl, cha argument 2849 ptp_ocp_show_output(const struct ocp_selector *tbl, u32 val, char *buf, int def_val) argument 2864 ptp_ocp_show_inputs(const struct ocp_selector *tbl, u32 val, char *buf, int def_val) argument 2889 sma_parse_inputs(const struct ocp_selector * const tbl[], const char *buf, enum ptp_ocp_sma_mode *mode) argument 2932 const struct ocp_selector * const *tbl; local 4576 const struct ocp_selector *tbl; local 4598 const struct ocp_selector *tbl; local [all...] |
/linux-master/drivers/net/wireless/realtek/rtw89/ |
H A D | phy.c | 1768 const struct rtw89_phy_reg3_tbl *tbl) 1773 for (i = 0; i < tbl->size; i++) { 1774 reg3 = &tbl->reg3[i]; 1818 const struct rtw89_txpwr_table *tbl) 1820 const struct rtw89_txpwr_byrate_cfg *cfg = tbl->data; 1821 const struct rtw89_txpwr_byrate_cfg *end = cfg + tbl->size; 3710 "[TSSI] tmeter tbl on subband: %u\n", subband); 6014 rtw89_rfk_parser(struct rtw89_dev *rtwdev, const struct rtw89_rfk_tbl *tbl) argument 6016 const struct rtw89_reg5_def *p = tbl->defs; 6017 const struct rtw89_reg5_def *end = tbl 1767 rtw89_phy_write_reg3_tbl(struct rtw89_dev *rtwdev, const struct rtw89_phy_reg3_tbl *tbl) argument 1817 rtw89_phy_load_txpwr_byrate(struct rtw89_dev *rtwdev, const struct rtw89_txpwr_table *tbl) argument [all...] |