/linux-master/net/ipv4/ |
H A D | inet_connection_sock.c | 241 /* This should be called only when the tb and tb2 hashbuckets' locks are held */ 243 const struct inet_bind_bucket *tb, 269 * in tb->owners and tb2->owners list belong 272 sk_for_each_bound_bhash(sk2, tb2, tb) { 340 struct inet_bind_bucket *tb; local 388 inet_bind_bucket_for_each(tb, &head->chain) 389 if (inet_bind_bucket_match(tb, net, port, l3mdev)) { 390 if (!inet_csk_bind_conflict(sk, tb, tb2, 396 tb = NULL; 421 *tb_ret = tb; 242 inet_csk_bind_conflict(const struct sock *sk, const struct inet_bind_bucket *tb, const struct inet_bind2_bucket *tb2, bool relax, bool reuseport_ok) argument 427 sk_reuseport_match(struct inet_bind_bucket *tb, struct sock *sk) argument 460 inet_csk_update_fastreuse(struct inet_bind_bucket *tb, struct sock *sk) argument 524 struct inet_bind_bucket *tb = NULL; local [all...] |
/linux-master/net/netfilter/ |
H A D | nft_nat.c | 172 const struct nlattr * const tb[]) 179 if (tb[NFTA_NAT_TYPE] == NULL || 180 (tb[NFTA_NAT_REG_ADDR_MIN] == NULL && 181 tb[NFTA_NAT_REG_PROTO_MIN] == NULL)) 184 switch (ntohl(nla_get_be32(tb[NFTA_NAT_TYPE]))) { 195 if (tb[NFTA_NAT_FAMILY] == NULL) 198 family = ntohl(nla_get_be32(tb[NFTA_NAT_FAMILY])); 210 if (tb[NFTA_NAT_REG_ADDR_MIN]) 216 if (tb[NFTA_NAT_REG_ADDR_MIN]) { 217 err = nft_parse_register_load(tb[NFTA_NAT_REG_ADDR_MI 171 nft_nat_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument [all...] |
H A D | nft_inner.c | 288 struct nlattr *tb[NFT_EXPR_MAXATTR + 1]; member in struct:nft_expr_info 293 const struct nlattr * const tb[]) 300 if (!tb[NFTA_INNER_FLAGS] || 301 !tb[NFTA_INNER_NUM] || 302 !tb[NFTA_INNER_HDRSIZE] || 303 !tb[NFTA_INNER_TYPE] || 304 !tb[NFTA_INNER_EXPR]) 307 flags = ntohl(nla_get_be32(tb[NFTA_INNER_FLAGS])); 311 num = ntohl(nla_get_be32(tb[NFTA_INNER_NUM])); 315 hdrsize = ntohl(nla_get_be32(tb[NFTA_INNER_HDRSIZ 291 nft_inner_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument [all...] |
H A D | nfnetlink_acct.c | 60 const struct nlattr * const tb[]) 68 if (!tb[NFACCT_NAME]) 71 acct_name = nla_data(tb[NFACCT_NAME]); 101 if (tb[NFACCT_FLAGS]) { 102 flags = ntohl(nla_get_be32(tb[NFACCT_FLAGS])); 109 if ((flags & NFACCT_F_QUOTA) && !tb[NFACCT_QUOTA]) 122 *quota = be64_to_cpu(nla_get_be64(tb[NFACCT_QUOTA])); 126 nla_strscpy(nfacct->name, tb[NFACCT_NAME], NFACCT_NAME_MAX); 128 if (tb[NFACCT_BYTES]) { 130 be64_to_cpu(nla_get_be64(tb[NFACCT_BYTE 59 nfnl_acct_new(struct sk_buff *skb, const struct nfnl_info *info, const struct nlattr * const tb[]) argument 248 struct nlattr *tb[NFACCT_FILTER_MAX + 1]; local 274 nfnl_acct_get(struct sk_buff *skb, const struct nfnl_info *info, const struct nlattr * const tb[]) argument 343 nfnl_acct_del(struct sk_buff *skb, const struct nfnl_info *info, const struct nlattr * const tb[]) argument [all...] |
H A D | nft_compat.c | 202 struct nlattr *tb[NFTA_RULE_COMPAT_MAX+1]; local 207 err = nla_parse_nested_deprecated(tb, NFTA_RULE_COMPAT_MAX, attr, 212 if (!tb[NFTA_RULE_COMPAT_PROTO] || !tb[NFTA_RULE_COMPAT_FLAGS]) 215 flags = ntohl(nla_get_be32(tb[NFTA_RULE_COMPAT_FLAGS])); 222 l4proto = ntohl(nla_get_be32(tb[NFTA_RULE_COMPAT_PROTO])); 244 const struct nlattr * const tb[]) 249 size_t size = XT_ALIGN(nla_len(tb[NFTA_TARGET_INFO])); 255 target_compat_from_user(target, nla_data(tb[NFTA_TARGET_INFO]), info); 498 const struct nlattr * const tb[], 243 nft_target_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument 497 __nft_match_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[], void *info) argument 525 nft_match_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument 532 nft_match_large_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument 681 nfnl_compat_get_rcu(struct sk_buff *skb, const struct nfnl_info *info, const struct nlattr * const tb[]) argument 785 nft_match_select_ops(const struct nft_ctx *ctx, const struct nlattr * const tb[]) argument 866 nft_target_select_ops(const struct nft_ctx *ctx, const struct nlattr * const tb[]) argument [all...] |
H A D | nft_counter.c | 59 static int nft_counter_do_init(const struct nlattr * const tb[], argument 71 if (tb[NFTA_COUNTER_PACKETS]) { 73 be64_to_cpu(nla_get_be64(tb[NFTA_COUNTER_PACKETS])); 75 if (tb[NFTA_COUNTER_BYTES]) { 77 be64_to_cpu(nla_get_be64(tb[NFTA_COUNTER_BYTES])); 85 const struct nlattr * const tb[], 90 return nft_counter_do_init(tb, priv); 214 const struct nlattr * const tb[]) 218 return nft_counter_do_init(tb, priv); 84 nft_counter_obj_init(const struct nft_ctx *ctx, const struct nlattr * const tb[], struct nft_object *obj) argument 212 nft_counter_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument
|
/linux-master/net/ipv6/ila/ |
H A D | ila_lwt.c | 135 struct nlattr *tb[ILA_ATTR_MAX + 1]; local 149 ret = nla_parse_nested_deprecated(tb, ILA_ATTR_MAX, nla, 154 if (!tb[ILA_ATTR_LOCATOR]) 159 if (tb[ILA_ATTR_IDENT_TYPE]) 160 ident_type = nla_get_u8(tb[ILA_ATTR_IDENT_TYPE]); 194 if (tb[ILA_ATTR_HOOK_TYPE]) 195 hook_type = nla_get_u8(tb[ILA_ATTR_HOOK_TYPE]); 208 if (tb[ILA_ATTR_CSUM_MODE]) 209 csum_mode = nla_get_u8(tb[ILA_ATTR_CSUM_MODE]); 236 p->locator.v64 = (__force __be64)nla_get_u64(tb[ILA_ATTR_LOCATO [all...] |
/linux-master/net/core/ |
H A D | fib_rules.c | 215 static struct fib_kuid_range nla_get_kuid_range(struct nlattr **tb) argument 220 in = (struct fib_rule_uid_range *)nla_data(tb[FRA_UID_RANGE]); 418 struct nlattr **tb, 486 if (!ops->compare(r, frh, tb)) 518 struct nlattr *tb[], 528 if (!tb[FRA_SRC] || 530 nla_len(tb[FRA_SRC]) != ops->addr_size) { 536 if (!tb[FRA_DST] || 538 nla_len(tb[FRA_DST]) != ops->addr_size) { 551 if (tb[FRA_PRIORIT 416 rule_find(struct fib_rules_ops *ops, struct fib_rule_hdr *frh, struct nlattr **tb, struct fib_rule *rule, bool user_priority) argument 515 fib_nl2rule(struct sk_buff *skb, struct nlmsghdr *nlh, struct netlink_ext_ack *extack, struct fib_rules_ops *ops, struct nlattr *tb[], struct fib_rule **rule, bool *user_priority) argument 684 rule_exists(struct fib_rules_ops *ops, struct fib_rule_hdr *frh, struct nlattr **tb, struct fib_rule *rule) argument 779 struct nlattr *tb[FRA_MAX + 1]; local 887 struct nlattr *tb[FRA_MAX+1]; local [all...] |
/linux-master/drivers/net/wireless/ath/ath10k/ |
H A D | testmode.c | 100 static int ath10k_tm_cmd_get_version(struct ath10k *ar, struct nlattr *tb[]) argument 226 static int ath10k_tm_cmd_utf_start(struct ath10k *ar, struct nlattr *tb[]) argument 347 static int ath10k_tm_cmd_utf_stop(struct ath10k *ar, struct nlattr *tb[]) argument 371 static int ath10k_tm_cmd_wmi(struct ath10k *ar, struct nlattr *tb[]) argument 385 if (!tb[ATH10K_TM_ATTR_DATA]) { 390 if (!tb[ATH10K_TM_ATTR_WMI_CMDID]) { 395 buf = nla_data(tb[ATH10K_TM_ATTR_DATA]); 396 buf_len = nla_len(tb[ATH10K_TM_ATTR_DATA]); 397 cmd_id = nla_get_u32(tb[ATH10K_TM_ATTR_WMI_CMDID]); 431 struct nlattr *tb[ATH10K_TM_ATTR_MA local [all...] |
/linux-master/net/sched/ |
H A D | act_bpf.c | 187 static int tcf_bpf_init_from_ops(struct nlattr **tb, struct tcf_bpf_cfg *cfg) argument 195 bpf_num_ops = nla_get_u16(tb[TCA_ACT_BPF_OPS_LEN]); 200 if (bpf_size != nla_len(tb[TCA_ACT_BPF_OPS])) 203 bpf_ops = kmemdup(nla_data(tb[TCA_ACT_BPF_OPS]), bpf_size, GFP_KERNEL); 224 static int tcf_bpf_init_from_efd(struct nlattr **tb, struct tcf_bpf_cfg *cfg) argument 230 bpf_fd = nla_get_u32(tb[TCA_ACT_BPF_FD]); 236 if (tb[TCA_ACT_BPF_NAME]) { 237 name = nla_memdup(tb[TCA_ACT_BPF_NAME], GFP_KERNEL); 286 struct nlattr *tb[TCA_ACT_BPF_MAX + 1]; local 298 ret = nla_parse_nested_deprecated(tb, TCA_ACT_BPF_MA [all...] |
H A D | act_police.c | 47 struct nlattr *tb[TCA_POLICE_MAX + 1]; local 62 err = nla_parse_nested_deprecated(tb, TCA_POLICE_MAX, nla, 67 if (tb[TCA_POLICE_TBF] == NULL) 69 size = nla_len(tb[TCA_POLICE_TBF]); 73 parm = nla_data(tb[TCA_POLICE_TBF]); 102 R_tab = qdisc_get_rtab(&parm->rate, tb[TCA_POLICE_RATE], NULL); 108 tb[TCA_POLICE_PEAKRATE], NULL); 122 } else if (tb[TCA_POLICE_AVRATE] && 129 if (tb[TCA_POLICE_RESULT]) { 130 tcfp_result = nla_get_u32(tb[TCA_POLICE_RESUL [all...] |
H A D | sch_codel.c | 107 struct nlattr *tb[TCA_CODEL_MAX + 1]; local 111 err = nla_parse_nested_deprecated(tb, TCA_CODEL_MAX, opt, 118 if (tb[TCA_CODEL_TARGET]) { 119 u32 target = nla_get_u32(tb[TCA_CODEL_TARGET]); 124 if (tb[TCA_CODEL_CE_THRESHOLD]) { 125 u64 val = nla_get_u32(tb[TCA_CODEL_CE_THRESHOLD]); 130 if (tb[TCA_CODEL_INTERVAL]) { 131 u32 interval = nla_get_u32(tb[TCA_CODEL_INTERVAL]); 136 if (tb[TCA_CODEL_LIMIT]) 137 sch->limit = nla_get_u32(tb[TCA_CODEL_LIMI [all...] |
H A D | sch_gred.c | 540 struct nlattr *tb[TCA_GRED_VQ_MAX + 1]; local 543 nla_parse_nested_deprecated(tb, TCA_GRED_VQ_MAX, entry, 546 dp = nla_get_u32(tb[TCA_GRED_VQ_DP]); 548 if (tb[TCA_GRED_VQ_FLAGS]) 549 table->tab[dp]->red_flags = nla_get_u32(tb[TCA_GRED_VQ_FLAGS]); 570 struct nlattr *tb[TCA_GRED_VQ_MAX + 1]; local 574 err = nla_parse_nested_deprecated(tb, TCA_GRED_VQ_MAX, entry, 579 if (!tb[TCA_GRED_VQ_DP]) { 583 dp = nla_get_u32(tb[TCA_GRED_VQ_DP]); 593 if (tb[TCA_GRED_VQ_FLAG 647 struct nlattr *tb[TCA_GRED_MAX + 1]; local 736 struct nlattr *tb[TCA_GRED_MAX + 1]; local [all...] |
/linux-master/include/linux/ |
H A D | thunderbolt.h | 61 * struct tb - main thunderbolt bus structure 75 struct tb { struct 191 * @tb: Pointer to the domain 239 struct tb *tb; member in struct:tb_xdomain 292 struct tb_xdomain *tb_xdomain_find_by_uuid(struct tb *tb, const uuid_t *uuid); 293 struct tb_xdomain *tb_xdomain_find_by_route(struct tb *tb, u64 route); 296 tb_xdomain_find_by_uuid_locked(struct tb *t argument 308 tb_xdomain_find_by_route_locked(struct tb *tb, u64 route) argument [all...] |
/linux-master/drivers/thunderbolt/ |
H A D | nhi.c | 26 #include "tb.h" 967 struct tb *tb = pci_get_drvdata(pdev); local 968 struct tb_nhi *nhi = tb->nhi; 971 ret = tb_domain_suspend_noirq(tb); 976 ret = nhi->ops->suspend_noirq(tb->nhi, wakeup); 992 struct tb *tb = pci_get_drvdata(pdev); local 994 return tb_domain_freeze_noirq(tb); 1000 struct tb *t local 1047 struct tb *tb = pci_get_drvdata(pdev); local 1073 struct tb *tb = pci_get_drvdata(pdev); local 1081 struct tb *tb = pci_get_drvdata(pdev); local 1097 struct tb *tb = pci_get_drvdata(pdev); local 1116 struct tb *tb = pci_get_drvdata(pdev); local 1308 struct tb *tb; local 1333 struct tb *tb; local 1418 struct tb *tb = pci_get_drvdata(pdev); local [all...] |
/linux-master/net/ethtool/ |
H A D | wol.c | 109 struct nlattr **tb = info->attrs; local 115 tb[ETHTOOL_A_WOL_MODES], wol_mode_names, 120 NL_SET_ERR_MSG_ATTR(info->extack, tb[ETHTOOL_A_WOL_MODES], 124 if (tb[ETHTOOL_A_WOL_SOPASS]) { 127 tb[ETHTOOL_A_WOL_SOPASS], 132 tb[ETHTOOL_A_WOL_SOPASS], &mod);
|
H A D | plca.c | 31 static void plca_update_sint(int *dst, struct nlattr **tb, u32 attrid, argument 34 const struct nlattr *attr = tb[attrid]; 146 struct nlattr **tb = info->attrs; local 160 plca_update_sint(&plca_cfg.enabled, tb, ETHTOOL_A_PLCA_ENABLED, &mod); 161 plca_update_sint(&plca_cfg.node_id, tb, ETHTOOL_A_PLCA_NODE_ID, &mod); 162 plca_update_sint(&plca_cfg.node_cnt, tb, ETHTOOL_A_PLCA_NODE_CNT, &mod); 163 plca_update_sint(&plca_cfg.to_tmr, tb, ETHTOOL_A_PLCA_TO_TMR, &mod); 164 plca_update_sint(&plca_cfg.burst_cnt, tb, ETHTOOL_A_PLCA_BURST_CNT, 166 plca_update_sint(&plca_cfg.burst_tmr, tb, ETHTOOL_A_PLCA_BURST_TMR,
|
/linux-master/net/ieee802154/6lowpan/ |
H A D | core.c | 122 static int lowpan_validate(struct nlattr *tb[], struct nlattr *data[], argument 125 if (tb[IFLA_ADDRESS]) { 126 if (nla_len(tb[IFLA_ADDRESS]) != IEEE802154_ADDR_LEN) 133 struct nlattr *tb[], struct nlattr *data[], 143 if (!tb[IFLA_LINK]) 146 wdev = dev_get_by_index(dev_net(ldev), nla_get_u32(tb[IFLA_LINK])); 132 lowpan_newlink(struct net *src_net, struct net_device *ldev, struct nlattr *tb[], struct nlattr *data[], struct netlink_ext_ack *extack) argument
|
/linux-master/lib/crypto/ |
H A D | arc4.c | 41 u32 ty, ta, tb; local 60 tb = S[ty]; 66 b = tb;
|
/linux-master/fs/smb/common/ |
H A D | cifs_arc4.c | 43 u32 ty, ta, tb; local 62 tb = S[ty]; 68 b = tb;
|
/linux-master/include/net/netfilter/ |
H A D | nft_reject.h | 23 const struct nlattr * const tb[]);
|
/linux-master/net/mctp/ |
H A D | device.c | 194 struct nlattr *tb[IFA_MAX + 1]; local 203 rc = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_mctp_policy, 210 if (tb[IFA_LOCAL]) 211 addr = nla_data(tb[IFA_LOCAL]); 212 else if (tb[IFA_ADDRESS]) 213 addr = nla_data(tb[IFA_ADDRESS]); 257 struct nlattr *tb[IFA_MAX + 1]; local 266 rc = nlmsg_parse(nlh, sizeof(*ifm), tb, IFA_MAX, ifa_mctp_policy, 273 if (tb[IFA_LOCAL]) 274 addr = nla_data(tb[IFA_LOCA 399 struct nlattr *tb[IFLA_MCTP_MAX + 1]; local [all...] |
H A D | neigh.c | 122 struct nlattr *tb[NDA_MAX + 1]; local 128 rc = nlmsg_parse(nlh, sizeof(*ndm), tb, NDA_MAX, nd_mctp_policy, 135 if (!tb[NDA_DST]) { 140 if (!tb[NDA_LLADDR]) { 145 eid = nla_get_u8(tb[NDA_DST]); 151 lladdr = nla_data(tb[NDA_LLADDR]); 152 lladdr_len = nla_len(tb[NDA_LLADDR]); 177 struct nlattr *tb[NDA_MAX + 1]; local 184 rc = nlmsg_parse(nlh, sizeof(*ndm), tb, NDA_MAX, nd_mctp_policy, 191 if (!tb[NDA_DS [all...] |
/linux-master/drivers/base/regmap/ |
H A D | regmap-spi-avmm.c | 249 char *tb, *tb_end, *pb, *pb_limit, *pb_eop = NULL; local 253 tb = br->trans_buf; 254 tb_end = tb + br->trans_len; 267 for (; pb < pb_limit && tb < tb_end; pb++) { 269 *pb = *tb++ ^ 0x20; 275 if (tb == tb_end - 1 && !pb_eop) { 285 switch (*tb) { 299 *pb = *tb++; 305 if (tb < tb_end) 364 char *pb, *tb_limit, *tb local [all...] |
/linux-master/drivers/tty/ |
H A D | tty_buffer.c | 310 struct tty_buffer *tb = port->buf.tail; local 315 memcpy(char_buf_ptr(tb, tb->used), chars, space); 318 memcpy(flag_buf_ptr(tb, tb->used), flags, space); 320 } else if (tb->flags) { 321 memset(flag_buf_ptr(tb, tb->used), flags[0], space); 323 /* tb->flags should be available once requested */ 327 tb 359 struct tty_buffer *tb = port->buf.tail; local [all...] |