/linux-master/net/dcb/ |
H A D | dcbnl.c | 242 u32 seq, struct nlattr **tb, struct sk_buff *skb) 244 /* if (!tb[DCB_ATTR_STATE] || !netdev->dcbnl_ops->getstate) */ 253 u32 seq, struct nlattr **tb, struct sk_buff *skb) 261 if (!tb[DCB_ATTR_PFC_CFG]) 268 tb[DCB_ATTR_PFC_CFG], 298 u32 seq, struct nlattr **tb, struct sk_buff *skb) 312 u32 seq, struct nlattr **tb, struct sk_buff *skb) 320 if (!tb[DCB_ATTR_CAP]) 327 tb[DCB_ATTR_CAP], dcbnl_cap_nest, 357 u32 seq, struct nlattr **tb, struc 241 dcbnl_getstate(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 252 dcbnl_getpfccfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 297 dcbnl_getperm_hwaddr(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 311 dcbnl_getcap(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 356 dcbnl_getnumtcs(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 403 dcbnl_setnumtcs(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 437 dcbnl_getpfcstate(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 447 dcbnl_setpfcstate(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 465 dcbnl_getapp(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 535 dcbnl_setapp(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 585 __dcbnl_pg_getcfg(struct net_device *netdev, struct nlmsghdr *nlh, struct nlattr **tb, struct sk_buff *skb, int dir) argument 721 dcbnl_pgtx_getcfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 727 dcbnl_pgrx_getcfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 733 dcbnl_setstate(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 750 dcbnl_setpfccfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 781 dcbnl_setall(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 799 __dcbnl_pg_setcfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb, int dir) argument 892 dcbnl_pgtx_setcfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 898 dcbnl_pgrx_setcfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 904 dcbnl_bcn_getcfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 966 dcbnl_bcn_setcfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 1560 dcbnl_ieee_set(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 1696 dcbnl_ieee_get(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 1707 dcbnl_ieee_del(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 1750 dcbnl_getdcbx(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 1760 dcbnl_setdcbx(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 1777 dcbnl_getfeatcfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 1823 dcbnl_setfeatcfg(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 1861 dcbnl_cee_get(struct net_device *netdev, struct nlmsghdr *nlh, u32 seq, struct nlattr **tb, struct sk_buff *skb) argument 1917 struct nlattr *tb[DCB_ATTR_MAX + 1]; local [all...] |
/linux-master/net/ipv6/ |
H A D | ioam6_iptunnel.c | 107 struct nlattr *tb[IOAM6_IPTUNNEL_MAX + 1]; local 119 err = nla_parse_nested(tb, IOAM6_IPTUNNEL_MAX, nla, 124 if ((!tb[IOAM6_IPTUNNEL_FREQ_K] && tb[IOAM6_IPTUNNEL_FREQ_N]) || 125 (tb[IOAM6_IPTUNNEL_FREQ_K] && !tb[IOAM6_IPTUNNEL_FREQ_N])) { 128 } else if (!tb[IOAM6_IPTUNNEL_FREQ_K] && !tb[IOAM6_IPTUNNEL_FREQ_N]) { 132 freq_k = nla_get_u32(tb[IOAM6_IPTUNNEL_FREQ_K]); 133 freq_n = nla_get_u32(tb[IOAM6_IPTUNNEL_FREQ_ [all...] |
/linux-master/net/sched/ |
H A D | sch_red.c | 234 static int __red_change(struct Qdisc *sch, struct nlattr **tb, argument 247 if (tb[TCA_RED_PARMS] == NULL || 248 tb[TCA_RED_STAB] == NULL) 251 max_P = tb[TCA_RED_MAX_P] ? nla_get_u32(tb[TCA_RED_MAX_P]) : 0; 253 ctl = nla_data(tb[TCA_RED_PARMS]); 254 stab = nla_data(tb[TCA_RED_STAB]); 260 tb[TCA_RED_FLAGS], TC_RED_SUPPORTED_FLAGS, 339 struct nlattr *tb[TCA_RED_MAX + 1]; local 349 err = nla_parse_nested_deprecated(tb, TCA_RED_MA 373 struct nlattr *tb[TCA_RED_MAX + 1]; local [all...] |
H A D | sch_pie.c | 142 struct nlattr *tb[TCA_PIE_MAX + 1]; local 146 err = nla_parse_nested_deprecated(tb, TCA_PIE_MAX, opt, pie_policy, 154 if (tb[TCA_PIE_TARGET]) { 156 u32 target = nla_get_u32(tb[TCA_PIE_TARGET]); 163 if (tb[TCA_PIE_TUPDATE]) 165 usecs_to_jiffies(nla_get_u32(tb[TCA_PIE_TUPDATE])); 167 if (tb[TCA_PIE_LIMIT]) { 168 u32 limit = nla_get_u32(tb[TCA_PIE_LIMIT]); 174 if (tb[TCA_PIE_ALPHA]) 175 q->params.alpha = nla_get_u32(tb[TCA_PIE_ALPH [all...] |
H A D | act_simple.c | 95 struct nlattr *tb[TCA_DEF_MAX + 1]; local 106 err = nla_parse_nested_deprecated(tb, TCA_DEF_MAX, nla, simple_policy, 111 if (tb[TCA_DEF_PARMS] == NULL) 114 parm = nla_data(tb[TCA_DEF_PARMS]); 123 if (tb[TCA_DEF_DATA] == NULL) { 145 err = alloc_defdata(d, tb[TCA_DEF_DATA]); 157 err = reset_policy(*a, tb[TCA_DEF_DATA], parm, tp, extack);
|
H A D | cls_flow.c | 398 struct nlattr *tb[TCA_FLOW_MAX + 1]; local 409 err = nla_parse_nested_deprecated(tb, TCA_FLOW_MAX, opt, flow_policy, 414 if (tb[TCA_FLOW_BASECLASS]) { 415 baseclass = nla_get_u32(tb[TCA_FLOW_BASECLASS]); 420 if (tb[TCA_FLOW_KEYS]) { 421 keymask = nla_get_u32(tb[TCA_FLOW_KEYS]); 439 err = tcf_em_tree_validate(tp, tb[TCA_FLOW_EMATCHES], &fnew->ematches); 447 err = tcf_exts_validate(net, tp, tb, tca[TCA_RATE], &fnew->exts, flags, 473 if (tb[TCA_FLOW_MODE]) 474 mode = nla_get_u32(tb[TCA_FLOW_MOD [all...] |
H A D | sch_fq.c | 991 struct nlattr *tb[TCA_FQ_MAX + 1]; local 996 err = nla_parse_nested_deprecated(tb, TCA_FQ_MAX, opt, fq_policy, 1005 if (tb[TCA_FQ_BUCKETS_LOG]) { 1006 u32 nval = nla_get_u32(tb[TCA_FQ_BUCKETS_LOG]); 1013 if (tb[TCA_FQ_PLIMIT]) 1014 sch->limit = nla_get_u32(tb[TCA_FQ_PLIMIT]); 1016 if (tb[TCA_FQ_FLOW_PLIMIT]) 1017 q->flow_plimit = nla_get_u32(tb[TCA_FQ_FLOW_PLIMIT]); 1019 if (tb[TCA_FQ_QUANTUM]) { 1020 u32 quantum = nla_get_u32(tb[TCA_FQ_QUANTU [all...] |
/linux-master/net/ethtool/ |
H A D | channels.c | 116 struct nlattr **tb = info->attrs; local 125 ethnl_update_u32(&channels.rx_count, tb[ETHTOOL_A_CHANNELS_RX_COUNT], 127 ethnl_update_u32(&channels.tx_count, tb[ETHTOOL_A_CHANNELS_TX_COUNT], 130 tb[ETHTOOL_A_CHANNELS_OTHER_COUNT], &mod); 132 tb[ETHTOOL_A_CHANNELS_COMBINED_COUNT], &mod_combined); 149 NL_SET_ERR_MSG_ATTR(info->extack, tb[err_attr], 164 NL_SET_ERR_MSG_ATTR(info->extack, tb[err_attr],
|
H A D | mm.c | 188 struct nlattr **tb = info->attrs; local 199 ethnl_update_bool(&cfg.verify_enabled, tb[ETHTOOL_A_MM_VERIFY_ENABLED], 201 ethnl_update_u32(&cfg.verify_time, tb[ETHTOOL_A_MM_VERIFY_TIME], &mod); 202 ethnl_update_bool(&cfg.tx_enabled, tb[ETHTOOL_A_MM_TX_ENABLED], &mod); 203 ethnl_update_bool(&cfg.pmac_enabled, tb[ETHTOOL_A_MM_PMAC_ENABLED], 206 tb[ETHTOOL_A_MM_TX_MIN_FRAG_SIZE], &mod); 212 NL_SET_ERR_MSG_ATTR(extack, tb[ETHTOOL_A_MM_VERIFY_TIME],
|
H A D | pause.c | 31 struct nlattr **tb, 37 if (tb[ETHTOOL_A_PAUSE_STATS_SRC]) { 44 src = nla_get_u32(tb[ETHTOOL_A_PAUSE_STATS_SRC]); 187 struct nlattr **tb = info->attrs; local 193 ethnl_update_bool32(¶ms.autoneg, tb[ETHTOOL_A_PAUSE_AUTONEG], &mod); 194 ethnl_update_bool32(¶ms.rx_pause, tb[ETHTOOL_A_PAUSE_RX], &mod); 195 ethnl_update_bool32(¶ms.tx_pause, tb[ETHTOOL_A_PAUSE_TX], &mod); 30 pause_parse_request(struct ethnl_req_info *req_base, struct nlattr **tb, struct netlink_ext_ack *extack) argument
|
/linux-master/net/netfilter/ |
H A D | nft_redir.c | 46 const struct nlattr * const tb[]) 53 if (tb[NFTA_REDIR_REG_PROTO_MIN]) { 54 err = nft_parse_register_load(tb[NFTA_REDIR_REG_PROTO_MIN], 59 if (tb[NFTA_REDIR_REG_PROTO_MAX]) { 60 err = nft_parse_register_load(tb[NFTA_REDIR_REG_PROTO_MAX], 72 if (tb[NFTA_REDIR_FLAGS]) 73 priv->flags = ntohl(nla_get_be32(tb[NFTA_REDIR_FLAGS])); 44 nft_redir_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument
|
H A D | nft_masq.c | 45 const struct nlattr * const tb[]) 51 if (tb[NFTA_MASQ_FLAGS]) 52 priv->flags = ntohl(nla_get_be32(tb[NFTA_MASQ_FLAGS])); 54 if (tb[NFTA_MASQ_REG_PROTO_MIN]) { 55 err = nft_parse_register_load(tb[NFTA_MASQ_REG_PROTO_MIN], 60 if (tb[NFTA_MASQ_REG_PROTO_MAX]) { 61 err = nft_parse_register_load(tb[NFTA_MASQ_REG_PROTO_MAX], 43 nft_masq_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument
|
H A D | nft_fib.c | 70 const struct nlattr * const tb[]) 76 if (!tb[NFTA_FIB_DREG] || !tb[NFTA_FIB_RESULT] || !tb[NFTA_FIB_FLAGS]) 79 priv->flags = ntohl(nla_get_be32(tb[NFTA_FIB_FLAGS])); 93 priv->result = ntohl(nla_get_be32(tb[NFTA_FIB_RESULT])); 113 err = nft_parse_register_store(ctx, tb[NFTA_FIB_DREG], &priv->dreg, 69 nft_fib_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument
|
H A D | nft_socket.c | 148 const struct nlattr * const tb[]) 153 if (!tb[NFTA_SOCKET_DREG] || !tb[NFTA_SOCKET_KEY]) 167 priv->key = ntohl(nla_get_be32(tb[NFTA_SOCKET_KEY])); 180 if (!tb[NFTA_SOCKET_LEVEL]) 183 level = ntohl(nla_get_be32(tb[NFTA_SOCKET_LEVEL])); 197 return nft_parse_register_store(ctx, tb[NFTA_SOCKET_DREG], &priv->dreg, 146 nft_socket_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument
|
H A D | nft_connlimit.c | 60 const struct nlattr * const tb[], 67 if (!tb[NFTA_CONNLIMIT_COUNT]) 70 limit = ntohl(nla_get_be32(tb[NFTA_CONNLIMIT_COUNT])); 72 if (tb[NFTA_CONNLIMIT_FLAGS]) { 73 flags = ntohl(nla_get_be32(tb[NFTA_CONNLIMIT_FLAGS])); 132 const struct nlattr * const tb[], 137 return nft_connlimit_do_init(ctx, tb, priv); 198 const struct nlattr * const tb[]) 202 return nft_connlimit_do_init(ctx, tb, priv); 59 nft_connlimit_do_init(const struct nft_ctx *ctx, const struct nlattr * const tb[], struct nft_connlimit *priv) argument 131 nft_connlimit_obj_init(const struct nft_ctx *ctx, const struct nlattr * const tb[], struct nft_object *obj) argument 196 nft_connlimit_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument
|
H A D | nft_xfrm.c | 35 const struct nlattr * const tb[]) 42 if (!tb[NFTA_XFRM_KEY] || !tb[NFTA_XFRM_DIR] || !tb[NFTA_XFRM_DREG]) 54 priv->key = ntohl(nla_get_be32(tb[NFTA_XFRM_KEY])); 72 dir = nla_get_u8(tb[NFTA_XFRM_DIR]); 82 if (tb[NFTA_XFRM_SPNUM]) 83 spnum = ntohl(nla_get_be32(tb[NFTA_XFRM_SPNUM])); 91 return nft_parse_register_store(ctx, tb[NFTA_XFRM_DREG], &priv->dreg, 33 nft_xfrm_get_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument
|
H A D | nft_dup_netdev.c | 36 const struct nlattr * const tb[]) 40 if (tb[NFTA_DUP_SREG_DEV] == NULL) 43 return nft_parse_register_load(tb[NFTA_DUP_SREG_DEV], &priv->sreg_dev, 34 nft_dup_netdev_init(const struct nft_ctx *ctx, const struct nft_expr *expr, const struct nlattr * const tb[]) argument
|
/linux-master/net/mpls/ |
H A D | mpls_iptunnel.c | 169 struct nlattr *tb[MPLS_IPTUNNEL_MAX + 1]; local 174 ret = nla_parse_nested_deprecated(tb, MPLS_IPTUNNEL_MAX, nla, 179 if (!tb[MPLS_IPTUNNEL_DST]) { 185 if (nla_get_labels(tb[MPLS_IPTUNNEL_DST], MAX_NEW_LABELS, 195 ret = nla_get_labels(tb[MPLS_IPTUNNEL_DST], n_labels, 203 if (tb[MPLS_IPTUNNEL_TTL]) { 204 tun_encap_info->default_ttl = nla_get_u8(tb[MPLS_IPTUNNEL_TTL]);
|
/linux-master/include/linux/netfilter/ipset/ |
H A D | ip_set.h | 163 int (*uadt)(struct ip_set *set, struct nlattr *tb[], 231 struct nlattr *tb[], u32 flags); 335 extern size_t ip_set_elem_len(struct ip_set *set, struct nlattr *tb[], 337 extern int ip_set_get_extensions(struct ip_set *set, struct nlattr *tb[], 377 ip_set_attr_netorder(struct nlattr *tb[], int type) argument 379 return tb[type] && (tb[type]->nla_type & NLA_F_NET_BYTEORDER); 383 ip_set_optattr_netorder(struct nlattr *tb[], int type) argument 385 return !tb[type] || (tb[typ 468 ip_set_timeout_uget(struct nlattr *tb) argument [all...] |
/linux-master/arch/sparc/kernel/ |
H A D | irq_64.c | 1005 struct trap_per_cpu *tb = &trap_block[this_cpu]; local 1007 register_one_mondo(tb->cpu_mondo_pa, HV_CPU_QUEUE_CPU_MONDO, 1008 tb->cpu_mondo_qmask); 1009 register_one_mondo(tb->dev_mondo_pa, HV_CPU_QUEUE_DEVICE_MONDO, 1010 tb->dev_mondo_qmask); 1011 register_one_mondo(tb->resum_mondo_pa, HV_CPU_QUEUE_RES_ERROR, 1012 tb->resum_qmask); 1013 register_one_mondo(tb->nonresum_mondo_pa, HV_CPU_QUEUE_NONRES_ERROR, 1014 tb->nonresum_qmask); 1036 static void __init init_cpu_send_mondo_info(struct trap_per_cpu *tb) argument 1069 struct trap_per_cpu *tb = &trap_block[cpu]; local 1086 struct trap_per_cpu *tb = &trap_block[cpu]; local [all...] |
H A D | smp_64.c | 295 struct trap_per_cpu *tb; local 314 tb = &trap_block[cpu]; 316 hdesc->fault_info_va = (unsigned long) &tb->fault_info; 317 hdesc->fault_info_pa = kimage_addr_to_ra(&tb->fault_info); 462 static void spitfire_xcall_deliver(struct trap_per_cpu *tb, int cnt) argument 470 cpu_list = __va(tb->cpu_list_pa); 471 mondo = __va(tb->cpu_mondo_block_pa); 483 static void cheetah_xcall_deliver(struct trap_per_cpu *tb, int cnt) argument 489 cpu_list = __va(tb->cpu_list_pa); 490 mondo = __va(tb 645 hypervisor_xcall_deliver(struct trap_per_cpu *tb, int cnt) argument 788 struct trap_per_cpu *tb; local 1313 struct trap_per_cpu *tb = &trap_block[cpu]; local [all...] |
/linux-master/drivers/thunderbolt/ |
H A D | retimer.c | 15 #include "tb.h" 43 if (!mutex_trylock(&rt->tb->lock)) { 49 mutex_unlock(&rt->tb->lock); 64 if (!mutex_trylock(&rt->tb->lock)) 68 mutex_unlock(&rt->tb->lock); 176 if (!mutex_trylock(&rt->tb->lock)) 186 mutex_unlock(&rt->tb->lock); 249 if (!mutex_trylock(&rt->tb->lock)) { 296 mutex_unlock(&rt->tb->lock); 313 if (!mutex_trylock(&rt->tb [all...] |
/linux-master/net/core/ |
H A D | net_namespace.c | 773 struct nlattr *tb[NETNSA_MAX + 1]; local 778 err = nlmsg_parse_deprecated(nlh, sizeof(struct rtgenmsg), tb, 782 if (!tb[NETNSA_NSID]) { 786 nsid = nla_get_s32(tb[NETNSA_NSID]); 788 if (tb[NETNSA_PID]) { 789 peer = get_net_ns_by_pid(nla_get_u32(tb[NETNSA_PID])); 790 nla = tb[NETNSA_PID]; 791 } else if (tb[NETNSA_FD]) { 792 peer = get_net_ns_by_fd(nla_get_u32(tb[NETNSA_FD])); 793 nla = tb[NETNSA_F 876 rtnl_net_valid_getid_req(struct sk_buff *skb, const struct nlmsghdr *nlh, struct nlattr **tb, struct netlink_ext_ack *extack) argument 917 struct nlattr *tb[NETNSA_MAX + 1]; local 1026 struct nlattr *tb[NETNSA_MAX + 1]; local [all...] |
/linux-master/drivers/net/wireless/ath/ath12k/ |
H A D | wmi.c | 349 const void **tb = data; local 352 tb[tag] = ptr; 357 static int ath12k_wmi_tlv_parse(struct ath12k_base *ar, const void **tb, argument 361 (void *)tb); 368 const void **tb; local 371 tb = kcalloc(WMI_TAG_MAX, sizeof(*tb), gfp); 372 if (!tb) 375 ret = ath12k_wmi_tlv_parse(ab, tb, skb->data, skb->len); 377 kfree(tb); 4441 const void **tb; local 4512 const void **tb; local 4805 const void **tb; local 4837 const void **tb; local 4865 const void **tb; local 4893 const void **tb; local 5037 const void **tb; local 5218 const void **tb; local 5251 const void **tb; local 5278 const void **tb; local 5330 const void **tb; local 5369 const void **tb; local 5409 const void **tb; local 5440 const void **tb; local 5469 const void **tb; local 6472 const void **tb; local 6536 const void **tb; local 6570 const void **tb; local 6650 const void **tb; local 6680 const void **tb; local 6712 const void **tb; local 6763 const void **tb; local 6802 const void **tb; local 6831 const void **tb; local [all...] |
/linux-master/net/can/ |
H A D | gw.c | 803 struct nlattr *tb[CGW_MAX + 1]; local 811 err = nlmsg_parse_deprecated(nlh, sizeof(struct rtcanmsg), tb, 816 if (tb[CGW_LIM_HOPS]) { 817 *limhops = nla_get_u8(tb[CGW_LIM_HOPS]); 827 if (tb[CGW_FDMOD_AND]) { 828 nla_memcpy(&mb, tb[CGW_FDMOD_AND], CGW_FDMODATTR_LEN); 846 if (tb[CGW_FDMOD_OR]) { 847 nla_memcpy(&mb, tb[CGW_FDMOD_OR], CGW_FDMODATTR_LEN); 865 if (tb[CGW_FDMOD_XOR]) { 866 nla_memcpy(&mb, tb[CGW_FDMOD_XO [all...] |