/linux-master/drivers/gpu/drm/amd/amdgpu/ |
H A D | amdgpu_xgmi.h | 38 struct task_barrier tb; member in struct:amdgpu_hive_info
|
/linux-master/net/ethtool/ |
H A D | strset.c | 163 struct nlattr *tb[ARRAY_SIZE(get_stringset_policy)]; local 166 ret = nla_parse_nested(tb, ARRAY_SIZE(get_stringset_policy) - 1, nest, 170 if (NL_REQ_ATTR_CHECK(extack, nest, tb, ETHTOOL_A_STRINGSET_ID)) 173 *val = nla_get_u32(tb[ETHTOOL_A_STRINGSET_ID]); 182 struct nlattr **tb, 186 struct nlattr *nest = tb[ETHTOOL_A_STRSET_STRINGSETS]; 198 req_info->counts_only = tb[ETHTOOL_A_STRSET_COUNTS_ONLY]; 181 strset_parse_request(struct ethnl_req_info *req_base, struct nlattr **tb, struct netlink_ext_ack *extack) argument
|
H A D | fec.c | 241 struct nlattr **tb = info->attrs; local 255 tb[ETHTOOL_A_FEC_MODES], 259 ethnl_update_u8(&fec_auto, tb[ETHTOOL_A_FEC_AUTO], &mod); 265 NL_SET_ERR_MSG_ATTR(info->extack, tb[ETHTOOL_A_FEC_MODES], 270 NL_SET_ERR_MSG_ATTR(info->extack, tb[ETHTOOL_A_FEC_MODES],
|
/linux-master/net/sched/ |
H A D | cls_matchall.c | 169 struct nlattr *tb[TCA_MATCHALL_MAX + 1]; local 181 err = nla_parse_nested_deprecated(tb, TCA_MATCHALL_MAX, 186 if (tb[TCA_MATCHALL_FLAGS]) { 187 userflags = nla_get_u32(tb[TCA_MATCHALL_FLAGS]); 210 err = tcf_exts_validate_ex(net, tp, tb, tca[TCA_RATE], 215 if (tb[TCA_MATCHALL_CLASSID]) { 216 new->res.classid = nla_get_u32(tb[TCA_MATCHALL_CLASSID]);
|
H A D | sch_ets.c | 204 struct nlattr *tb[TCA_ETS_MAX + 1]; local 221 err = nla_parse_nested(tb, TCA_ETS_MAX, opt, ets_class_policy, extack); 225 if (!tb[TCA_ETS_QUANTA_BAND]) 234 err = ets_quantum_parse(sch, tb[TCA_ETS_QUANTA_BAND], &quantum, 582 struct nlattr *tb[TCA_ETS_MAX + 1]; local 590 err = nla_parse_nested(tb, TCA_ETS_MAX, opt, ets_policy, extack); 594 if (!tb[TCA_ETS_NBANDS]) { 598 nbands = nla_get_u8(tb[TCA_ETS_NBANDS]); 606 if (tb[TCA_ETS_NSTRICT]) { 607 nstrict = nla_get_u8(tb[TCA_ETS_NSTRIC [all...] |
H A D | act_mirred.c | 110 struct nlattr *tb[TCA_MIRRED_MAX + 1]; local 123 ret = nla_parse_nested_deprecated(tb, TCA_MIRRED_MAX, nla, 127 if (!tb[TCA_MIRRED_PARMS]) { 131 parm = nla_data(tb[TCA_MIRRED_PARMS]); 140 if (tb[TCA_MIRRED_BLOCKID] && parm->ifindex) { 167 if (!parm->ifindex && !tb[TCA_MIRRED_BLOCKID]) { 209 } else if (tb[TCA_MIRRED_BLOCKID]) { 212 m->tcfm_blockid = nla_get_u32(tb[TCA_MIRRED_BLOCKID]);
|
/linux-master/net/core/ |
H A D | page_pool_user.c | 176 struct nlattr *tb[ARRAY_SIZE(netdev_page_pool_info_nl_policy)]; local 185 err = nla_parse_nested(tb, ARRAY_SIZE(tb) - 1, nest, 191 if (NL_REQ_ATTR_CHECK(info->extack, nest, tb, NETDEV_A_PAGE_POOL_ID)) 193 if (tb[NETDEV_A_PAGE_POOL_IFINDEX]) { 195 tb[NETDEV_A_PAGE_POOL_IFINDEX], 200 id = nla_get_uint(tb[NETDEV_A_PAGE_POOL_ID]);
|
/linux-master/arch/powerpc/boot/ |
H A D | devtree.c | 63 void dt_fixup_cpu_clocks(u32 cpu, u32 tb, u32 bus) argument 68 printf("CPU timebase-frequency <- 0x%x (%dMHz)\n\r", tb, MHZ(tb)); 74 setprop_val(devp, "timebase-frequency", cpu_to_be32(tb)); 79 timebase_period_ns = 1000000000 / tb;
|
/linux-master/net/netfilter/ |
H A D | nf_conntrack_proto_gre.c | 248 static int gre_timeout_nlattr_to_obj(struct nlattr *tb[], argument 260 if (tb[CTA_TIMEOUT_GRE_UNREPLIED]) { 262 ntohl(nla_get_be32(tb[CTA_TIMEOUT_GRE_UNREPLIED])) * HZ; 264 if (tb[CTA_TIMEOUT_GRE_REPLIED]) { 266 ntohl(nla_get_be32(tb[CTA_TIMEOUT_GRE_REPLIED])) * HZ;
|
H A D | nf_conntrack_netlink.c | 900 struct nlattr *tb[CTA_FILTER_MAX + 1]; local 903 ret = nla_parse_nested(tb, CTA_FILTER_MAX, attr, cta_filter_nla_policy, 908 if (tb[CTA_FILTER_ORIG_FLAGS]) { 909 filter->orig_flags = nla_get_u32(tb[CTA_FILTER_ORIG_FLAGS]); 914 if (tb[CTA_FILTER_REPLY_FLAGS]) { 915 filter->reply_flags = nla_get_u32(tb[CTA_FILTER_REPLY_FLAGS]); 1284 static int ipv4_nlattr_to_tuple(struct nlattr *tb[], argument 1289 if (!tb[CTA_IP_V4_SRC]) 1292 t->src.u3.ip = nla_get_in_addr(tb[CTA_IP_V4_SRC]); 1296 if (!tb[CTA_IP_V4_DS 1305 ipv6_nlattr_to_tuple(struct nlattr *tb[], struct nf_conntrack_tuple *t, u_int32_t flags) argument 1330 struct nlattr *tb[CTA_IP_MAX+1]; local 1359 struct nlattr *tb[CTA_PROTO_MAX+1]; local 1450 struct nlattr *tb[CTA_TUPLE_MAX+1]; local 1523 struct nlattr *tb[CTA_HELP_MAX+1]; local 2029 struct nlattr *tb[CTA_PROTOINFO_MAX+1]; local 2128 struct nlattr *tb[CTA_SYNPROXY_MAX + 1]; local 3483 struct nlattr *tb[CTA_EXPECT_NAT_MAX+1]; local [all...] |
/linux-master/net/ipv6/ |
H A D | rpl_iptunnel.c | 72 struct nlattr *tb[RPL_IPTUNNEL_MAX + 1]; local 81 err = nla_parse_nested(tb, RPL_IPTUNNEL_MAX, nla, 86 if (!tb[RPL_IPTUNNEL_SRH]) 89 srh = nla_data(tb[RPL_IPTUNNEL_SRH]); 90 srh_len = nla_len(tb[RPL_IPTUNNEL_SRH]);
|
H A D | ip6_fib.c | 220 static void fib6_link_table(struct net *net, struct fib6_table *tb) argument 228 spin_lock_init(&tb->tb6_lock); 229 h = tb->tb6_id & (FIB6_TABLE_HASHSZ - 1); 235 hlist_add_head_rcu(&tb->tb6_hlist, &net->ipv6.fib_table_hash[h]); 259 struct fib6_table *tb; local 263 tb = fib6_get_table(net, id); 264 if (tb) 265 return tb; 267 tb = fib6_alloc_table(net, id); 268 if (tb) 277 struct fib6_table *tb; local 355 struct fib6_table *tb; local 469 fib6_table_dump(struct net *net, struct fib6_table *tb, struct fib6_walker *w) argument 502 struct fib6_table *tb; local 633 struct fib6_table *tb; local 2485 struct fib6_table *tb; local [all...] |
/linux-master/net/netfilter/ipset/ |
H A D | ip_set_hash_gen.h | 1433 IPSET_TOKEN(MTYPE, _uadt)(struct ip_set *set, struct nlattr *tb[], 1459 struct nlattr *tb[], u32 flags) 1487 if (unlikely(!ip_set_optattr_netorder(tb, IPSET_ATTR_HASHSIZE) || 1488 !ip_set_optattr_netorder(tb, IPSET_ATTR_MAXELEM) || 1489 !ip_set_optattr_netorder(tb, IPSET_ATTR_TIMEOUT) || 1490 !ip_set_optattr_netorder(tb, IPSET_ATTR_CADT_FLAGS))) 1495 if (unlikely(!ip_set_optattr_netorder(tb, IPSET_ATTR_MARKMASK))) 1499 if (tb[IPSET_ATTR_MARKMASK]) { 1500 markmask = ntohl(nla_get_be32(tb[IPSET_ATTR_MARKMASK])); 1507 if (tb[IPSET_ATTR_NETMAS [all...] |
/linux-master/crypto/ |
H A D | ccm.c | 447 struct rtattr **tb, 458 err = crypto_check_attr_type(tb, CRYPTO_ALG_TYPE_AEAD, &mask); 529 static int crypto_ccm_create(struct crypto_template *tmpl, struct rtattr **tb) argument 535 cipher_name = crypto_attr_alg_name(tb[1]); 547 return crypto_ccm_create_common(tmpl, tb, ctr_name, mac_name); 551 struct rtattr **tb) 556 ctr_name = crypto_attr_alg_name(tb[1]); 560 mac_name = crypto_attr_alg_name(tb[2]); 564 return crypto_ccm_create_common(tmpl, tb, ctr_name, mac_name); 705 struct rtattr **tb) 446 crypto_ccm_create_common(struct crypto_template *tmpl, struct rtattr **tb, const char *ctr_name, const char *mac_name) argument 550 crypto_ccm_base_create(struct crypto_template *tmpl, struct rtattr **tb) argument 704 crypto_rfc4309_create(struct crypto_template *tmpl, struct rtattr **tb) argument 858 cbcmac_create(struct crypto_template *tmpl, struct rtattr **tb) argument [all...] |
H A D | hctr2.c | 405 struct rtattr **tb, 419 err = crypto_check_attr_type(tb, CRYPTO_ALG_TYPE_SKCIPHER, &mask); 516 static int hctr2_create_base(struct crypto_template *tmpl, struct rtattr **tb) argument 521 xctr_name = crypto_attr_alg_name(tb[1]); 525 polyval_name = crypto_attr_alg_name(tb[2]); 529 return hctr2_create_common(tmpl, tb, xctr_name, polyval_name); 532 static int hctr2_create(struct crypto_template *tmpl, struct rtattr **tb) argument 537 blockcipher_name = crypto_attr_alg_name(tb[1]); 545 return hctr2_create_common(tmpl, tb, xctr_name, "polyval"); 404 hctr2_create_common(struct crypto_template *tmpl, struct rtattr **tb, const char *xctr_name, const char *polyval_name) argument
|
H A D | echainiv.c | 113 struct rtattr **tb) 118 inst = aead_geniv_alloc(tmpl, tb); 112 echainiv_aead_create(struct crypto_template *tmpl, struct rtattr **tb) argument
|
H A D | keywrap.c | 264 static int crypto_kw_create(struct crypto_template *tmpl, struct rtattr **tb) argument 270 inst = skcipher_alloc_instance_simple(tmpl, tb);
|
/linux-master/net/bridge/ |
H A D | br_mdb.c | 1080 struct nlattr *tb[MDBE_SRCATTR_MAX + 1]; local 1083 err = nla_parse_nested(tb, MDBE_SRCATTR_MAX, src_entry, 1088 if (NL_REQ_ATTR_CHECK(extack, src_entry, tb, MDBE_SRCATTR_ADDRESS)) 1091 if (!is_valid_mdb_source(tb[MDBE_SRCATTR_ADDRESS], proto, extack)) 1095 nla_memcpy(&src->addr.src, tb[MDBE_SRCATTR_ADDRESS], 1096 nla_len(tb[MDBE_SRCATTR_ADDRESS])); 1214 struct nlattr *tb[], u16 nlmsg_flags, 1236 cfg->entry = nla_data(tb[MDBA_SET_ENTRY]); 1265 if (tb[MDBA_SET_ENTRY_ATTRS]) 1266 return br_mdb_config_attrs_init(tb[MDBA_SET_ENTRY_ATTR 1213 br_mdb_config_init(struct br_mdb_config *cfg, struct net_device *dev, struct nlattr *tb[], u16 nlmsg_flags, struct netlink_ext_ack *extack) argument 1279 br_mdb_add(struct net_device *dev, struct nlattr *tb[], u16 nlmsg_flags, struct netlink_ext_ack *extack) argument 1381 br_mdb_del(struct net_device *dev, struct nlattr *tb[], struct netlink_ext_ack *extack) argument 1428 br_mdb_flush_desc_init(struct br_mdb_flush_desc *desc, struct nlattr *tb[], struct netlink_ext_ack *extack) argument 1532 br_mdb_del_bulk(struct net_device *dev, struct nlattr *tb[], struct netlink_ext_ack *extack) argument 1554 br_mdb_get_parse(struct net_device *dev, struct nlattr *tb[], struct br_ip *group, struct netlink_ext_ack *extack) argument 1657 br_mdb_get(struct net_device *dev, struct nlattr *tb[], u32 portid, u32 seq, struct netlink_ext_ack *extack) argument [all...] |
/linux-master/block/partitions/ |
H A D | ldm.c | 355 struct tocblock *tb[4]; local 364 tb[0] = &ldb->toc; 365 tb[1] = kmalloc_array(3, sizeof(*tb[1]), GFP_KERNEL); 366 if (!tb[1]) { 370 tb[2] = (struct tocblock*)((u8*)tb[1] + sizeof(*tb[1])); 371 tb[3] = (struct tocblock*)((u8*)tb[ [all...] |
/linux-master/drivers/net/vxlan/ |
H A D | vxlan_mdb.c | 417 struct nlattr *tb[MDBE_SRCATTR_MAX + 1]; local 421 err = nla_parse_nested(tb, MDBE_SRCATTR_MAX, src_entry, 426 if (NL_REQ_ATTR_CHECK(extack, src_entry, tb, MDBE_SRCATTR_ADDRESS)) 429 if (!vxlan_mdb_is_valid_source(tb[MDBE_SRCATTR_ADDRESS], proto, 437 err = vxlan_nla_get_addr(&src->addr, tb[MDBE_SRCATTR_ADDRESS]); 601 struct net_device *dev, struct nlattr *tb[], 605 struct br_mdb_entry *entry = nla_data(tb[MDBA_SET_ENTRY]); 648 if (NL_REQ_ATTR_CHECK(extack, NULL, tb, MDBA_SET_ENTRY_ATTRS)) { 653 return vxlan_mdb_config_attrs_init(cfg, entry, tb[MDBA_SET_ENTRY_ATTRS], 1281 int vxlan_mdb_add(struct net_device *dev, struct nlattr *tb[], u1 argument 600 vxlan_mdb_config_init(struct vxlan_mdb_config *cfg, struct net_device *dev, struct nlattr *tb[], u16 nlmsg_flags, struct netlink_ext_ack *extack) argument 1299 vxlan_mdb_del(struct net_device *dev, struct nlattr *tb[], struct netlink_ext_ack *extack) argument 1329 vxlan_mdb_flush_desc_init(struct vxlan_dev *vxlan, struct vxlan_mdb_flush_desc *desc, struct nlattr *tb[], struct netlink_ext_ack *extack) argument 1438 vxlan_mdb_del_bulk(struct net_device *dev, struct nlattr *tb[], struct netlink_ext_ack *extack) argument 1463 vxlan_mdb_get_parse(struct net_device *dev, struct nlattr *tb[], struct vxlan_mdb_entry_key *group, struct netlink_ext_ack *extack) argument 1568 vxlan_mdb_get(struct net_device *dev, struct nlattr *tb[], u32 portid, u32 seq, struct netlink_ext_ack *extack) argument [all...] |
/linux-master/arch/powerpc/kvm/ |
H A D | book3s_hv_p9_entry.c | 224 u64 tb = mftb() - vc->tb_offset_applied; local 232 vcpu->arch.cur_tb_start = tb; 237 delta = tb - prev_tb; 532 int kvmhv_vcpu_entry_p9(struct kvm_vcpu *vcpu, u64 time_limit, unsigned long lpcr, u64 *tb) argument 554 hdec = time_limit - *tb; 597 u64 new_tb = *tb + vc->tb_offset; 603 *tb = new_tb; 711 mtspr(SPRN_DEC, vcpu->arch.dec_expires - *tb); 848 *tb = mftb(); 849 vcpu->arch.dec_expires = dec + *tb; [all...] |
H A D | book3s_64_mmu.c | 41 if (vcpu->arch.slb[i].tb) 55 vcpu->arch.slb[i].tb ? 't' : ' ', 65 return slbe->tb ? SID_SHIFT_1T : SID_SHIFT; 129 ssize = slbe->tb ? MMU_SEGSIZE_1T : MMU_SEGSIZE_256M; 233 if (slbe->tb) 381 slbe->tb = (rs & SLB_VSID_B_1T) ? 1 : 0; 382 slbe->esid = slbe->tb ? esid_1t : esid; 587 if (slb->tb) {
|
/linux-master/drivers/net/wwan/ |
H A D | wwan_core.c | 910 static int wwan_rtnl_validate(struct nlattr *tb[], struct nlattr *data[], argument 916 if (!tb[IFLA_PARENT_DEV_NAME]) 927 static struct net_device *wwan_rtnl_alloc(struct nlattr *tb[], argument 933 const char *devname = nla_data(tb[IFLA_PARENT_DEV_NAME]); 963 struct nlattr *tb[], struct nlattr *data[], 1054 struct nlattr *tb[IFLA_MAX + 1], *linkinfo[IFLA_INFO_MAX + 1]; local 1073 tb[IFLA_LINKINFO] = nla_nest_start(msg, IFLA_LINKINFO); 1074 if (!tb[IFLA_LINKINFO]) 1082 nla_nest_end(msg, tb[IFLA_LINKINFO]); 1087 nlmsg_parse_deprecated(nlh, 0, tb, IFLA_MA 962 wwan_rtnl_newlink(struct net *src_net, struct net_device *dev, struct nlattr *tb[], struct nlattr *data[], struct netlink_ext_ack *extack) argument [all...] |
/linux-master/drivers/media/platform/nvidia/tegra-vde/ |
H A D | h264.c | 702 struct tegra_m2m_buffer *tb = vb_to_tegra_buf(vb); local 728 if (!tb->aux || tb->aux->size < csize) { 730 tb->aux ? tb->aux->size : -1, csize); 743 if (tb->b_frame) 747 vde->frames[id].y_addr = tb->dma_addr[0]; 748 vde->frames[id].cb_addr = tb->dma_addr[1]; 749 vde->frames[id].cr_addr = tb->dma_addr[2]; 750 vde->frames[id].aux_addr = tb 764 struct tegra_m2m_buffer *tb = vb_to_tegra_buf(&dst->vb2_buf); local [all...] |
/linux-master/drivers/net/ |
H A D | netkit.c | 281 static int netkit_check_policy(int policy, struct nlattr *tb, argument 289 NL_SET_ERR_MSG_ATTR(extack, tb, 295 static int netkit_check_mode(int mode, struct nlattr *tb, argument 303 NL_SET_ERR_MSG_ATTR(extack, tb, 309 static int netkit_validate(struct nlattr *tb[], struct nlattr *data[], argument 312 struct nlattr *attr = tb[IFLA_ADDRESS]; 326 struct nlattr *tb[], struct nlattr *data[], 329 struct nlattr *peer_tb[IFLA_MAX + 1], **tbp = tb, *attr; 384 (tb[IFLA_ADDRESS] || tbp[IFLA_ADDRESS])) 423 if (mode == NETKIT_L2 && !tb[IFLA_ADDRES 325 netkit_new_link(struct net *src_net, struct net_device *dev, struct nlattr *tb[], struct nlattr *data[], struct netlink_ext_ack *extack) argument 850 netkit_change_link(struct net_device *dev, struct nlattr *tb[], struct nlattr *data[], struct netlink_ext_ack *extack) argument [all...] |