/linux-master/include/asm-generic/ |
H A D | percpu.h | 65 #define raw_cpu_generic_read(pcp) \ 67 *raw_cpu_ptr(&(pcp)); \ 70 #define raw_cpu_generic_to_op(pcp, val, op) \ 72 *raw_cpu_ptr(&(pcp)) op val; \ 75 #define raw_cpu_generic_add_return(pcp, val) \ 77 typeof(pcp) *__p = raw_cpu_ptr(&(pcp)); \ 83 #define raw_cpu_generic_xchg(pcp, nval) \ 85 typeof(pcp) *__p = raw_cpu_ptr(&(pcp)); \ [all...] |
/linux-master/arch/s390/include/asm/ |
H A D | percpu.h | 27 #define arch_this_cpu_to_op_simple(pcp, val, op) \ 29 typedef typeof(pcp) pcp_op_T__; \ 33 ptr__ = raw_cpu_ptr(&(pcp)); \ 44 #define this_cpu_add_1(pcp, val) arch_this_cpu_to_op_simple(pcp, val, +) 45 #define this_cpu_add_2(pcp, val) arch_this_cpu_to_op_simple(pcp, val, +) 46 #define this_cpu_add_return_1(pcp, val) arch_this_cpu_to_op_simple(pcp, val, +) 47 #define this_cpu_add_return_2(pcp, va [all...] |
/linux-master/arch/loongarch/include/asm/ |
H A D | percpu.h | 161 #define _protect_cmpxchg_local(pcp, o, n) \ 163 typeof(*raw_cpu_ptr(&(pcp))) __ret; \ 165 __ret = cmpxchg_local(raw_cpu_ptr(&(pcp)), o, n); \ 170 #define _percpu_read(pcp) \ 172 typeof(pcp) __retval; \ 173 __retval = (typeof(pcp))__percpu_read(&(pcp), sizeof(pcp)); \ 177 #define _percpu_write(pcp, val) \ 179 __percpu_write(&(pcp), (unsigne [all...] |
/linux-master/arch/arm64/include/asm/ |
H A D | percpu.h | 149 #define _pcp_protect(op, pcp, ...) \ 152 op(raw_cpu_ptr(&(pcp)), __VA_ARGS__); \ 156 #define _pcp_protect_return(op, pcp, args...) \ 158 typeof(pcp) __retval; \ 160 __retval = (typeof(pcp))op(raw_cpu_ptr(&(pcp)), ##args); \ 165 #define this_cpu_read_1(pcp) \ 166 _pcp_protect_return(__percpu_read_8, pcp) 167 #define this_cpu_read_2(pcp) \ 168 _pcp_protect_return(__percpu_read_16, pcp) [all...] |
/linux-master/arch/x86/include/asm/ |
H A D | percpu.h | 316 #define raw_cpu_cmpxchg64(pcp, oval, nval) percpu_cmpxchg64_op(8, , pcp, oval, nval) 317 #define this_cpu_cmpxchg64(pcp, oval, nval) percpu_cmpxchg64_op(8, volatile, pcp, oval, nval) 349 #define raw_cpu_try_cmpxchg64(pcp, ovalp, nval) percpu_try_cmpxchg64_op(8, , pcp, ovalp, nval) 350 #define this_cpu_try_cmpxchg64(pcp, ovalp, nval) percpu_try_cmpxchg64_op(8, volatile, pcp, ovalp, nval) 354 #define raw_cpu_cmpxchg64(pcp, oval, nval) percpu_cmpxchg_op(8, , pcp, ova [all...] |
/linux-master/include/linux/ |
H A D | percpu-defs.h | 408 #define raw_cpu_read(pcp) __pcpu_size_call_return(raw_cpu_read_, pcp) 409 #define raw_cpu_write(pcp, val) __pcpu_size_call(raw_cpu_write_, pcp, val) 410 #define raw_cpu_add(pcp, val) __pcpu_size_call(raw_cpu_add_, pcp, val) 411 #define raw_cpu_and(pcp, val) __pcpu_size_call(raw_cpu_and_, pcp, val) 412 #define raw_cpu_or(pcp, val) __pcpu_size_call(raw_cpu_or_, pcp, va [all...] |
H A D | gfp.h | 339 int decay_pcp_high(struct zone *zone, struct per_cpu_pages *pcp); 340 void drain_zone_pages(struct zone *zone, struct per_cpu_pages *pcp);
|
/linux-master/lib/ |
H A D | percpu_test.c | 4 /* validate @native and @pcp counter values match @expected */ 5 #define CHECK(native, pcp, expected) \ 11 WARN(__this_cpu_read(pcp) != (expected), \ 12 "pcp %ld (0x%lx) != expected %lld (0x%llx)", \ 13 __this_cpu_read(pcp), __this_cpu_read(pcp), \
|
/linux-master/mm/ |
H A D | page_alloc.c | 62 /* Free Page Internal flags: for internal, non-pcp variants of free_pages(). */ 109 * Locking a pcp requires a PCP lookup followed by a spinlock. To avoid 563 if (pcp_allowed_order(order)) /* Via pcp? */ 1189 struct per_cpu_pages *pcp, 1201 count = min(pcp->count, count); 1217 list = &pcp->lists[pindex]; 1228 /* must delete to avoid corrupting pcp list */ 1231 pcp->count -= nr_pages; 2164 int decay_pcp_high(struct zone *zone, struct per_cpu_pages *pcp) argument 2169 high_min = READ_ONCE(pcp 1188 free_pcppages_bulk(struct zone *zone, int count, struct per_cpu_pages *pcp, int pindex) argument 2200 drain_zone_pages(struct zone *zone, struct per_cpu_pages *pcp) argument 2219 struct per_cpu_pages *pcp; local 2292 struct per_cpu_pages *pcp; local 2355 nr_pcp_free(struct per_cpu_pages *pcp, int batch, int high, bool free_high) argument 2380 nr_pcp_high(struct per_cpu_pages *pcp, struct zone *zone, int batch, bool free_high) argument 2428 free_unref_page_commit(struct zone *zone, struct per_cpu_pages *pcp, struct page *page, int migratetype, unsigned int order) argument 2482 struct per_cpu_pages *pcp; local 2524 struct per_cpu_pages *pcp = NULL; local 2761 nr_pcp_alloc(struct per_cpu_pages *pcp, struct zone *zone, int order) argument 2813 __rmqueue_pcplist(struct zone *zone, unsigned int order, int migratetype, unsigned int alloc_flags, struct per_cpu_pages *pcp, struct list_head *list) argument 2848 struct per_cpu_pages *pcp; local 4396 struct per_cpu_pages *pcp; local 5482 pageset_update(struct per_cpu_pages *pcp, unsigned long high_min, unsigned long high_max, unsigned long batch) argument 5490 per_cpu_pages_init(struct per_cpu_pages *pcp, struct per_cpu_zonestat *pzstats) argument 5516 struct per_cpu_pages *pcp; local 5571 struct per_cpu_pages *pcp; local 5595 struct per_cpu_pages *pcp; local [all...] |
H A D | vmstat.c | 345 struct per_cpu_zonestat __percpu *pcp = zone->per_cpu_zonestats; local 346 s8 __percpu *p = pcp->vm_stat_diff + item; 361 t = __this_cpu_read(pcp->stat_threshold); 376 struct per_cpu_nodestat __percpu *pcp = pgdat->per_cpu_nodestats; local 377 s8 __percpu *p = pcp->vm_node_stat_diff + item; 397 t = __this_cpu_read(pcp->stat_threshold); 434 struct per_cpu_zonestat __percpu *pcp = zone->per_cpu_zonestats; local 435 s8 __percpu *p = pcp->vm_stat_diff + item; 442 t = __this_cpu_read(pcp->stat_threshold); 455 struct per_cpu_nodestat __percpu *pcp local 490 struct per_cpu_zonestat __percpu *pcp = zone->per_cpu_zonestats; local 511 struct per_cpu_nodestat __percpu *pcp = pgdat->per_cpu_nodestats; local 560 struct per_cpu_zonestat __percpu *pcp = zone->per_cpu_zonestats; local 618 struct per_cpu_nodestat __percpu *pcp = pgdat->per_cpu_nodestats; local 819 struct per_cpu_pages __percpu *pcp = zone->per_cpu_pageset; local 1738 struct per_cpu_pages *pcp; local [all...] |
/linux-master/drivers/net/dsa/sja1105/ |
H A D | sja1105_flower.c | 123 rule->key.tc.pcp = tc; 187 key->tc.pcp, rate_bytes_per_sec, 205 u16 pcp = U16_MAX; local 274 pcp = match.key->vlan_priority; 277 if (is_bcast_dmac && vid == U16_MAX && pcp == U16_MAX) { 281 if (dmac == U64_MAX && vid == U16_MAX && pcp != U16_MAX) { 283 key->tc.pcp = pcp; 286 if (dmac != U64_MAX && vid != U16_MAX && pcp != U16_MAX) { 290 key->vl.pcp [all...] |
H A D | sja1105.h | 182 int pcp; member in struct:sja1105_key::__anon51::__anon52 190 u16 pcp; member in struct:sja1105_key::__anon51::__anon53
|
H A D | sja1105_vl.c | 408 vl_lookup[k].vlanprior = rule->key.vl.pcp; 692 ipv = key->vl.pcp; 750 vl_lookup[i].vlanprior == key->vl.pcp)
|
/linux-master/net/dsa/ |
H A D | tag_ocelot_8021q.c | 68 u8 pcp = netdev_txq_to_tc(netdev, queue_mapping); local 76 ((pcp << VLAN_PRIO_SHIFT) | tx_vid));
|
H A D | tag_sja1105.c | 241 struct sk_buff *skb, u8 pcp) 262 return vlan_insert_tag(skb, xmit_tpid, (pcp << VLAN_PRIO_SHIFT) | 271 u8 pcp = netdev_txq_to_tc(netdev, queue_mapping); local 282 skb = sja1105_pvid_tag_control_pkt(dp, skb, pcp); 290 ((pcp << VLAN_PRIO_SHIFT) | tx_vid)); 299 u8 pcp = netdev_txq_to_tc(netdev, queue_mapping); local 314 ((pcp << VLAN_PRIO_SHIFT) | tx_vid)); 316 skb = sja1105_pvid_tag_control_pkt(dp, skb, pcp); 333 *tx_trailer = cpu_to_be32(SJA1110_TX_TRAILER_PRIO(pcp) | 240 sja1105_pvid_tag_control_pkt(struct dsa_port *dp, struct sk_buff *skb, u8 pcp) argument
|
/linux-master/drivers/net/ethernet/microchip/lan966x/ |
H A D | lan966x_port.c | 402 u8 pcp, dp; local 410 pcp = *(pcp_itr + i); 413 lan_rmw(ANA_PCP_DEI_CFG_QOS_PCP_DEI_VAL_SET(pcp) | 457 /* Set default pcp and dei for untagged frames */ 471 u8 pcp, dei; local 485 /* Map each value to pcp and dei */ 487 pcp = qos->map[i]; 488 if (pcp > LAN966X_PORT_QOS_PCP_COUNT) 494 REW_PCP_DEI_CFG_PCP_QOS_VAL_SET(pcp), 540 lan966x_port_qos_pcp_set(port, &qos->pcp); [all...] |
H A D | lan966x_dcb.c | 21 [LAN966X_DCB_APPTRUST_PCP] = "pcp", 22 [LAN966X_DCB_APPTRUST_DSCP_PCP] = "dscp pcp" 57 /* Get pcp ingress mapping */ 58 for (int i = 0; i < ARRAY_SIZE(qos.pcp.map); i++) { 61 qos.pcp.map[i] = dcb_getapp(dev, &app_itr); 76 /* Get pcp rewrite mapping */ 96 /* Enable use of pcp for queue classification */ 98 qos.pcp.enable = true;
|
/linux-master/drivers/infiniband/ulp/opa_vnic/ |
H A D | opa_vnic_encap.c | 371 u8 pcp = OPA_VNIC_VLAN_PCP(vlan_tci); local 374 sc = info->vport.pcp_to_sc_mc[pcp]; 376 sc = info->vport.pcp_to_sc_uc[pcp]; 394 u8 pcp = skb_vlan_tag_get(skb) >> VLAN_PRIO_SHIFT; local 397 vl = info->vport.pcp_to_vl_mc[pcp]; 399 vl = info->vport.pcp_to_vl_uc[pcp];
|
/linux-master/drivers/net/ethernet/microchip/sparx5/ |
H A D | sparx5_dcb.c | 27 [SPARX5_DCB_APPTRUST_PCP] = "pcp", 28 [SPARX5_DCB_APPTRUST_DSCP_PCP] = "dscp pcp" 150 pcp_map = &qos.pcp.map; 164 /* Get pcp ingress mapping */ 171 /* Get pcp rewrite mapping */ 200 /* Enable use of pcp for queue classification ? */ 202 qos.pcp.qos_enable = true; 203 qos.pcp.dp_enable = qos.pcp.qos_enable; 394 /* Initialize [dscp, pcp] defaul [all...] |
H A D | sparx5_port.h | 150 struct sparx5_port_qos_pcp pcp; member in struct:sparx5_port_qos
|
H A D | sparx5_port.c | 1158 sparx5_port_qos_pcp_set(port, &qos->pcp); 1171 u8 pcp, dei; local 1187 pcp = qos->map.map[i]; 1188 if (pcp > SPARX5_PORT_QOS_PCP_COUNT) 1197 * 0:0nd - prio=0 and dp:0 => pcp=0 and dei=0 1198 * 0:0de - prio=0 and dp:1 => pcp=0 and dei=1 1201 spx5_rmw(REW_PCP_MAP_DE1_PCP_DE1_SET(pcp), 1209 spx5_rmw(REW_PCP_MAP_DE0_PCP_DE0_SET(pcp), 1227 u8 pcp, dp; local 1230 /* Enable/disable pcp an [all...] |
/linux-master/arch/sparc/kernel/ |
H A D | pcic.c | 608 struct pcidev_cookie *pcp; local 631 pcp = pci_devcookie_alloc(); 632 pcp->pbm = &pcic->pbm; 633 pcp->prom_node = of_find_node_by_phandle(node); 634 dev->sysdata = pcp;
|
/linux-master/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | core_acl_flex_actions.h | 73 u16 vid, u8 pcp, u8 et,
|
/linux-master/include/soc/mscc/ |
H A D | ocelot_vcap.h | 499 struct ocelot_vcap_u8 pcp; /* PCP (3 bit) */ member in struct:ocelot_vcap_key_vlan 642 u8 pcp; member in struct:ocelot_vcap_action::__anon3251::__anon3253
|
/linux-master/drivers/net/ethernet/mscc/ |
H A D | ocelot_vcap.c | 385 tag->pcp.value[0], tag->pcp.mask[0]); 660 vcap_action_set(vcap, data, VCAP_IS1_ACT_PCP_VAL, a->pcp); 701 tag->pcp.value[0], tag->pcp.mask[0]); 839 tag->pcp.value[0], tag->pcp.mask[0]);
|