Lines Matching defs:req

965 						struct init_ets_req* req,
984 struct init_ets_tc_req *tc_req = &req->tc_req[tc];
1009 struct init_ets_tc_req *tc_req = &req->tc_req[tc];
1022 ecore_wr(p_hwfn, p_ptt, tc_bound_base_addr + tc_bound_addr_diff * tc_client_offset, NIG_ETS_UP_BOUND(byte_weight, req->mtu));
1028 struct init_nig_lb_rl_req* req)
1038 if (req->lb_mac_rate) {
1042 inc_val = NIG_RL_INC_VAL(req->lb_mac_rate);
1044 ecore_wr(p_hwfn, p_ptt, NIG_REG_TX_LB_GLBRATELIMIT_MAX_VALUE, NIG_RL_MAX_VAL(inc_val, req->mtu));
1056 if (req->lb_rate) {
1060 inc_val = NIG_RL_INC_VAL(req->lb_rate);
1062 ecore_wr(p_hwfn, p_ptt, NIG_REG_LB_BRBRATELIMIT_MAX_VALUE, NIG_RL_MAX_VAL(inc_val, req->mtu));
1077 if (!req->tc_rate[tc])
1082 inc_val = NIG_RL_INC_VAL(req->tc_rate[tc]);
1084 ecore_wr(p_hwfn, p_ptt, NIG_REG_LB_TCRATELIMIT_MAX_VALUE_0 + reg_offset, NIG_RL_MAX_VAL(inc_val, req->mtu));
1094 struct init_nig_pri_tc_map_req* req)
1101 if (!req->pri[pri].valid)
1104 pri_tc_mask |= (req->pri[pri].tc_id << (pri * NIG_PRIORITY_MAP_TC_BITS));
1105 tc_pri_mask[req->pri[pri].tc_id] |= (1 << pri);
1129 struct init_ets_req* req)
1138 struct init_ets_tc_req *tc_req = &req->tc_req[tc];
1163 struct init_ets_tc_req *tc_req = &req->tc_req[tc];
1176 ecore_wr(p_hwfn, p_ptt, PRS_REG_ETS_ARB_CREDIT_UPPER_BOUND_0 + tc * tc_bound_addr_diff, PRS_ETS_UP_BOUND(byte_weight, req->mtu));
1194 struct init_brb_ram_req* req)
1200 tc_headroom_blocks = (u32)DIV_ROUND_UP(req->headroom_per_tc, BRB_BLOCK_SIZE);
1201 min_pkt_size_blocks = (u32)DIV_ROUND_UP(req->min_pkt_size, BRB_BLOCK_SIZE);
1206 if (req->num_active_tcs[port])
1211 for (port = 0; port < req->max_ports_per_engine; port++) {
1218 tc_guaranteed_blocks = (u32)DIV_ROUND_UP(req->guranteed_per_tc, BRB_BLOCK_SIZE);
1219 port_blocks = req->num_active_tcs[port] ? active_port_blocks : 0;
1220 port_guaranteed_blocks = req->num_active_tcs[port] * tc_guaranteed_blocks;
1222 full_xoff_th = req->num_active_tcs[port] * BRB_MIN_BLOCKS_PER_TC;
1235 if (tc == req->num_active_tcs[port]) {