/linux-master/kernel/sched/ |
H A D | fair.c | 173 lw->weight += inc; 179 lw->weight -= dec; 185 lw->weight = w; 244 w = scale_load_down(lw->weight); 255 * delta_exec * weight / lw.weight 257 * (delta_exec * (weight * lw->inv_weight)) >> WMULT_SHIFT 259 * Either weight := NICE_0_LOAD and lw \e sched_prio_to_wmult[], in which case 263 * Or, weight =< lw.weight (becaus 266 __calc_delta(u64 delta_exec, unsigned long weight, struct load_weight *lw) argument 630 unsigned long weight = scale_load_down(se->load.weight); local 640 unsigned long weight = scale_load_down(se->load.weight); local 667 unsigned long weight = scale_load_down(curr->load.weight); local 740 unsigned long weight = scale_load_down(curr->load.weight); local 1968 unsigned int weight; member in struct:numa_stats 3685 reweight_eevdf(struct sched_entity *se, u64 avruntime, unsigned long weight) argument 3791 reweight_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, unsigned long weight) argument 3849 unsigned long weight = scale_load(sched_prio_to_weight[prio]); local 7193 int weight; local [all...] |
H A D | sched.h | 122 * low-weight task groups (eg. nice +19 on an autogroup), deeper taskgroup 150 * Task weight (visible to users) and its load (invisible to users) have 437 * A weight of 0 or 1 can cause arithmetics problems. 438 * A weight of a cfs_rq is the sum of weights of which entities 439 * are queued on this cfs_rq, so a weight of a entity should not be 441 * (The default weight is 1024 - so there's no practical 630 * h_load = weight * f(tg) 632 * Where f(tg) is the recursive weight fraction assigned to 826 return scale_load_down(se->load.weight);
|
H A D | core.c | 1333 * SCHED_IDLE tasks get minimal weight: 1336 load->weight = scale_load(WEIGHT_IDLEPRIO); 1343 * weight 1348 load->weight = scale_load(sched_prio_to_weight[prio]); 10006 * based on each entity's (task or task-group's) weight 10007 * (se->load.weight). 10009 * In other words, if root_task_group has 10 tasks of weight 10010 * 1024) and two child groups A0 and A1 (of weight 1024 each), 11301 u64 weight = scale_load_down(tg->shares); local 11303 return DIV_ROUND_CLOSEST_ULL(weight * CGROUP_WEIGHT_DF 11306 cpu_weight_write_u64(struct cgroup_subsys_state *css, struct cftype *cft, u64 weight) argument 11327 unsigned long weight = scale_load_down(css_tg(css)->shares); local 11345 unsigned long weight; local 11911 sched_mm_cid_remote_clear_weight(struct mm_struct *mm, int cpu, int weight) argument 11930 int weight, cpu; local [all...] |
/linux-master/drivers/thunderbolt/ |
H A D | tb.h | 408 * @weight: Weight of the path inside the priority group 430 int weight:4; member in struct:tb_path
|
H A D | tunnel.c | 270 path->weight = TB_PCI_WEIGHT; 1239 path->weight = TB_DP_AUX_WEIGHT; 1286 path->weight = TB_DP_VIDEO_WEIGHT; 1563 path->weight = TB_DMA_WEIGHT; 1597 path->weight = TB_DMA_WEIGHT; 1939 path->weight = TB_USB3_WEIGHT;
|
H A D | tb_regs.h | 515 u32 weight:4; member in struct:tb_regs_hop
|
/linux-master/block/ |
H A D | blk-iocost.c | 67 * of equal weight, each gets 50% share. If then B starts issuing IOs, B 74 * For example, with 12.5% weight, A0's time runs 8 times slower (100/12.5) 137 * each active cgroup is actually using and yield part of its weight if 139 * A's weight will be lowered so that it hovers above the actual usage and 142 * As we don't want to penalize a cgroup for donating its weight, the 143 * surplus weight adjustment factors in a margin and has an immediate 161 * active weight hweight% inflt% dbt delay usages% 168 * - weight : Surplus-adjusted and configured weights 221 * serves as its IO credit buffer. Surplus weight adjustment is 467 * A iocg can get its weight fro 485 u32 weight; member in struct:ioc_gq 1251 u32 weight; local [all...] |
/linux-master/net/core/ |
H A D | dev.c | 4415 int weight_p __read_mostly = 64; /* old backlog weight */ 4416 int dev_weight_rx_bias __read_mostly = 1; /* bias for backlog weight */ 5981 napi->weight = READ_ONCE(net_hotdata.dev_rx_weight); 6508 int (*poll)(struct napi_struct *, int), int weight) 6522 if (weight > NAPI_POLL_WEIGHT) 6523 netdev_err_once(dev, "%s() called with weight %d\n", __func__, 6524 weight); 6525 napi->weight = weight; 6626 int work, weight; local 6507 netif_napi_add_weight(struct net_device *dev, struct napi_struct *napi, int (*poll)(struct napi_struct *, int), int weight) argument [all...] |
/linux-master/drivers/net/ethernet/stmicro/stmmac/ |
H A D | stmmac_main.c | 3159 * stmmac_set_tx_queue_weight - Set TX queue weight 3161 * Description: It is used for setting TX queues weight 3166 u32 weight; local 3170 weight = priv->plat->tx_queues_cfg[queue].weight; 3171 stmmac_set_mtl_tx_queue_weight(priv, priv->hw, weight, queue);
|
H A D | dwxgmac2_core.c | 246 u32 weight, u32 queue) 250 writel(weight, ioaddr + XGMAC_MTL_TCx_QUANTUM_WEIGHT(queue)); 244 dwxgmac2_set_mtl_tx_queue_weight(struct stmmac_priv *priv, struct mac_device_info *hw, u32 weight, u32 queue) argument
|
H A D | dwmac4_core.c | 238 u32 weight, u32 queue) 246 value |= weight & MTL_TXQ_WEIGHT_ISCQW_MASK; 297 /* configure idle slope (same register as tx weight) */ 236 dwmac4_set_mtl_tx_queue_weight(struct stmmac_priv *priv, struct mac_device_info *hw, u32 weight, u32 queue) argument
|
/linux-master/drivers/vhost/ |
H A D | vhost.c | 470 pkts >= dev->weight) { 505 int iov_limit, int weight, int byte_weight, 521 dev->weight = weight; 503 vhost_dev_init(struct vhost_dev *dev, struct vhost_virtqueue **vqs, int nvqs, int iov_limit, int weight, int byte_weight, bool use_worker, int (*msg_handler)(struct vhost_dev *dev, u32 asid, struct vhost_iotlb_msg *msg)) argument
|
/linux-master/arch/x86/events/ |
H A D | core.c | 785 int weight; member in struct:sched_state 821 if (constraints[idx]->weight == wmin) 825 sched->state.event = idx; /* start with min weight */ 826 sched->state.weight = wmin; 928 * Take events with the least weight first. Return true on success. 941 /* next weight */ 943 sched->state.weight++; 944 if (sched->state.weight > sched->max_weight) 948 } while (c->weight != sched->state.weight); [all...] |
/linux-master/fs/bcachefs/ |
H A D | util.h | 335 #define ewma_add(ewma, val, weight) \ 338 typeof(weight) _weight = (weight); \
|
H A D | mean_and_variance_test.c | 93 unsigned weight, 105 mean_and_variance_weighted_update(&vw, initial_value, false, weight); 109 KUNIT_EXPECT_EQ(test, mean_and_variance_weighted_get_mean(vw, weight), initial_value); 110 KUNIT_EXPECT_EQ(test, mean_and_variance_weighted_get_stddev(vw, weight),0); 115 mean_and_variance_weighted_update(&vw, data[i], true, weight); 119 KUNIT_EXPECT_EQ(test, mean_and_variance_weighted_get_mean(vw, weight), weighted_mean[i]); 120 KUNIT_EXPECT_EQ(test, mean_and_variance_weighted_get_stddev(vw, weight),weighted_stddev[i]); 89 do_mean_and_variance_test(struct kunit *test, s64 initial_value, s64 initial_n, s64 n, unsigned weight, s64 *data, s64 *mean, s64 *stddev, s64 *weighted_mean, s64 *weighted_stddev) argument
|
/linux-master/drivers/net/ |
H A D | virtio_net.c | 53 * at once, the weight is chosen so that the EWMA will be insensitive to short- 535 if (napi->weight) 861 bool use_napi = sq->napi.weight; 2069 if (!napi->weight) 2076 napi->weight = 0; 2085 if (napi->weight) 2168 if (!sq->napi.weight || is_xdp_raw_buffer_queue(vi, index)) 2410 bool use_napi = sq->napi.weight; 3614 static int virtnet_should_update_vq_weight(int dev_flags, int weight, argument 3617 if (weight [all...] |
/linux-master/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | qos.c | 73 u32 weight; local 75 weight = quantum / pfvf->hw.dwrr_mtu; 77 weight += 1; 79 return weight; 1101 /* Max Round robin weight supported by octeontx2 and CN10K
|
/linux-master/arch/x86/hyperv/ |
H A D | hv_apic.c | 163 unsigned int weight; local 167 weight = cpumask_weight(mask); 174 if (weight == 0 || 175 (exclude_self && weight == 1 && cpumask_test_cpu(this_cpu, mask)))
|
/linux-master/arch/x86/events/intel/ |
H A D | ds.c | 1205 * + For RTM TSX weight we need GPRs for the abort code. 1637 * Use latency for weight (only avail with PEBS-LL) 1640 data->weight.full = pebs->lat; 1737 /* Only set the TSX weight when no memory weight. */ 1739 data->weight.full = intel_get_tsx_weight(pebs->tsx_tuning); 1835 data->weight.var3_w = format_size >> PEBS_RETIRE_LATENCY_OFFSET & PEBS_LATENCY_MASK; 1862 u64 weight = meminfo->latency; local 1865 data->weight.var2_w = weight [all...] |
/linux-master/drivers/thermal/ |
H A D | gov_power_allocator.c | 409 u32 power_range, weight; local 431 weight = 1 << FRAC_BITS; 433 weight = instance->weight; 435 pa->weighted_req_power = frac_to_int(weight * pa->req_power); 655 params->total_weight += instance->weight;
|
/linux-master/arch/x86/events/amd/ |
H A D | core.c | 547 nb->event_constraints[i].weight = 1;
|
/linux-master/drivers/scsi/cxlflash/ |
H A D | main.c | 2853 * irqpoll_weight_show() - presents the current IRQ poll weight for the host 2855 * @attr: Device attribute representing the IRQ poll weight. 2857 * weight in ASCII. 2859 * An IRQ poll weight of 0 indicates polling is disabled. 2873 * irqpoll_weight_store() - sets the current IRQ poll weight for the host 2875 * @attr: Device attribute representing the IRQ poll weight. 2877 * weight in ASCII. 2880 * An IRQ poll weight of 0 indicates polling is disabled. 2892 u32 weight; local 2895 rc = kstrtouint(buf, 10, &weight); [all...] |
/linux-master/net/ipv4/ |
H A D | nexthop.c | 154 info->nh_grp->nh_entries[i].weight = nhge->weight; 892 p->weight = nhg->nh_entries[i].weight - 1; 1286 if (nhg[i].weight > 254) { 1287 NL_SET_ERR_MSG(extack, "Invalid value for weight"); 1890 total += nhg->nh_entries[i].weight; 1896 w += nhge->weight; 1969 total += nhg->nh_entries[i].weight; 1975 w += nhge->weight; [all...] |
/linux-master/drivers/net/ethernet/microchip/ |
H A D | lan743x_main.c | 2229 static int lan743x_tx_napi_poll(struct napi_struct *napi, int weight) argument 2761 static int lan743x_rx_napi_poll(struct napi_struct *napi, int weight) argument 2774 for (count = 0; count < weight; count++) { 2780 if (count == weight || result == RX_PROCESS_RESULT_BUFFER_RECEIVED) 2781 return weight;
|
/linux-master/include/linux/ |
H A D | netdevice.h | 356 int weight; member in struct:napi_struct 2613 /* Default NAPI poll() weight 2619 int (*poll)(struct napi_struct *, int), int weight); 2641 int weight) 2644 netif_napi_add_weight(dev, napi, poll, weight); 2638 netif_napi_add_tx_weight(struct net_device *dev, struct napi_struct *napi, int (*poll)(struct napi_struct *, int), int weight) argument
|