Searched refs:reclaimable (Results 1 - 11 of 11) sorted by relevance
/linux-master/include/trace/events/ |
H A D | oom.h | 38 unsigned long reclaimable, 44 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check), 50 __field( unsigned long, reclaimable) 61 __entry->reclaimable = reclaimable; 68 TP_printk("node=%d zone=%-8s order=%d reclaimable=%lu available=%lu min_wmark=%lu no_progress_loops=%d wmark_check=%d", 71 __entry->reclaimable, __entry->available, __entry->min_wmark,
|
/linux-master/mm/ |
H A D | show_mem.c | 37 unsigned long reclaimable; local 60 * Part of the reclaimable slab and other kernel memory consists of 64 reclaimable = global_node_page_state_pages(NR_SLAB_RECLAIMABLE_B) + 66 reclaimable -= min(reclaimable / 2, wmark_low); 67 available += reclaimable;
|
H A D | page-writeback.c | 250 * free and reclaimable pages, minus some zone reserves to protect 2084 unsigned long reclaimable; local 2100 reclaimable = wb_stat_sum(wb, WB_RECLAIMABLE); 2102 reclaimable = wb_stat(wb, WB_RECLAIMABLE); 2104 if (reclaimable > thresh) 2120 reclaimable = wb_stat_sum(wb, WB_RECLAIMABLE); 2122 reclaimable = wb_stat(wb, WB_RECLAIMABLE); 2124 if (reclaimable > thresh)
|
H A D | vmscan.c | 136 /* There is easily reclaimable cold cache in the current node */ 468 int reclaimable = 0, write_pending = 0; local 489 reclaimable += zone_reclaimable_pages(zone); 493 if (2 * write_pending <= reclaimable) 840 * The supposedly reclaimable folio was found to be in a VM_LOCKED vma. 4887 unsigned long reclaimable; local 4896 reclaimable = node_page_state(pgdat, NR_INACTIVE_FILE); 4898 reclaimable += node_page_state(pgdat, NR_INACTIVE_ANON); 4900 /* round down reclaimable and round up sc->nr_to_reclaim */ 4901 priority = fls_long(reclaimable) 5891 bool reclaimable = false; local [all...] |
H A D | page_alloc.c | 638 * This might let an unmovable request use a reclaimable pageblock 1694 * reclaimable and unmovable allocations, we steal regardless of page size, 1696 * is worse than movable allocations stealing from unmovable and reclaimable 2491 * We only track unmovable, reclaimable and movable on pcp lists. 3966 * request even if all reclaimable pages are considered then we are 3972 unsigned long reclaimable; local 3976 available = reclaimable = zone_reclaimable_pages(zone); 3981 * reclaimable pages? 3985 trace_reclaim_retry_zone(z, order, reclaimable,
|
/linux-master/drivers/net/ethernet/chelsio/cxgb4vf/ |
H A D | sge.c | 409 * Return the number of reclaimable descriptors in a TX queue. 411 static inline int reclaimable(const struct sge_txq *tq) function 414 int reclaimable = hw_cidx - tq->cidx; local 415 if (reclaimable < 0) 416 reclaimable += tq->size; 417 return reclaimable; 434 int avail = reclaimable(tq); 2133 if (reclaimable(&txq->q) && __netif_tx_trylock(txq->txq)) { 2134 int avail = reclaimable(&txq->q); 2155 * If we found too many reclaimable packet [all...] |
/linux-master/fs/xfs/ |
H A D | xfs_icache.c | 186 * Queue background inode reclaim work if there are reclaimable inodes and there 297 * When we recycle a reclaimable inode, we need to re-initialise the VFS inode 514 * reclaimable state, wait for the initialisation to complete 1016 * Return the number of reclaimable inodes in the filesystem for 1025 long reclaimable = 0; local 1029 reclaimable += pag->pag_ici_reclaimable; 1032 return reclaimable; 2163 * it does make inodes reclaimable, which eventually frees memory.
|
/linux-master/drivers/md/ |
H A D | raid5-cache.c | 28 * log->max_free_space is min(1/4 disk size, 10G reclaimable space). 1502 sector_t reclaimable; local 1511 * reclaimable/unreclaimable io_unit can be mixed in the list, we 1515 reclaimable = r5l_reclaimable_space(log); 1516 if (reclaimable >= reclaim_target || 1525 r5l_reclaimable_space(log) > reclaimable, 1532 if (reclaimable == 0 || !write_super)
|
/linux-master/include/net/ |
H A D | sock.h | 259 * @sk_reserved_mem: space reserved and non-reclaimable for the socket 1612 int reclaimable; local 1617 reclaimable = sk->sk_forward_alloc - sk_unused_reserved_mem(sk); 1619 if (reclaimable >= (int)PAGE_SIZE) 1620 __sk_mem_reclaim(sk, reclaimable);
|
/linux-master/net/mptcp/ |
H A D | protocol.c | 176 int reclaimable; local 179 reclaimable = msk->rmem_fwd_alloc - sk_unused_reserved_mem(sk); 182 if (unlikely(reclaimable >= PAGE_SIZE)) 183 __mptcp_rmem_reclaim(sk, reclaimable);
|
/linux-master/drivers/net/ethernet/chelsio/cxgb4/ |
H A D | sge.c | 339 * Return the number of reclaimable descriptors in a Tx queue. 341 static inline int reclaimable(const struct sge_txq *q) function 362 int reclaim = reclaimable(q); 4290 avail = reclaimable(&q->q); 4313 /* If we found too many reclaimable packets schedule a timer 4318 /* We reclaimed all reclaimable TX Descriptors, so reschedule
|
Completed in 501 milliseconds