/linux-master/tools/testing/memblock/linux/ |
H A D | mmzone.h | 8 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat); 10 #define for_each_online_pgdat(pgdat) \ 11 for (pgdat = first_online_pgdat(); \ 12 pgdat; \ 13 pgdat = next_online_pgdat(pgdat))
|
/linux-master/mm/ |
H A D | shuffle.h | 11 extern void __shuffle_free_memory(pg_data_t *pgdat); 13 static inline void __meminit shuffle_free_memory(pg_data_t *pgdat) argument 17 __shuffle_free_memory(pgdat); 40 static inline void shuffle_free_memory(pg_data_t *pgdat) argument
|
H A D | mmzone.c | 18 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) argument 20 int nid = next_online_node(pgdat->node_id); 32 pg_data_t *pgdat = zone->zone_pgdat; local 34 if (zone < pgdat->node_zones + MAX_NR_ZONES - 1) 37 pgdat = next_online_pgdat(pgdat); 38 if (pgdat) 39 zone = pgdat->node_zones;
|
H A D | show_mem.c | 95 pg_data_t *pgdat = NODE_DATA(nid); local 98 managed_pages += zone_managed_pages(&pgdat->node_zones[zone_type]); 100 val->sharedram = node_page_state(pgdat, NR_SHMEM); 104 struct zone *zone = &pgdat->node_zones[zone_type]; 168 static bool node_has_managed_zones(pg_data_t *pgdat, int max_zone_idx) argument 172 if (zone_managed_pages(pgdat->node_zones + zone_idx)) 191 pg_data_t *pgdat; local 232 for_each_online_pgdat(pgdat) { 233 if (show_mem_node_skip(filter, pgdat->node_id, nodemask)) 235 if (!node_has_managed_zones(pgdat, max_zone_id [all...] |
H A D | vmstat.c | 274 struct pglist_data *pgdat; local 279 /* Zero current pgdat thresholds */ 280 for_each_online_pgdat(pgdat) { 282 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold = 0; 287 struct pglist_data *pgdat = zone->zone_pgdat; local 299 pgdat_threshold = per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold; 300 per_cpu_ptr(pgdat->per_cpu_nodestats, cpu)->stat_threshold 317 void set_pgdat_percpu_threshold(pg_data_t *pgdat, argument 325 for (i = 0; i < pgdat->nr_zones; i++) { 326 zone = &pgdat 373 __mod_node_page_state(struct pglist_data *pgdat, enum node_stat_item item, long delta) argument 453 __inc_node_state(struct pglist_data *pgdat, enum node_stat_item item) argument 509 __dec_node_state(struct pglist_data *pgdat, enum node_stat_item item) argument 615 mod_node_state(struct pglist_data *pgdat, enum node_stat_item item, int delta, int overstep_mode) argument 665 mod_node_page_state(struct pglist_data *pgdat, enum node_stat_item item, long delta) argument 672 inc_node_state(struct pglist_data *pgdat, enum node_stat_item item) argument 725 inc_node_state(struct pglist_data *pgdat, enum node_stat_item item) argument 735 mod_node_page_state(struct pglist_data *pgdat, enum node_stat_item item, long delta) argument 749 struct pglist_data *pgdat; local 810 struct pglist_data *pgdat; local 898 struct pglist_data *pgdat; local 1016 node_page_state_pages(struct pglist_data *pgdat, enum node_stat_item item) argument 1027 node_page_state(struct pglist_data *pgdat, enum node_stat_item item) argument 1424 pg_data_t *pgdat; local 1437 pg_data_t *pgdat = (pg_data_t *)arg; local 1451 walk_zones_in_node(struct seq_file *m, pg_data_t *pgdat, bool assert_populated, bool nolock, void (*print)(struct seq_file *m, pg_data_t *, struct zone *)) argument 1473 frag_show_print(struct seq_file *m, pg_data_t *pgdat, struct zone *zone) argument 1493 pg_data_t *pgdat = (pg_data_t *)arg; local 1498 pagetypeinfo_showfree_print(struct seq_file *m, pg_data_t *pgdat, struct zone *zone) argument 1544 pg_data_t *pgdat = (pg_data_t *)arg; local 1555 pagetypeinfo_showblockcount_print(struct seq_file *m, pg_data_t *pgdat, struct zone *zone) argument 1591 pg_data_t *pgdat = (pg_data_t *)arg; local 1607 pagetypeinfo_showmixedcount(struct seq_file *m, pg_data_t *pgdat) argument 1633 pg_data_t *pgdat = (pg_data_t *)arg; local 1663 is_zone_first_populated(pg_data_t *pgdat, struct zone *zone) argument 1677 zoneinfo_show_print(struct seq_file *m, pg_data_t *pgdat, struct zone *zone) argument 1773 pg_data_t *pgdat = (pg_data_t *)arg; local 2174 unusable_show_print(struct seq_file *m, pg_data_t *pgdat, struct zone *zone) argument 2204 pg_data_t *pgdat = (pg_data_t *)arg; local 2224 extfrag_show_print(struct seq_file *m, pg_data_t *pgdat, struct zone *zone) argument 2250 pg_data_t *pgdat = (pg_data_t *)arg; local [all...] |
H A D | mm_init.c | 48 pg_data_t *pgdat = NODE_DATA(nid); local 60 zonelist = &pgdat->node_zonelists[listid]; 61 zone = &pgdat->node_zones[zoneid]; 653 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) argument 655 pgdat->first_deferred_pfn = ULONG_MAX; 708 pg_data_t *pgdat; local 714 pgdat = NODE_DATA(nid); 717 struct zone *zone = &pgdat->node_zones[zid]; 725 static inline void pgdat_set_deferred_range(pg_data_t *pgdat) {} argument 1070 struct pglist_data *pgdat local 1250 reset_memoryless_node_totalpages(struct pglist_data *pgdat) argument 1268 calculate_node_totalpages(struct pglist_data *pgdat, unsigned long node_start_pfn, unsigned long node_end_pfn) argument 1332 pgdat_init_split_queue(struct pglist_data *pgdat) argument 1341 pgdat_init_split_queue(struct pglist_data *pgdat) argument 1345 pgdat_init_kcompactd(struct pglist_data *pgdat) argument 1350 pgdat_init_kcompactd(struct pglist_data *pgdat) argument 1353 pgdat_init_internals(struct pglist_data *pgdat) argument 1402 struct pglist_data *pgdat = zone->zone_pgdat; local 1502 free_area_init_core_hotplug(struct pglist_data *pgdat) argument 1554 free_area_init_core(struct pglist_data *pgdat) argument 1635 alloc_node_mem_map(struct pglist_data *pgdat) argument 1672 alloc_node_mem_map(struct pglist_data *pgdat) argument 1705 pg_data_t *pgdat = NODE_DATA(nid); local 1738 check_for_memory(pg_data_t *pgdat) argument 1868 pg_data_t *pgdat; local 2177 pg_data_t *pgdat = data; local 2270 pg_data_t *pgdat = zone->zone_pgdat; local [all...] |
H A D | vmscan.c | 466 static bool skip_throttle_noprogress(pg_data_t *pgdat) argument 475 if (pgdat->kswapd_failures >= MAX_RECLAIM_RETRIES) 484 struct zone *zone = pgdat->node_zones + i; 499 void reclaim_throttle(pg_data_t *pgdat, enum vmscan_throttle_state reason) argument 501 wait_queue_head_t *wqh = &pgdat->reclaim_wait[reason]; 530 if (atomic_inc_return(&pgdat->nr_writeback_throttled) == 1) { 531 WRITE_ONCE(pgdat->nr_reclaim_start, 532 node_page_state(pgdat, NR_THROTTLED_WRITTEN)); 539 if (skip_throttle_noprogress(pgdat)) { 561 atomic_dec(&pgdat 573 __acct_reclaim_writeback(pg_data_t *pgdat, struct folio *folio, int nr_throttled) argument 954 demote_folio_list(struct list_head *demote_folios, struct pglist_data *pgdat) argument 1011 shrink_folio_list(struct list_head *folio_list, struct pglist_data *pgdat, struct scan_control *sc, struct reclaim_stat *stat, bool ignore_references) argument 1754 too_many_isolated(struct pglist_data *pgdat, int file, struct scan_control *sc) argument 1891 struct pglist_data *pgdat = lruvec_pgdat(lruvec); local 2012 struct pglist_data *pgdat = lruvec_pgdat(lruvec); local 2093 reclaim_folio_list(struct list_head *folio_list, struct pglist_data *pgdat, bool ignore_references) argument 2220 prepare_scan_control(pg_data_t *pgdat, struct scan_control *sc) argument 2337 struct pglist_data *pgdat = lruvec_pgdat(lruvec); local 2535 can_age_anon_pages(struct pglist_data *pgdat, struct scan_control *sc) argument 2591 struct pglist_data *pgdat = NODE_DATA(nid); local 2612 struct pglist_data *pgdat = lruvec_pgdat(lruvec); local 2762 struct pglist_data *pgdat = lruvec_pgdat(walk->lruvec); local 3299 get_pfn_folio(unsigned long pfn, struct mem_cgroup *memcg, struct pglist_data *pgdat, bool can_swap) argument 3341 struct pglist_data *pgdat = lruvec_pgdat(walk->lruvec); local 3409 struct pglist_data *pgdat = lruvec_pgdat(walk->lruvec); local 3516 struct pglist_data *pgdat = lruvec_pgdat(walk->lruvec); local 3650 set_mm_walk(struct pglist_data *pgdat, bool force_alloc) argument 3938 lru_gen_age_node(struct pglist_data *pgdat, struct scan_control *sc) argument 4001 struct pglist_data *pgdat = folio_pgdat(folio); local 4109 struct pglist_data *pgdat = lruvec_pgdat(lruvec); local 4158 struct pglist_data *pgdat = NODE_DATA(nid); local 4193 struct pglist_data *pgdat = NODE_DATA(nid); local 4521 struct pglist_data *pgdat = lruvec_pgdat(lruvec); local 4757 struct pglist_data *pgdat = lruvec_pgdat(lruvec); local 4793 shrink_many(struct pglist_data *pgdat, struct scan_control *sc) argument 4884 set_initial_priority(struct pglist_data *pgdat, struct scan_control *sc) argument 4906 lru_gen_shrink_node(struct pglist_data *pgdat, struct scan_control *sc) argument 5533 lru_gen_init_pgdat(struct pglist_data *pgdat) argument 5624 lru_gen_age_node(struct pglist_data *pgdat, struct scan_control *sc) argument 5634 lru_gen_shrink_node(struct pglist_data *pgdat, struct scan_control *sc) argument 5777 should_continue_reclaim(struct pglist_data *pgdat, unsigned long nr_reclaimed, struct scan_control *sc) argument 5829 shrink_node_memcgs(pg_data_t *pgdat, struct scan_control *sc) argument 5887 shrink_node(pg_data_t *pgdat, struct scan_control *sc) argument 6035 consider_reclaim_throttle(pg_data_t *pgdat, struct scan_control *sc) argument 6165 snapshot_refaults(struct mem_cgroup *target_memcg, pg_data_t *pgdat) argument 6285 allow_direct_reclaim(pg_data_t *pgdat) argument 6339 pg_data_t *pgdat = NULL; local 6460 mem_cgroup_shrink_node(struct mem_cgroup *memcg, gfp_t gfp_mask, bool noswap, pg_data_t *pgdat, unsigned long *nr_scanned) argument 6539 kswapd_age_node(struct pglist_data *pgdat, struct scan_control *sc) argument 6565 pgdat_watermark_boosted(pg_data_t *pgdat, int highest_zoneidx) argument 6593 pgdat_balanced(pg_data_t *pgdat, int order, int highest_zoneidx) argument 6629 clear_pgdat_congested(pg_data_t *pgdat) argument 6645 prepare_kswapd_sleep(pg_data_t *pgdat, int order, int highest_zoneidx) argument 6684 kswapd_shrink_node(pg_data_t *pgdat, struct scan_control *sc) argument 6721 update_reclaim_active(pg_data_t *pgdat, int highest_zoneidx, bool active) argument 6740 set_reclaim_active(pg_data_t *pgdat, int highest_zoneidx) argument 6746 clear_reclaim_active(pg_data_t *pgdat, int highest_zoneidx) argument 6764 balance_pgdat(pg_data_t *pgdat, int order, int highest_zoneidx) argument 6992 kswapd_highest_zoneidx(pg_data_t *pgdat, enum zone_type prev_highest_zoneidx) argument 7000 kswapd_try_to_sleep(pg_data_t *pgdat, int alloc_order, int reclaim_order, unsigned int highest_zoneidx) argument 7101 pg_data_t *pgdat = (pg_data_t *)p; local 7185 pg_data_t *pgdat; local 7271 pg_data_t *pgdat = NODE_DATA(nid); local 7293 pg_data_t *pgdat = NODE_DATA(nid); local 7345 node_unmapped_file_pages(struct pglist_data *pgdat) argument 7360 node_pagecache_reclaimable(struct pglist_data *pgdat) argument 7390 __node_reclaim(struct pglist_data *pgdat, gfp_t gfp_mask, unsigned int order) argument 7443 node_reclaim(struct pglist_data *pgdat, gfp_t gfp_mask, unsigned int order) argument [all...] |
H A D | bootmem_info.c | 102 void __init register_page_bootmem_info_node(struct pglist_data *pgdat) argument 105 int node = pgdat->node_id; 109 page = virt_to_page(pgdat); 114 pfn = pgdat->node_start_pfn; 115 end_pfn = pgdat_end_pfn(pgdat);
|
H A D | workingset.c | 199 static void *pack_shadow(int memcgid, pg_data_t *pgdat, unsigned long eviction, argument 204 eviction = (eviction << NODES_SHIFT) | pgdat->node_id; 210 static void unpack_shadow(void *shadow, int *memcgidp, pg_data_t **pgdat, argument 225 *pgdat = NODE_DATA(nid); 244 struct pglist_data *pgdat = folio_pgdat(folio); local 248 lruvec = mem_cgroup_lruvec(memcg, pgdat); 256 return pack_shadow(mem_cgroup_id(memcg), pgdat, token, refs); 269 struct pglist_data *pgdat; local 271 unpack_shadow(shadow, &memcg_id, &pgdat, token, workingset); 274 *lruvec = mem_cgroup_lruvec(memcg, pgdat); 384 struct pglist_data *pgdat = folio_pgdat(folio); local 426 struct pglist_data *pgdat; local 533 struct pglist_data *pgdat; local [all...] |
H A D | compaction.c | 453 void reset_isolation_suitable(pg_data_t *pgdat) argument 458 struct zone *zone = &pgdat->node_zones[zoneid]; 817 pg_data_t *pgdat = cc->zone->zone_pgdat; local 822 inactive = node_page_state(pgdat, NR_INACTIVE_FILE) + 823 node_page_state(pgdat, NR_INACTIVE_ANON); 824 active = node_page_state(pgdat, NR_ACTIVE_FILE) + 825 node_page_state(pgdat, NR_ACTIVE_ANON); 826 isolated = node_page_state(pgdat, NR_ISOLATED_FILE) + 827 node_page_state(pgdat, NR_ISOLATED_ANON); 842 wake_throttle_isolated(pgdat); 894 pg_data_t *pgdat = cc->zone->zone_pgdat; local 2203 kswapd_is_running(pg_data_t *pgdat) argument 2248 fragmentation_score_node(pg_data_t *pgdat) argument 2278 should_proactive_compact_node(pg_data_t *pgdat) argument 2317 pg_data_t *pgdat; local 2900 compact_node(pg_data_t *pgdat, bool proactive) argument 2964 pg_data_t *pgdat = NODE_DATA(nid); local 3030 kcompactd_work_requested(pg_data_t *pgdat) argument 3036 kcompactd_node_suitable(pg_data_t *pgdat) argument 3059 kcompactd_do_work(pg_data_t *pgdat) argument 3137 wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx) argument 3169 pg_data_t *pgdat = (pg_data_t *)p; local 3244 pg_data_t *pgdat = NODE_DATA(nid); local 3281 pg_data_t *pgdat = NODE_DATA(nid); local [all...] |
H A D | shuffle.c | 151 * @pgdat: node page data 153 void __meminit __shuffle_free_memory(pg_data_t *pgdat) argument 157 for (z = pgdat->node_zones; z < pgdat->node_zones + MAX_NR_ZONES; z++)
|
H A D | memory-tiers.c | 235 pg_data_t *pgdat; local 237 pgdat = NODE_DATA(node); 238 if (!pgdat) 245 return rcu_dereference_check(pgdat->memtier, 253 pg_data_t *pgdat; local 256 pgdat = NODE_DATA(node); 257 if (!pgdat) 261 memtier = rcu_dereference(pgdat->memtier); 275 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) argument 285 memtier = rcu_dereference(pgdat 509 pg_data_t *pgdat = NODE_DATA(node); local 536 pg_data_t *pgdat; local [all...] |
H A D | memory_hotplug.c | 499 static void update_pgdat_span(struct pglist_data *pgdat) argument 504 for (zone = pgdat->node_zones; 505 zone < pgdat->node_zones + MAX_NR_ZONES; zone++) { 523 pgdat->node_start_pfn = node_start_pfn; 524 pgdat->node_spanned_pages = node_end_pfn - node_start_pfn; 532 struct pglist_data *pgdat = zone->zone_pgdat; local 557 update_pgdat_span(pgdat); 716 static void __meminit resize_pgdat_range(struct pglist_data *pgdat, unsigned long start_pfn, argument 719 unsigned long old_end_pfn = pgdat_end_pfn(pgdat); 721 if (!pgdat 754 struct pglist_data *pgdat = zone->zone_pgdat; local 849 pg_data_t *pgdat = NODE_DATA(nid); local 900 struct pglist_data *pgdat = NODE_DATA(nid); local 1239 struct pglist_data *pgdat; local 1274 pg_data_t *pgdat; local 1887 struct pglist_data *pgdat = zone->zone_pgdat; local [all...] |
H A D | sparse.c | 321 static inline phys_addr_t pgdat_to_phys(struct pglist_data *pgdat) argument 324 VM_BUG_ON(pgdat != &contig_page_data); 327 return __pa(pgdat); 332 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, argument 342 * a pgdat can prevent a section being removed. If section A 343 * contains a pgdat and section B contains the usemap, both 345 * from the same section as the pgdat where possible to avoid 348 goal = pgdat_to_phys(pgdat) & (PAGE_SECTION_MASK << PAGE_SHIFT); 366 struct pglist_data *pgdat = NODE_DATA(nid); local 376 pgdat_snr = pfn_to_section_nr(pgdat_to_phys(pgdat) >> PAGE_SHIF 404 sparse_early_usemaps_alloc_pgdat_section(struct pglist_data *pgdat, unsigned long size) argument [all...] |
/linux-master/include/linux/ |
H A D | memory_hotplug.h | 21 * For supporting node-hotadd, we have to allocate a new pgdat. 30 extern void arch_refresh_nodedata(int nid, pg_data_t *pgdat); 39 * Because, pgdat for the new node is not allocated/initialized yet itself. 44 memblock_alloc(sizeof(*pgdat), SMP_CACHE_BYTES); \ 48 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) argument 50 node_data[nid] = pgdat; 61 static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) argument 230 static inline void pgdat_kswapd_lock(pg_data_t *pgdat) argument 232 mutex_lock(&pgdat->kswapd_lock); 235 static inline void pgdat_kswapd_unlock(pg_data_t *pgdat) argument 240 pgdat_kswapd_lock_init(pg_data_t *pgdat) argument 287 pgdat_kswapd_lock(pg_data_t *pgdat) argument 288 pgdat_kswapd_unlock(pg_data_t *pgdat) argument 289 pgdat_kswapd_lock_init(pg_data_t *pgdat) argument 304 pgdat_resize_lock(struct pglist_data *pgdat, unsigned long *flags) argument 309 pgdat_resize_unlock(struct pglist_data *pgdat, unsigned long *flags) argument 314 pgdat_resize_init(struct pglist_data *pgdat) argument 324 pgdat_resize_init(struct pglist_data *pgdat) argument [all...] |
H A D | compaction.h | 91 extern void reset_isolation_suitable(pg_data_t *pgdat); 103 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx); 106 static inline void reset_isolation_suitable(pg_data_t *pgdat) argument 123 static inline void wakeup_kcompactd(pg_data_t *pgdat, argument
|
H A D | bootmem_info.h | 21 void __init register_page_bootmem_info_node(struct pglist_data *pgdat); 48 static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) argument
|
H A D | memory-tiers.h | 53 void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets); 61 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) argument 104 static inline void node_get_allowed_targets(pg_data_t *pgdat, nodemask_t *targets) argument
|
H A D | node.h | 131 struct pglist_data *pgdat = NODE_DATA(nid); local 132 unsigned long start_pfn = pgdat->node_start_pfn; 133 unsigned long end_pfn = start_pfn + pgdat->node_spanned_pages;
|
H A D | page_ext.h | 58 extern void pgdat_page_ext_init(struct pglist_data *pgdat); 105 static inline void pgdat_page_ext_init(struct pglist_data *pgdat) argument
|
H A D | vmstat.h | 172 static inline void node_page_state_add(long x, struct pglist_data *pgdat, argument 175 atomic_long_add(x, &pgdat->vm_stat[item]); 261 extern unsigned long node_page_state(struct pglist_data *pgdat, 263 extern unsigned long node_page_state_pages(struct pglist_data *pgdat, 311 void set_pgdat_percpu_threshold(pg_data_t *pgdat, 325 static inline void __mod_node_page_state(struct pglist_data *pgdat, argument 339 node_page_state_add(delta, pgdat, item); 348 static inline void __inc_node_state(struct pglist_data *pgdat, enum node_stat_item item) argument 350 atomic_long_inc(&pgdat->vm_stat[item]); 360 static inline void __dec_node_state(struct pglist_data *pgdat, enu argument [all...] |
H A D | memcontrol.h | 59 pg_data_t *pgdat; member in struct:mem_cgroup_reclaim_cookie 731 * @pgdat: pglist_data 734 * @pgdat combination. This can be the node lruvec, if the memory 738 struct pglist_data *pgdat) 744 lruvec = &pgdat->__lruvec; 751 mz = memcg->nodeinfo[pgdat->node_id]; 756 * we have to be prepared to initialize lruvec->pgdat here; 759 if (unlikely(lruvec->pgdat != pgdat)) 760 lruvec->pgdat 737 mem_cgroup_lruvec(struct mem_cgroup *memcg, struct pglist_data *pgdat) argument 1314 mem_cgroup_lruvec(struct mem_cgroup *memcg, struct pglist_data *pgdat) argument 1322 struct pglist_data *pgdat = folio_pgdat(folio); local 1378 struct pglist_data *pgdat = folio_pgdat(folio); local 1386 struct pglist_data *pgdat = folio_pgdat(folio); local 1395 struct pglist_data *pgdat = folio_pgdat(folio); local 1627 mem_cgroup_soft_limit_reclaim(pg_data_t *pgdat, int order, gfp_t gfp_mask, unsigned long *total_scanned) argument [all...] |
H A D | mmzone.h | 555 void lru_gen_init_pgdat(struct pglist_data *pgdat); 568 static inline void lru_gen_init_pgdat(struct pglist_data *pgdat) argument 632 struct pglist_data *pgdat; member in struct:lruvec 1301 * Also synchronizes pgdat->first_deferred_pfn during deferred page 1423 static inline unsigned long pgdat_end_pfn(pg_data_t *pgdat) argument 1425 return pgdat->node_start_pfn + pgdat->node_spanned_pages; 1430 void build_all_zonelists(pg_data_t *pgdat); 1458 return lruvec->pgdat; 1572 extern struct pglist_data *next_online_pgdat(struct pglist_data *pgdat); [all...] |
/linux-master/tools/testing/memblock/ |
H A D | mmzone.c | 9 struct pglist_data *next_online_pgdat(struct pglist_data *pgdat) argument
|
/linux-master/drivers/base/ |
H A D | node.c | 375 struct pglist_data *pgdat = NODE_DATA(nid); local 381 sreclaimable = node_page_state_pages(pgdat, NR_SLAB_RECLAIMABLE_B); 382 sunreclaimable = node_page_state_pages(pgdat, NR_SLAB_UNRECLAIMABLE_B); 384 swapcached = node_page_state_pages(pgdat, NR_SWAPCACHE); 403 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON) + 404 node_page_state(pgdat, NR_ACTIVE_FILE)), 405 nid, K(node_page_state(pgdat, NR_INACTIVE_ANON) + 406 node_page_state(pgdat, NR_INACTIVE_FILE)), 407 nid, K(node_page_state(pgdat, NR_ACTIVE_ANON)), 408 nid, K(node_page_state(pgdat, NR_INACTIVE_ANO 520 struct pglist_data *pgdat = NODE_DATA(nid); local [all...] |