Lines Matching refs:nid

42 				       struct shrinker_info *old, int nid)
50 unit = kzalloc_node(sizeof(*unit), GFP_KERNEL, nid);
66 int nid;
68 for_each_node(nid) {
69 pn = memcg->nodeinfo[nid];
80 int nid, ret = 0;
85 for_each_node(nid) {
86 info = kvzalloc_node(sizeof(*info) + array_size, GFP_KERNEL, nid);
90 if (shrinker_unit_alloc(info, NULL, nid))
92 rcu_assign_pointer(memcg->nodeinfo[nid]->shrinker_info, info);
105 int nid)
107 return rcu_dereference_protected(memcg->nodeinfo[nid]->shrinker_info,
116 int nid;
118 for_each_node(nid) {
119 pn = memcg->nodeinfo[nid];
120 old = shrinker_info_protected(memcg, nid);
129 new = kvzalloc_node(sizeof(*new) + new_size, GFP_KERNEL, nid);
136 if (shrinker_unit_alloc(new, old, nid)) {
194 void set_shrinker_bit(struct mem_cgroup *memcg, int nid, int shrinker_id)
201 info = rcu_dereference(memcg->nodeinfo[nid]->shrinker_info);
250 static long xchg_nr_deferred_memcg(int nid, struct shrinker *shrinker,
258 info = rcu_dereference(memcg->nodeinfo[nid]->shrinker_info);
266 static long add_nr_deferred_memcg(long nr, int nid, struct shrinker *shrinker,
274 info = rcu_dereference(memcg->nodeinfo[nid]->shrinker_info);
285 int nid, index, offset;
297 for_each_node(nid) {
298 child_info = shrinker_info_protected(memcg, nid);
299 parent_info = shrinker_info_protected(parent, nid);
321 static long xchg_nr_deferred_memcg(int nid, struct shrinker *shrinker,
327 static long add_nr_deferred_memcg(long nr, int nid, struct shrinker *shrinker,
337 int nid = sc->nid;
340 nid = 0;
344 return xchg_nr_deferred_memcg(nid, shrinker,
347 return atomic_long_xchg(&shrinker->nr_deferred[nid], 0);
354 int nid = sc->nid;
357 nid = 0;
361 return add_nr_deferred_memcg(nr, nid, shrinker,
364 return atomic_long_add_return(nr, &shrinker->nr_deferred[nid]);
462 trace_mm_shrink_slab_end(shrinker, shrinkctl->nid, freed, nr, new_nr, total_scan);
467 static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid,
514 info = rcu_dereference(memcg->nodeinfo[nid]->shrinker_info);
528 .nid = nid,
571 set_shrinker_bit(memcg, nid, shrinker_id);
585 static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid,
595 * @nid: node whose slab caches to target
601 * @nid is passed along to shrinkers with SHRINKER_NUMA_AWARE set,
612 unsigned long shrink_slab(gfp_t gfp_mask, int nid, struct mem_cgroup *memcg,
626 return shrink_slab_memcg(gfp_mask, nid, memcg, priority);
653 .nid = nid,