• Home
  • History
  • Annotate
  • Raw
  • Download
  • only in /netgear-R7000-V1.0.7.12_1.2.5/components/opensource/linux/linux-2.6.36/mm/

Lines Matching refs:early_node_map

204   static struct node_active_region __meminitdata early_node_map[MAX_ACTIVE_REGIONS];
3566 if (nid == MAX_NUMNODES || early_node_map[i].nid == nid)
3579 if (nid == MAX_NUMNODES || early_node_map[index].nid == nid)
3597 unsigned long start_pfn = early_node_map[i].start_pfn;
3598 unsigned long end_pfn = early_node_map[i].end_pfn;
3601 return early_node_map[i].nid;
3631 /* Basic iterator support to walk early_node_map[] */
3652 unsigned long end_pfn = early_node_map[i].end_pfn;
3654 if (early_node_map[i].start_pfn >= max_low_pfn)
3660 size_pages = end_pfn - early_node_map[i].start_pfn;
3661 free_bootmem_node(NODE_DATA(early_node_map[i].nid),
3662 PFN_PHYS(early_node_map[i].start_pfn),
3673 /* need to go over early_node_map to find out good range for node */
3675 start = early_node_map[i].start_pfn;
3676 end = early_node_map[i].end_pfn;
3692 /* need to go over early_node_map to find out good range for node */
3697 ei_last = early_node_map[i].end_pfn;
3699 ei_start = early_node_map[i].start_pfn;
3730 ret = work_fn(early_node_map[i].start_pfn,
3731 early_node_map[i].end_pfn, data);
3749 memory_present(early_node_map[i].nid,
3750 early_node_map[i].start_pfn,
3751 early_node_map[i].end_pfn);
3773 *start_pfn = min(*start_pfn, early_node_map[i].start_pfn);
3774 *end_pfn = max(*end_pfn, early_node_map[i].end_pfn);
3886 prev_end_pfn = min(early_node_map[i].start_pfn, range_end_pfn);
3889 if (early_node_map[i].start_pfn > range_start_pfn)
3900 start_pfn = min(early_node_map[i].start_pfn, range_end_pfn);
3908 prev_end_pfn = early_node_map[i].end_pfn;
4249 * These ranges are stored in an early_node_map[] and later used by
4270 if (early_node_map[i].nid != nid)
4274 if (start_pfn >= early_node_map[i].start_pfn &&
4275 end_pfn <= early_node_map[i].end_pfn)
4279 if (start_pfn <= early_node_map[i].end_pfn &&
4280 end_pfn > early_node_map[i].end_pfn) {
4281 early_node_map[i].end_pfn = end_pfn;
4286 if (start_pfn < early_node_map[i].start_pfn &&
4287 end_pfn >= early_node_map[i].start_pfn) {
4288 early_node_map[i].start_pfn = start_pfn;
4293 /* Check that early_node_map is large enough */
4300 early_node_map[i].nid = nid;
4301 early_node_map[i].start_pfn = start_pfn;
4302 early_node_map[i].end_pfn = end_pfn;
4328 if (early_node_map[i].start_pfn >= start_pfn &&
4329 early_node_map[i].end_pfn <= end_pfn) {
4331 early_node_map[i].start_pfn = 0;
4332 early_node_map[i].end_pfn = 0;
4336 if (early_node_map[i].start_pfn < start_pfn &&
4337 early_node_map[i].end_pfn > start_pfn) {
4338 unsigned long temp_end_pfn = early_node_map[i].end_pfn;
4339 early_node_map[i].end_pfn = start_pfn;
4344 if (early_node_map[i].start_pfn >= start_pfn &&
4345 early_node_map[i].end_pfn > end_pfn &&
4346 early_node_map[i].start_pfn < end_pfn) {
4347 early_node_map[i].start_pfn = end_pfn;
4357 if (early_node_map[i].nid != nid)
4359 if (early_node_map[i].end_pfn)
4363 memcpy(&early_node_map[j], &early_node_map[j+1],
4364 sizeof(early_node_map[j]));
4366 memset(&early_node_map[j], 0, sizeof(early_node_map[j]));
4380 memset(early_node_map, 0, sizeof(early_node_map));
4402 sort(early_node_map, (size_t)nr_nodemap_entries,
4415 min_pfn = min(min_pfn, early_node_map[i].start_pfn);
4448 unsigned long pages = early_node_map[i].end_pfn -
4449 early_node_map[i].start_pfn;
4452 node_set_state(early_node_map[i].nid, N_HIGH_MEMORY);
4527 start_pfn = max(early_node_map[i].start_pfn,
4529 end_pfn = early_node_map[i].end_pfn;
4634 /* Sort early_node_map as initialisation assumes it is sorted */
4681 /* Print out the early_node_map[] */
4682 printk("early_node_map[%d] active PFN ranges\n", nr_nodemap_entries);
4684 printk(" %3d: %0#10lx -> %0#10lx\n", early_node_map[i].nid,
4685 early_node_map[i].start_pfn,
4686 early_node_map[i].end_pfn);