/linux-master/kernel/module/ |
H A D | decompress.c | 19 struct page **new_pages; local 21 new_pages = kvmalloc_array(info->max_pages + extent, 23 if (!new_pages) 26 memcpy(new_pages, info->pages, info->max_pages * sizeof(info->pages)); 28 info->pages = new_pages;
|
/linux-master/drivers/block/drbd/ |
H A D | drbd_bitmap.c | 381 struct page **new_pages, *page; local 397 new_pages = kzalloc(bytes, GFP_NOIO | __GFP_NOWARN); 398 if (!new_pages) { 399 new_pages = __vmalloc(bytes, GFP_NOIO | __GFP_ZERO); 400 if (!new_pages) 406 new_pages[i] = old_pages[i]; 410 bm_free_pages(new_pages + have, i - have); 411 bm_vk_free(new_pages); 417 new_pages[i] = page; 421 new_pages[ [all...] |
/linux-master/arch/s390/kvm/ |
H A D | pci.c | 207 unsigned long page_limit, cur_pages, new_pages; local 213 new_pages = cur_pages + nr_pages; 214 if (new_pages > page_limit) 217 new_pages) != cur_pages);
|
/linux-master/kernel/trace/ |
H A D | ring_buffer.c | 489 struct list_head new_pages; /* new pages to add */ member in struct:ring_buffer_per_cpu 1620 INIT_LIST_HEAD(&cpu_buffer->new_pages); 1916 struct list_head *pages = &cpu_buffer->new_pages; 1985 list_for_each_entry_safe(bpage, tmp, &cpu_buffer->new_pages, 2083 INIT_LIST_HEAD(&cpu_buffer->new_pages); 2085 &cpu_buffer->new_pages)) { 2154 INIT_LIST_HEAD(&cpu_buffer->new_pages); 2157 &cpu_buffer->new_pages)) { 2220 if (list_empty(&cpu_buffer->new_pages)) 2223 list_for_each_entry_safe(bpage, tmp, &cpu_buffer->new_pages, [all...] |
/linux-master/drivers/base/firmware_loader/ |
H A D | main.c | 264 struct page **new_pages; local 266 new_pages = kvmalloc_array(new_array_size, sizeof(void *), 268 if (!new_pages) 270 memcpy(new_pages, fw_priv->pages, 272 memset(&new_pages[fw_priv->page_array_size], 0, sizeof(void *) * 275 fw_priv->pages = new_pages;
|
/linux-master/io_uring/ |
H A D | rsrc.c | 43 unsigned long page_limit, cur_pages, new_pages; local 53 new_pages = cur_pages + nr_pages; 54 if (new_pages > page_limit) 57 &cur_pages, new_pages));
|
/linux-master/drivers/virtio/ |
H A D | virtio_mem.c | 398 int new_pages = PFN_UP(new_bytes); local 401 if (vm->bbm.bb_states && old_pages == new_pages) 404 new_array = vzalloc(new_pages * PAGE_SIZE); 464 int new_pages = PFN_UP(vm->sbm.next_mb_id - vm->sbm.first_mb_id + 1); local 467 if (vm->sbm.mb_states && old_pages == new_pages) 470 new_array = vzalloc(new_pages * PAGE_SIZE); 587 int new_pages = PFN_UP(BITS_TO_LONGS(new_nb_bits) * sizeof(long)); local 590 if (vm->sbm.sb_states && old_pages == new_pages) 593 new_bitmap = vzalloc(new_pages * PAGE_SIZE);
|
/linux-master/arch/s390/kernel/ |
H A D | debug.c | 1309 int rc, new_pages; local 1323 new_pages = debug_get_uint(str); 1324 if (new_pages < 0) { 1328 rc = debug_set_size(id, id->nr_areas, new_pages);
|
/linux-master/drivers/md/dm-vdo/ |
H A D | encodings.c | 450 block_count_t new_pages; local 454 new_pages = level_size; 456 new_pages -= old_sizes->levels[height]; 457 total_pages += (new_pages * root_count);
|
H A D | block-map.c | 2337 static int make_segment(struct forest *old_forest, block_count_t new_pages, argument 2359 result = vdo_allocate(new_pages, struct tree_page, 2449 block_count_t new_pages; local 2455 new_pages = vdo_compute_new_forest_pages(map->root_count, old_boundary, 2457 if (new_pages == 0) { 2469 result = make_segment(old_forest, new_pages, &new_boundary, forest);
|
/linux-master/drivers/iommu/iommufd/ |
H A D | pages.c | 808 unsigned long new_pages; local 814 new_pages = cur_pages + npages; 815 if (new_pages > lock_limit) 818 new_pages) != cur_pages);
|
/linux-master/drivers/gpu/drm/imagination/ |
H A D | pvr_free_list.c | 387 resp->new_pages = free_list->current_pages + free_list->ready_pages;
|
H A D | pvr_rogue_fwif_check.h | 264 OFFSET_CHECK(struct rogue_fwif_freelist_gs_data, new_pages, 8);
|
H A D | pvr_rogue_fwif.h | 1027 u32 new_pages; member in struct:rogue_fwif_freelist_gs_data
|