/linux-master/tools/testing/selftests/damon/ |
H A D | access_memory.c | 13 char **regions; local 30 regions = malloc(sizeof(*regions) * nr_regions); 32 regions[i] = malloc(sz_region); 38 memset(regions[i], i, 1024 * 1024 * 10);
|
/linux-master/arch/powerpc/mm/nohash/ |
H A D | kaslr_booke.c | 23 struct regions { struct 38 struct regions __initdata regions; variable in typeref:struct:__initdata 113 if (regions.reserved_mem < 0) 117 for (subnode = fdt_first_subnode(fdt, regions.reserved_mem); 125 while (len >= (regions.reserved_mem_addr_cells + 126 regions.reserved_mem_size_cells)) { 128 if (regions.reserved_mem_addr_cells == 2) 131 reg += regions.reserved_mem_addr_cells; 132 len -= 4 * regions [all...] |
/linux-master/mm/damon/ |
H A D | vaddr-test.h | 44 * discontiguous regions which cover every mapped areas. However, the three 45 * regions should not include the two biggest unmapped areas in the original 47 * heap and the mmap()-ed regions, and 2) the mmap()-ed regions and stack. 54 * three regions and returns. For more detail, refer to the comment of 60 * mapped. To cover every mappings, the three regions should start with 10, 63 * unmapped areas, and thus it should be converted to three regions of 10-25, 69 struct damon_addr_range regions[3] = {0,}; local 84 __damon_va_three_regions(&mm, regions); 86 KUNIT_EXPECT_EQ(test, 10ul, regions[ 129 damon_do_test_apply_three_regions(struct kunit *test, unsigned long *regions, int nr_regions, struct damon_addr_range *three_regions, unsigned long *expected, int nr_expected) argument 164 unsigned long regions[] = {10, 20, 20, 30, 50, 55, 55, 57, 57, 59, local 186 unsigned long regions[] = {10, 20, 20, 30, 50, 55, 55, 57, 57, 59, local 210 unsigned long regions[] = {10, 20, 20, 30, 50, 55, 55, 57, 57, 59, local 235 unsigned long regions[] = {10, 20, 20, 30, 50, 55, 55, 57, 57, 59, local [all...] |
/linux-master/drivers/virt/nitro_enclaves/ |
H A D | ne_misc_dev_test.c | 23 * regions = {} 34 * regions = {} 45 * regions = { 58 * regions = { 72 * regions = { 87 * regions = { 102 * regions = { 117 phys_contig_mem_regions.regions = kunit_kcalloc(test, MAX_PHYS_REGIONS, 118 sizeof(*phys_contig_mem_regions.regions), 120 KUNIT_ASSERT_TRUE(test, phys_contig_mem_regions.regions); [all...] |
/linux-master/drivers/mtd/chips/ |
H A D | jedec_probe.c | 275 const uint32_t regions[6]; member in struct:amd_flash_info 307 .regions = { 319 .regions = { 334 .regions = { 349 .regions = { 364 .regions = { 379 .regions = { 395 .regions = { 412 .regions = { 429 .regions [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/nvfw/ |
H A D | acr.c | 130 hdr->regions.no_regions); 132 for (i = 0; i < ARRAY_SIZE(hdr->regions.region_props); i++) { 135 hdr->regions.region_props[i].start_addr); 137 hdr->regions.region_props[i].end_addr); 139 hdr->regions.region_props[i].region_id); 141 hdr->regions.region_props[i].read_mask); 143 hdr->regions.region_props[i].write_mask); 145 hdr->regions.region_props[i].client_mask); 173 hdr->regions.no_regions); 175 for (i = 0; i < ARRAY_SIZE(hdr->regions [all...] |
/linux-master/arch/mips/include/asm/ |
H A D | yamon-dt.h | 38 * @regions: zero size terminated array of physical memory regions 41 * by YAMON in its environment and the @regions array. 46 const struct yamon_mem_region *regions);
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
H A D | nv1a.c | 31 .tile.regions = 8,
|
H A D | nv4e.c | 32 .tile.regions = 12,
|
H A D | nv49.c | 33 .tile.regions = 15,
|
H A D | nv47.c | 33 .tile.regions = 15,
|
/linux-master/drivers/vfio/cdx/ |
H A D | private.h | 25 struct vfio_cdx_region *regions; member in struct:vfio_cdx_device
|
H A D | main.c | 19 vdev->regions = kcalloc(count, sizeof(struct vfio_cdx_region), 21 if (!vdev->regions) 27 vdev->regions[i].addr = res->start; 28 vdev->regions[i].size = resource_size(res); 29 vdev->regions[i].type = res->flags; 31 * Only regions addressed with PAGE granularity may be 34 if (!(vdev->regions[i].addr & ~PAGE_MASK) && 35 !(vdev->regions[i].size & ~PAGE_MASK)) 36 vdev->regions[i].flags |= 38 vdev->regions[ [all...] |
/linux-master/drivers/net/dsa/sja1105/ |
H A D | sja1105_devlink.c | 7 /* Since devlink regions have a fixed size and the static config has a variable 85 priv->regions = kcalloc(num_regions, sizeof(struct devlink_region *), 87 if (!priv->regions) 97 dsa_devlink_region_destroy(priv->regions[i]); 99 kfree(priv->regions); 103 priv->regions[i] = region; 115 dsa_devlink_region_destroy(priv->regions[i]); 117 kfree(priv->regions);
|
/linux-master/drivers/vfio/platform/ |
H A D | vfio_platform_common.c | 144 vdev->regions = kcalloc(cnt, sizeof(struct vfio_platform_region), 146 if (!vdev->regions) 153 vdev->regions[i].addr = res->start; 154 vdev->regions[i].size = resource_size(res); 155 vdev->regions[i].flags = 0; 159 vdev->regions[i].type = VFIO_PLATFORM_REGION_TYPE_MMIO; 160 vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_READ; 162 vdev->regions[i].flags |= 166 * Only regions addressed with PAGE granularity may be 169 if (!(vdev->regions[ [all...] |
/linux-master/tools/testing/memblock/tests/ |
H A D | alloc_nid_api.c | 66 struct memblock_region *rgn = &memblock.reserved.regions[0]; 118 struct memblock_region *rgn = &memblock.reserved.regions[0]; 169 struct memblock_region *rgn = &memblock.reserved.regions[0]; 221 struct memblock_region *rgn = &memblock.reserved.regions[0]; 307 * Expect a merge of both regions. Only the region size gets updated. 311 struct memblock_region *rgn = &memblock.reserved.regions[0]; 359 * Expect a merge of regions. Only the region size gets updated. 363 struct memblock_region *rgn = &memblock.reserved.regions[0]; 399 * there are two reserved regions at the borders, with a gap big enough to fit 416 struct memblock_region *rgn1 = &memblock.reserved.regions[ [all...] |
H A D | basic_api.c | 17 ASSERT_NE(memblock.memory.regions, NULL); 22 ASSERT_NE(memblock.reserved.regions, NULL); 37 * and size to the collection of available memory regions (memblock.memory). 45 rgn = &memblock.memory.regions[0]; 70 * NUMA node and memory flags to the collection of available memory regions. 78 rgn = &memblock.memory.regions[0]; 114 * available memory regions (memblock.memory). The total size and 121 rgn1 = &memblock.memory.regions[0]; 122 rgn2 = &memblock.memory.regions[1]; 167 * and has size of two regions minu [all...] |
H A D | alloc_exact_nid_api.c | 30 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; 31 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; 82 struct memblock_region *new_rgn = &memblock.reserved.regions[1]; 83 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; 143 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; 144 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; 200 struct memblock_region *new_rgn = &memblock.reserved.regions[0]; 201 struct memblock_region *req_node = &memblock.memory.regions[nid_req]; 202 struct memblock_region *node2 = &memblock.memory.regions[6]; 258 struct memblock_region *new_rgn = &memblock.reserved.regions[ [all...] |
H A D | alloc_helpers_api.c | 20 struct memblock_region *rgn = &memblock.reserved.regions[0]; 63 struct memblock_region *rgn = &memblock.reserved.regions[0]; 110 struct memblock_region *rgn = &memblock.reserved.regions[0]; 149 * regions get merged into one. 153 struct memblock_region *rgn = &memblock.reserved.regions[0]; 186 * Expect successful allocation and merge of both regions. 190 struct memblock_region *rgn = &memblock.reserved.regions[0]; 236 struct memblock_region *rgn = &memblock.reserved.regions[0]; 278 struct memblock_region *rgn = &memblock.reserved.regions[0]; 314 struct memblock_region *rgn = &memblock.reserved.regions[ [all...] |
/linux-master/drivers/md/ |
H A D | dm-bio-prison-v1.c | 29 struct prison_region regions[] __counted_by(num_locks); 47 prison = kzalloc(struct_size(prison, regions, num_locks), GFP_KERNEL); 53 spin_lock_init(&prison->regions[i].lock); 54 prison->regions[i].cell = RB_ROOT; 184 spin_lock_irq(&prison->regions[l].lock); 185 r = __bio_detain(&prison->regions[l].cell, key, inmate, cell_prealloc, cell_result); 186 spin_unlock_irq(&prison->regions[l].lock); 232 spin_lock_irq(&prison->regions[l].lock); 233 __cell_release(&prison->regions[l].cell, cell, bios); 234 spin_unlock_irq(&prison->regions[ [all...] |
/linux-master/drivers/fpga/ |
H A D | dfl-afu-region.c | 21 INIT_LIST_HEAD(&afu->regions); 25 list_for_each_entry((region), &(afu)->regions, node) 79 list_add(®ion->node, &afu->regions); 93 * afu_mmio_region_destroy - destroy all mmio regions under given feature dev. 101 list_for_each_entry_safe(region, tmp, &afu->regions, node)
|
/linux-master/drivers/vfio/fsl-mc/ |
H A D | vfio_fsl_mc.c | 30 vdev->regions = kcalloc(count, sizeof(struct vfio_fsl_mc_region), 32 if (!vdev->regions) 36 struct resource *res = &mc_dev->regions[i]; 39 vdev->regions[i].addr = res->start; 40 vdev->regions[i].size = resource_size(res); 41 vdev->regions[i].type = mc_dev->regions[i].flags & IORESOURCE_BITS; 43 * Only regions addressed with PAGE granularity may be 46 if (!no_mmap && !(vdev->regions[i].addr & ~PAGE_MASK) && 47 !(vdev->regions[ [all...] |
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/acr/ |
H A D | gp102.c | 203 desc->regions.no_regions = 2; 204 desc->regions.region_props[0].start_addr = acr->wpr_start >> 8; 205 desc->regions.region_props[0].end_addr = acr->wpr_end >> 8; 206 desc->regions.region_props[0].region_id = 1; 207 desc->regions.region_props[0].read_mask = 0xf; 208 desc->regions.region_props[0].write_mask = 0xc; 209 desc->regions.region_props[0].client_mask = 0x2; 210 desc->regions.region_props[0].shadow_mem_start_addr = acr->shadow_start >> 8;
|
/linux-master/mm/ |
H A D | memblock.c | 39 * Memblock is a method of managing memory regions during the early 44 * regions. There are several types of these collections: 50 * * ``reserved`` - describes the regions that were allocated 58 * which contains an array of memory regions along with 66 * arrays during addition of new regions. This feature should be used 116 .memory.regions = memblock_memory_init_regions, 121 .reserved.regions = memblock_reserved_init_regions, 132 .regions = memblock_physmem_init_regions, 148 for (i = 0, rgn = &memblock_type->regions[0]; \ 150 i++, rgn = &memblock_type->regions[ [all...] |
/linux-master/arch/mips/generic/ |
H A D | yamon-dt.c | 44 const struct yamon_mem_region *regions, 53 for (mr = regions; mr->size && memsize; ++mr) { 55 pr_warn("Number of regions exceeds max %u\n", 76 const struct yamon_mem_region *regions) 136 mem_entries = gen_fdt_mem_array(regions, mem_array, 145 mem_entries = gen_fdt_mem_array(regions, mem_array, 43 gen_fdt_mem_array( const struct yamon_mem_region *regions, __be32 *mem_array, unsigned int max_entries, unsigned long memsize) argument 75 yamon_dt_append_memory(void *fdt, const struct yamon_mem_region *regions) argument
|