Searched refs:regions (Results 1 - 25 of 169) sorted by relevance

1234567

/linux-master/tools/testing/selftests/damon/
H A Daccess_memory.c13 char **regions; local
30 regions = malloc(sizeof(*regions) * nr_regions);
32 regions[i] = malloc(sz_region);
38 memset(regions[i], i, 1024 * 1024 * 10);
/linux-master/arch/powerpc/mm/nohash/
H A Dkaslr_booke.c23 struct regions { struct
38 struct regions __initdata regions; variable in typeref:struct:__initdata
113 if (regions.reserved_mem < 0)
117 for (subnode = fdt_first_subnode(fdt, regions.reserved_mem);
125 while (len >= (regions.reserved_mem_addr_cells +
126 regions.reserved_mem_size_cells)) {
128 if (regions.reserved_mem_addr_cells == 2)
131 reg += regions.reserved_mem_addr_cells;
132 len -= 4 * regions
[all...]
/linux-master/mm/damon/
H A Dvaddr-test.h44 * discontiguous regions which cover every mapped areas. However, the three
45 * regions should not include the two biggest unmapped areas in the original
47 * heap and the mmap()-ed regions, and 2) the mmap()-ed regions and stack.
54 * three regions and returns. For more detail, refer to the comment of
60 * mapped. To cover every mappings, the three regions should start with 10,
63 * unmapped areas, and thus it should be converted to three regions of 10-25,
69 struct damon_addr_range regions[3] = {0,}; local
84 __damon_va_three_regions(&mm, regions);
86 KUNIT_EXPECT_EQ(test, 10ul, regions[
129 damon_do_test_apply_three_regions(struct kunit *test, unsigned long *regions, int nr_regions, struct damon_addr_range *three_regions, unsigned long *expected, int nr_expected) argument
164 unsigned long regions[] = {10, 20, 20, 30, 50, 55, 55, 57, 57, 59, local
186 unsigned long regions[] = {10, 20, 20, 30, 50, 55, 55, 57, 57, 59, local
210 unsigned long regions[] = {10, 20, 20, 30, 50, 55, 55, 57, 57, 59, local
235 unsigned long regions[] = {10, 20, 20, 30, 50, 55, 55, 57, 57, 59, local
[all...]
/linux-master/drivers/virt/nitro_enclaves/
H A Dne_misc_dev_test.c23 * regions = {}
34 * regions = {}
45 * regions = {
58 * regions = {
72 * regions = {
87 * regions = {
102 * regions = {
117 phys_contig_mem_regions.regions = kunit_kcalloc(test, MAX_PHYS_REGIONS,
118 sizeof(*phys_contig_mem_regions.regions),
120 KUNIT_ASSERT_TRUE(test, phys_contig_mem_regions.regions);
[all...]
/linux-master/drivers/mtd/chips/
H A Djedec_probe.c275 const uint32_t regions[6]; member in struct:amd_flash_info
307 .regions = {
319 .regions = {
334 .regions = {
349 .regions = {
364 .regions = {
379 .regions = {
395 .regions = {
412 .regions = {
429 .regions
[all...]
/linux-master/drivers/gpu/drm/nouveau/nvkm/nvfw/
H A Dacr.c130 hdr->regions.no_regions);
132 for (i = 0; i < ARRAY_SIZE(hdr->regions.region_props); i++) {
135 hdr->regions.region_props[i].start_addr);
137 hdr->regions.region_props[i].end_addr);
139 hdr->regions.region_props[i].region_id);
141 hdr->regions.region_props[i].read_mask);
143 hdr->regions.region_props[i].write_mask);
145 hdr->regions.region_props[i].client_mask);
173 hdr->regions.no_regions);
175 for (i = 0; i < ARRAY_SIZE(hdr->regions
[all...]
/linux-master/arch/mips/include/asm/
H A Dyamon-dt.h38 * @regions: zero size terminated array of physical memory regions
41 * by YAMON in its environment and the @regions array.
46 const struct yamon_mem_region *regions);
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
H A Dnv1a.c31 .tile.regions = 8,
H A Dnv4e.c32 .tile.regions = 12,
H A Dnv49.c33 .tile.regions = 15,
H A Dnv47.c33 .tile.regions = 15,
/linux-master/drivers/vfio/cdx/
H A Dprivate.h25 struct vfio_cdx_region *regions; member in struct:vfio_cdx_device
H A Dmain.c19 vdev->regions = kcalloc(count, sizeof(struct vfio_cdx_region),
21 if (!vdev->regions)
27 vdev->regions[i].addr = res->start;
28 vdev->regions[i].size = resource_size(res);
29 vdev->regions[i].type = res->flags;
31 * Only regions addressed with PAGE granularity may be
34 if (!(vdev->regions[i].addr & ~PAGE_MASK) &&
35 !(vdev->regions[i].size & ~PAGE_MASK))
36 vdev->regions[i].flags |=
38 vdev->regions[
[all...]
/linux-master/drivers/net/dsa/sja1105/
H A Dsja1105_devlink.c7 /* Since devlink regions have a fixed size and the static config has a variable
85 priv->regions = kcalloc(num_regions, sizeof(struct devlink_region *),
87 if (!priv->regions)
97 dsa_devlink_region_destroy(priv->regions[i]);
99 kfree(priv->regions);
103 priv->regions[i] = region;
115 dsa_devlink_region_destroy(priv->regions[i]);
117 kfree(priv->regions);
/linux-master/drivers/vfio/platform/
H A Dvfio_platform_common.c144 vdev->regions = kcalloc(cnt, sizeof(struct vfio_platform_region),
146 if (!vdev->regions)
153 vdev->regions[i].addr = res->start;
154 vdev->regions[i].size = resource_size(res);
155 vdev->regions[i].flags = 0;
159 vdev->regions[i].type = VFIO_PLATFORM_REGION_TYPE_MMIO;
160 vdev->regions[i].flags |= VFIO_REGION_INFO_FLAG_READ;
162 vdev->regions[i].flags |=
166 * Only regions addressed with PAGE granularity may be
169 if (!(vdev->regions[
[all...]
/linux-master/tools/testing/memblock/tests/
H A Dalloc_nid_api.c66 struct memblock_region *rgn = &memblock.reserved.regions[0];
118 struct memblock_region *rgn = &memblock.reserved.regions[0];
169 struct memblock_region *rgn = &memblock.reserved.regions[0];
221 struct memblock_region *rgn = &memblock.reserved.regions[0];
307 * Expect a merge of both regions. Only the region size gets updated.
311 struct memblock_region *rgn = &memblock.reserved.regions[0];
359 * Expect a merge of regions. Only the region size gets updated.
363 struct memblock_region *rgn = &memblock.reserved.regions[0];
399 * there are two reserved regions at the borders, with a gap big enough to fit
416 struct memblock_region *rgn1 = &memblock.reserved.regions[
[all...]
H A Dbasic_api.c17 ASSERT_NE(memblock.memory.regions, NULL);
22 ASSERT_NE(memblock.reserved.regions, NULL);
37 * and size to the collection of available memory regions (memblock.memory).
45 rgn = &memblock.memory.regions[0];
70 * NUMA node and memory flags to the collection of available memory regions.
78 rgn = &memblock.memory.regions[0];
114 * available memory regions (memblock.memory). The total size and
121 rgn1 = &memblock.memory.regions[0];
122 rgn2 = &memblock.memory.regions[1];
167 * and has size of two regions minu
[all...]
H A Dalloc_exact_nid_api.c30 struct memblock_region *new_rgn = &memblock.reserved.regions[0];
31 struct memblock_region *req_node = &memblock.memory.regions[nid_req];
82 struct memblock_region *new_rgn = &memblock.reserved.regions[1];
83 struct memblock_region *req_node = &memblock.memory.regions[nid_req];
143 struct memblock_region *new_rgn = &memblock.reserved.regions[0];
144 struct memblock_region *req_node = &memblock.memory.regions[nid_req];
200 struct memblock_region *new_rgn = &memblock.reserved.regions[0];
201 struct memblock_region *req_node = &memblock.memory.regions[nid_req];
202 struct memblock_region *node2 = &memblock.memory.regions[6];
258 struct memblock_region *new_rgn = &memblock.reserved.regions[
[all...]
H A Dalloc_helpers_api.c20 struct memblock_region *rgn = &memblock.reserved.regions[0];
63 struct memblock_region *rgn = &memblock.reserved.regions[0];
110 struct memblock_region *rgn = &memblock.reserved.regions[0];
149 * regions get merged into one.
153 struct memblock_region *rgn = &memblock.reserved.regions[0];
186 * Expect successful allocation and merge of both regions.
190 struct memblock_region *rgn = &memblock.reserved.regions[0];
236 struct memblock_region *rgn = &memblock.reserved.regions[0];
278 struct memblock_region *rgn = &memblock.reserved.regions[0];
314 struct memblock_region *rgn = &memblock.reserved.regions[
[all...]
/linux-master/drivers/md/
H A Ddm-bio-prison-v1.c29 struct prison_region regions[] __counted_by(num_locks);
47 prison = kzalloc(struct_size(prison, regions, num_locks), GFP_KERNEL);
53 spin_lock_init(&prison->regions[i].lock);
54 prison->regions[i].cell = RB_ROOT;
184 spin_lock_irq(&prison->regions[l].lock);
185 r = __bio_detain(&prison->regions[l].cell, key, inmate, cell_prealloc, cell_result);
186 spin_unlock_irq(&prison->regions[l].lock);
232 spin_lock_irq(&prison->regions[l].lock);
233 __cell_release(&prison->regions[l].cell, cell, bios);
234 spin_unlock_irq(&prison->regions[
[all...]
/linux-master/drivers/fpga/
H A Ddfl-afu-region.c21 INIT_LIST_HEAD(&afu->regions);
25 list_for_each_entry((region), &(afu)->regions, node)
79 list_add(&region->node, &afu->regions);
93 * afu_mmio_region_destroy - destroy all mmio regions under given feature dev.
101 list_for_each_entry_safe(region, tmp, &afu->regions, node)
/linux-master/drivers/vfio/fsl-mc/
H A Dvfio_fsl_mc.c30 vdev->regions = kcalloc(count, sizeof(struct vfio_fsl_mc_region),
32 if (!vdev->regions)
36 struct resource *res = &mc_dev->regions[i];
39 vdev->regions[i].addr = res->start;
40 vdev->regions[i].size = resource_size(res);
41 vdev->regions[i].type = mc_dev->regions[i].flags & IORESOURCE_BITS;
43 * Only regions addressed with PAGE granularity may be
46 if (!no_mmap && !(vdev->regions[i].addr & ~PAGE_MASK) &&
47 !(vdev->regions[
[all...]
/linux-master/drivers/gpu/drm/nouveau/nvkm/subdev/acr/
H A Dgp102.c203 desc->regions.no_regions = 2;
204 desc->regions.region_props[0].start_addr = acr->wpr_start >> 8;
205 desc->regions.region_props[0].end_addr = acr->wpr_end >> 8;
206 desc->regions.region_props[0].region_id = 1;
207 desc->regions.region_props[0].read_mask = 0xf;
208 desc->regions.region_props[0].write_mask = 0xc;
209 desc->regions.region_props[0].client_mask = 0x2;
210 desc->regions.region_props[0].shadow_mem_start_addr = acr->shadow_start >> 8;
/linux-master/mm/
H A Dmemblock.c39 * Memblock is a method of managing memory regions during the early
44 * regions. There are several types of these collections:
50 * * ``reserved`` - describes the regions that were allocated
58 * which contains an array of memory regions along with
66 * arrays during addition of new regions. This feature should be used
116 .memory.regions = memblock_memory_init_regions,
121 .reserved.regions = memblock_reserved_init_regions,
132 .regions = memblock_physmem_init_regions,
148 for (i = 0, rgn = &memblock_type->regions[0]; \
150 i++, rgn = &memblock_type->regions[
[all...]
/linux-master/arch/mips/generic/
H A Dyamon-dt.c44 const struct yamon_mem_region *regions,
53 for (mr = regions; mr->size && memsize; ++mr) {
55 pr_warn("Number of regions exceeds max %u\n",
76 const struct yamon_mem_region *regions)
136 mem_entries = gen_fdt_mem_array(regions, mem_array,
145 mem_entries = gen_fdt_mem_array(regions, mem_array,
43 gen_fdt_mem_array( const struct yamon_mem_region *regions, __be32 *mem_array, unsigned int max_entries, unsigned long memsize) argument
75 yamon_dt_append_memory(void *fdt, const struct yamon_mem_region *regions) argument

Completed in 196 milliseconds

1234567