Searched refs:ZONE_DMA (Results 1 - 20 of 20) sorted by relevance
/linux-master/arch/m68k/mm/ |
H A D | init.c | 75 max_zone_pfn[ZONE_DMA] = end_mem >> PAGE_SHIFT;
|
H A D | sun3mmu.c | 91 max_zone_pfn[ZONE_DMA] = ((unsigned long)high_memory) >> PAGE_SHIFT;
|
H A D | mcfmmu.c | 82 max_zone_pfn[ZONE_DMA] = PFN_DOWN(_ramend);
|
H A D | motorola.c | 511 max_zone_pfn[ZONE_DMA] = memblock_end_of_DRAM();
|
/linux-master/arch/microblaze/mm/ |
H A D | init.c | 88 zones_size[ZONE_DMA] = max_low_pfn; 91 zones_size[ZONE_DMA] = max_pfn;
|
/linux-master/drivers/crypto/ |
H A D | omap-crypto.c | 105 if (page_zonenum(sg_page(sg)) != ZONE_DMA)
|
H A D | omap-sham.c | 780 if (page_zonenum(sg_page(sg_tmp)) != ZONE_DMA) {
|
/linux-master/arch/alpha/mm/ |
H A D | init.c | 249 max_zone_pfn[ZONE_DMA] = dma_pfn;
|
/linux-master/arch/loongarch/mm/ |
H A D | init.c | 68 max_zone_pfns[ZONE_DMA] = MAX_DMA_PFN;
|
/linux-master/arch/powerpc/mm/ |
H A D | mem.c | 213 * We setup ZONE_DMA to be 31-bits on all platforms and ZONE_NORMAL to be 215 * ZONE_DMA. 220 * ZONE_DMA. 258 max_zone_pfns[ZONE_DMA] = min(max_low_pfn,
|
/linux-master/arch/s390/mm/ |
H A D | init.c | 100 max_zone_pfns[ZONE_DMA] = virt_to_pfn(MAX_DMA_ADDRESS);
|
/linux-master/arch/arm64/mm/ |
H A D | init.c | 60 * If the corresponding config options are enabled, we create both ZONE_DMA 61 * and ZONE_DMA32. By default ZONE_DMA covers the 32-bit addressable memory 146 max_zone_pfns[ZONE_DMA] = PFN_DOWN(arm64_dma_phys_limit); 375 * If no bouncing needed for ZONE_DMA, reduce the swiotlb
|
/linux-master/include/trace/events/ |
H A D | mmflags.h | 265 IFDEF_ZONE_DMA( EM (ZONE_DMA, "DMA")) \
|
/linux-master/include/linux/ |
H A D | gfp.h | 47 #define OPT_ZONE_DMA ZONE_DMA
|
H A D | mmzone.h | 729 * ZONE_DMA and ZONE_DMA32 are used when there are peripherals not able 732 * space ZONE_DMA32 is used. ZONE_DMA is left for the ones with smaller 739 ZONE_DMA, enumerator in enum:zone_type 1471 * zone_idx() returns 0 for the ZONE_DMA zone, 1 for the ZONE_NORMAL zone, etc.
|
/linux-master/arch/arm/mm/ |
H A D | init.c | 115 max_zone_pfn[ZONE_DMA] = min(arm_dma_pfn_limit, max_low);
|
/linux-master/arch/mips/mm/ |
H A D | init.c | 411 max_zone_pfns[ZONE_DMA] = MAX_DMA_PFN;
|
/linux-master/arch/x86/mm/ |
H A D | init.c | 1047 max_zone_pfns[ZONE_DMA] = min(MAX_DMA_PFN, max_low_pfn);
|
/linux-master/arch/sparc/mm/ |
H A D | srmmu.c | 978 max_zone_pfn[ZONE_DMA] = max_low_pfn;
|
/linux-master/mm/ |
H A D | page_alloc.c | 238 * NORMAL allocation will leave 784M/256 of ram reserved in the ZONE_DMA 240 * HIGHMEM allocation will leave (224M+784M)/256 of ram reserved in ZONE_DMA 247 [ZONE_DMA] = 256, 3124 * are worse than fragmentation. If the next zone is ZONE_DMA then it is 4943 * watermark within ZONE_DMA and ZONE_NORMAL. 4945 * Return: number of pages beyond high watermark within ZONE_DMA and 6785 struct zone *zone = &pgdat->node_zones[ZONE_DMA];
|
Completed in 235 milliseconds