/arch/sh/mm/ |
H A D | numa.c | 30 unsigned long start_pfn, end_pfn; local 37 end_pfn = end >> PAGE_SHIFT; 44 __add_active_range(nid, start_pfn, end_pfn); 53 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; 56 bootmap_pages = bootmem_bootmap_pages(end_pfn - start_pfn); 60 start_pfn, end_pfn); local 62 free_bootmem_with_active_regions(nid, end_pfn);
|
H A D | init.c | 195 unsigned long start_pfn, end_pfn; local 200 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); 204 SMP_CACHE_BYTES, end_pfn << PAGE_SHIFT); 219 NODE_DATA(nid)->node_spanned_pages = end_pfn - start_pfn; 225 unsigned long end_pfn; local 234 end_pfn = p->node_start_pfn + p->node_spanned_pages; 242 init_bootmem_node(p, paddr >> PAGE_SHIFT, p->node_start_pfn, end_pfn); 244 free_bootmem_with_active_regions(nid, end_pfn); 270 unsigned long start_pfn, end_pfn; local 272 end_pfn [all...] |
/arch/sh/kernel/ |
H A D | swsusp.c | 25 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; local 27 return (pfn >= begin_pfn) && (pfn < end_pfn);
|
H A D | setup.c | 192 unsigned long end_pfn) 200 end = end_pfn << PAGE_SHIFT; 209 start_pfn, end_pfn); 234 PFN_PHYS(end_pfn - start_pfn), nid); 191 __add_active_range(unsigned int nid, unsigned long start_pfn, unsigned long end_pfn) argument
|
/arch/x86/mm/ |
H A D | init.c | 100 unsigned long start_pfn, unsigned long end_pfn, 103 if (start_pfn < end_pfn) { 107 mr[nr_range].end = end_pfn<<PAGE_SHIFT; 124 unsigned long start_pfn, end_pfn; local 175 end_pfn = 1<<(PMD_SHIFT - PAGE_SHIFT); 177 end_pfn = ((pos + (PMD_SIZE - 1))>>PMD_SHIFT) 180 end_pfn = ((pos + (PMD_SIZE - 1)) >> PMD_SHIFT) 183 if (end_pfn > (end >> PAGE_SHIFT)) 184 end_pfn = end >> PAGE_SHIFT; 185 if (start_pfn < end_pfn) { 99 save_mr(struct map_range *mr, int nr_range, unsigned long start_pfn, unsigned long end_pfn, unsigned long page_size_mask) argument [all...] |
H A D | numa_32.c | 51 printk(KERN_INFO "Node: %d, start_pfn: %lx, end_pfn: %lx\n", 63 unsigned long end_pfn) 65 unsigned long nr_pages = end_pfn - start_pfn; 183 unsigned long end_pfn = end >> PAGE_SHIFT; local 193 nid, start_pfn, end_pfn); 196 size = node_memmap_size_bytes(nid, start_pfn, end_pfn); 62 node_memmap_size_bytes(int nid, unsigned long start_pfn, unsigned long end_pfn) argument
|
H A D | init_32.c | 246 unsigned long start_pfn, end_pfn; local 257 end_pfn = end >> PAGE_SHIFT; 286 if (pfn >= end_pfn) 294 for (; pmd_idx < PTRS_PER_PMD && pfn < end_pfn; 333 for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; 427 unsigned long start_pfn, unsigned long end_pfn) 434 start_pfn, end_pfn); 436 start_pfn, end_pfn); 426 add_highpages_with_active_regions(int nid, unsigned long start_pfn, unsigned long end_pfn) argument
|
/arch/sh/include/asm/ |
H A D | mmzone.h | 42 unsigned long end_pfn);
|
/arch/powerpc/mm/ |
H A D | numa.c | 82 static int __cpuinit fake_numa_create_new_node(unsigned long end_pfn, argument 113 if ((end_pfn << PAGE_SHIFT) > mem) { 138 unsigned long start_pfn, end_pfn; local 141 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { 142 if (pfn >= start_pfn && pfn < end_pfn) { 145 node_ar->end_pfn = end_pfn; 805 unsigned long start_pfn, end_pfn; local 816 end_pfn = memblock_region_memory_end_pfn(reg); 818 fake_numa_create_new_node(end_pfn, 901 careful_zallocation(int nid, unsigned long size, unsigned long align, unsigned long end_pfn) argument 959 unsigned long end_pfn = PFN_UP(physbase + size); local 1032 unsigned long start_pfn, end_pfn; local [all...] |
H A D | mem.c | 149 unsigned long end_pfn = start_pfn + nr_pages; local 155 tend = min(end_pfn, memblock_region_memory_end_pfn(reg)); 200 unsigned long start_pfn, end_pfn; local 202 end_pfn = memblock_region_memory_end_pfn(reg);
|
/arch/x86/include/asm/ |
H A D | highmem.h | 73 unsigned long end_pfn);
|
H A D | mtrr.h | 135 extern int mtrr_trim_uncached_memory(unsigned long end_pfn); 165 static inline int mtrr_trim_uncached_memory(unsigned long end_pfn) argument
|
/arch/unicore32/kernel/ |
H A D | hibernate.c | 147 unsigned long end_pfn = PAGE_ALIGN(__pa(&__nosave_end)) >> PAGE_SHIFT; local 149 return (pfn >= begin_pfn) && (pfn < end_pfn);
|
/arch/unicore32/mm/ |
H A D | init.c | 132 unsigned long end_pfn) 143 boot_pages = bootmem_bootmap_pages(end_pfn - start_pfn); 145 __pfn_to_phys(end_pfn)); 153 init_bootmem_node(pgdat, __phys_to_pfn(bitmap), start_pfn, end_pfn); 160 if (end >= end_pfn) 161 end = end_pfn; 173 if (end >= end_pfn) 174 end = end_pfn; 335 free_memmap(unsigned long start_pfn, unsigned long end_pfn) argument 341 * Convert start_pfn/end_pfn t 131 uc32_bootmem_init(unsigned long start_pfn, unsigned long end_pfn) argument [all...] |
/arch/s390/kernel/ |
H A D | early.c | 93 unsigned int i, stext_pfn, eshared_pfn, end_pfn, min_size; local 126 end_pfn = PFN_UP(__pa(&_end)); 127 min_size = end_pfn << 2; 132 eshared_pfn - 1, eshared_pfn, end_pfn); 211 unsigned long end_pfn, init_pfn; local 213 end_pfn = PFN_UP(__pa(&_end)); 215 for (init_pfn = 0 ; init_pfn < end_pfn; init_pfn++)
|
/arch/mips/sgi-ip27/ |
H A D | ip27-memory.c | 396 pfn_t start_pfn, end_pfn; local 398 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); 407 NODE_DATA(node)->node_spanned_pages = end_pfn - start_pfn; 415 start_pfn, end_pfn); 416 free_bootmem_with_active_regions(node, end_pfn); 469 pfn_t start_pfn, end_pfn; local 471 get_pfn_range_for_nid(node, &start_pfn, &end_pfn); 473 if (end_pfn > max_low_pfn) 474 max_low_pfn = end_pfn;
|
/arch/sparc/mm/ |
H A D | init_32.c | 121 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; local 123 if (end_pfn <= max_low_pfn) 129 nr += end_pfn - start_pfn; 384 static void map_high_region(unsigned long start_pfn, unsigned long end_pfn) argument 389 printk("mapping high region %08lx - %08lx\n", start_pfn, end_pfn); 392 for (tmp = start_pfn; tmp < end_pfn; tmp++) { 444 unsigned long end_pfn = (sp_banks[i].base_addr + sp_banks[i].num_bytes) >> PAGE_SHIFT; local 448 if (end_pfn <= highstart_pfn) 454 map_high_region(start_pfn, end_pfn);
|
H A D | init_64.c | 823 unsigned long paddr, num_pages, start_pfn, end_pfn; local 840 get_pfn_range_for_nid(nid, &start_pfn, &end_pfn); 842 p->node_spanned_pages = end_pfn - start_pfn; 1338 unsigned long end_pfn; local 1340 end_pfn = p->node_start_pfn + p->node_spanned_pages; 1343 nid, paddr >> PAGE_SHIFT, p->node_start_pfn, end_pfn); 1346 p->node_start_pfn, end_pfn); 1349 nid, end_pfn); 1350 free_bootmem_with_active_regions(nid, end_pfn); 1362 unsigned long end_pfn; local 1702 unsigned long end_pfn, shift, phys_base; local [all...] |
/arch/mn10300/kernel/ |
H A D | setup.c | 123 unsigned long kstart_pfn, start_pfn, free_pfn, end_pfn; local 143 end_pfn = PFN_DOWN(__pa(memory_end)); 148 end_pfn); 155 PFN_PHYS(end_pfn - free_pfn));
|
/arch/arm/mm/ |
H A D | init.c | 152 unsigned long end_pfn) 163 boot_pages = bootmem_bootmap_pages(end_pfn - start_pfn); 165 __pfn_to_phys(end_pfn)); 173 init_bootmem_node(pgdat, __phys_to_pfn(bitmap), start_pfn, end_pfn); 180 if (end >= end_pfn) 181 end = end_pfn; 193 if (end >= end_pfn) 194 end = end_pfn; 442 free_memmap(unsigned long start_pfn, unsigned long end_pfn) argument 448 * Convert start_pfn/end_pfn t 151 arm_bootmem_init(unsigned long start_pfn, unsigned long end_pfn) argument [all...] |
/arch/x86/xen/ |
H A D | setup.c | 148 unsigned long end_pfn = PFN_UP(end); local 151 end_pfn = PFN_UP(entry->addr); 153 if (start_pfn < end_pfn) { 156 start_pfn, min(end_pfn, nr_pages)); 159 start_pfn, end_pfn);
|
/arch/x86/kernel/ |
H A D | e820.c | 766 unsigned long end_pfn; local 772 end_pfn = (ei->addr + ei->size) >> PAGE_SHIFT; 776 if (end_pfn > limit_pfn) { 780 if (end_pfn > last_pfn) 781 last_pfn = end_pfn; 1086 unsigned long start_pfn, end_pfn; local 1096 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { 1098 end_pfn = min_t(unsigned long, end_pfn, MAX_DMA_PFN); 1099 nr_pages += end_pfn [all...] |
/arch/alpha/mm/ |
H A D | numa.c | 308 unsigned long end_pfn = bdata->node_low_pfn; local 310 if (dma_local_pfn >= end_pfn - start_pfn) 311 zones_size[ZONE_DMA] = end_pfn - start_pfn; 314 zones_size[ZONE_NORMAL] = (end_pfn - start_pfn) - dma_local_pfn;
|
/arch/microblaze/mm/ |
H A D | init.c | 202 unsigned long start_pfn, end_pfn; local 205 end_pfn = memblock_region_memory_end_pfn(reg); 207 (end_pfn - start_pfn) << PAGE_SHIFT, 0);
|
/arch/parisc/mm/ |
H A D | init.c | 233 unsigned long end_pfn; local 237 end_pfn = pmem_ranges[0].start_pfn + pmem_ranges[0].pages; 240 hole_pages = pmem_ranges[i].start_pfn - end_pfn; 242 pmem_holes[npmem_holes].start_pfn = end_pfn; 244 end_pfn += hole_pages; 246 end_pfn += pmem_ranges[i].pages; 249 pmem_ranges[0].pages = end_pfn - pmem_ranges[0].start_pfn;
|