/arch/s390/mm/ |
H A D | page-states.c | 95 struct zone *zone; local 101 for_each_populated_zone(zone) { 102 spin_lock_irqsave(&zone->lock, flags); 104 list_for_each(l, &zone->free_area[order].free_list[t]) { 112 spin_unlock_irqrestore(&zone->lock, flags);
|
H A D | init.c | 178 struct zone *zone; local 184 for_each_zone(zone) { 185 if (zone_idx(zone) != ZONE_MOVABLE) { 186 /* Add range within existing zone limits */ 187 zone_start_pfn = zone->zone_start_pfn; 188 zone_end_pfn = zone->zone_start_pfn + 189 zone->spanned_pages; 199 rc = __add_pages(nid, zone, start_pfn, nr_pages);
|
/arch/x86/mm/ |
H A D | highmem_32.c | 122 struct zone *zone; local 126 * Explicitly reset zone->managed_pages because set_highmem_pages_init() 130 for_each_zone(zone) { 133 if (!is_highmem(zone)) 136 zone_start_pfn = zone->zone_start_pfn; 137 zone_end_pfn = zone_start_pfn + zone->spanned_pages; 139 nid = zone_to_nid(zone); 141 zone->name, nid, zone_start_pfn, zone_end_pfn);
|
H A D | init_32.c | 828 struct zone *zone = pgdata->node_zones + local 833 return __add_pages(nid, zone, start_pfn, nr_pages); 841 struct zone *zone; local 843 zone = page_zone(pfn_to_page(start_pfn)); 844 return __remove_pages(zone, start_pfn, nr_pages);
|
H A D | init_64.c | 699 * Memory is added always to NORMAL zone. This means you will never get 705 struct zone *zone = pgdat->node_zones + local 713 ret = __add_pages(nid, zone, start_pfn, nr_pages); 1033 struct zone *zone; local 1036 zone = page_zone(pfn_to_page(start_pfn)); 1038 ret = __remove_pages(zone, start_pfn, nr_pages);
|
/arch/m68k/mm/ |
H A D | mcfmmu.c | 42 enum zone_type zone; local 78 for (zone = 0; zone < MAX_NR_ZONES; zone++) 79 zones_size[zone] = 0x0;
|
/arch/ia64/mm/ |
H A D | init.c | 479 unsigned long zone; member in struct:memmap_init_callback_data 508 args->nid, args->zone, page_to_pfn(map_start), 514 memmap_init (unsigned long size, int nid, unsigned long zone, argument 518 memmap_init_zone(size, nid, zone, start_pfn, MEMMAP_EARLY); 527 args.zone = zone; 658 struct zone *zone; local 665 zone = pgdat->node_zones + 667 ret = __add_pages(nid, zone, start_pf 681 struct zone *zone; local 747 struct zone *zone = &pgdat->node_zones[zoneid]; local [all...] |
/arch/powerpc/mm/ |
H A D | mem.c | 120 struct zone *zone; local 131 zone = pgdata->node_zones + 134 return __add_pages(nid, zone, start_pfn, nr_pages); 142 struct zone *zone; local 145 zone = page_zone(pfn_to_page(start_pfn)); 146 ret = __remove_pages(zone, start_pfn, nr_pages); 277 * Restrict the specified zone and all more restrictive zones 281 void __init limit_zone_pfn(enum zone_type zone, unsigne argument [all...] |
/arch/tile/mm/ |
H A D | pgtable.c | 45 struct zone *zone; local 66 for_each_zone(zone) { 69 if (!populated_zone(zone)) 72 spin_lock_irqsave(&zone->lock, flags); 74 int nr = zone->free_area[order].nr_free; 79 spin_unlock_irqrestore(&zone->lock, flags); 81 zone_to_nid(zone), zone->name,
|
H A D | init.c | 700 struct zone *z; 875 struct zone *zone = pgdata->node_zones + MAX_NR_ZONES-1; local 879 return __add_pages(zone, start_pfn, nr_pages);
|
/arch/powerpc/kernel/ |
H A D | dma.c | 58 int zone; 66 zone = dma_pfn_limit_to_zone(pfn); 67 if (zone < 0) { 68 dev_err(dev, "%s: No suitable zone for pfn %#llx\n", 73 switch (zone) {
|
/arch/tile/include/asm/ |
H A D | homecache.h | 26 struct zone;
|
/arch/parisc/mm/ |
H A D | init.c | 660 struct zone *zone = &pgdat->node_zones[zoneid]; local 661 if (!populated_zone(zone)) 664 total += zone->present_pages; 665 reserved = zone->present_pages - zone->managed_pages; 682 struct zone *zone; local 684 printk("Zone list for zone %d on node %d: ", j, i); 685 for_each_zone_zonelist(zone, [all...] |
/arch/sh/mm/ |
H A D | init.c | 522 struct zone *zone; local 525 zone = page_zone(pfn_to_page(start_pfn)); 526 ret = __remove_pages(zone, start_pfn, nr_pages);
|
/arch/c6x/ |
H A D | Kconfig | 88 int "Maximum zone order" 92 blocks into "zones", where each zone is a power of two number of
|
/arch/powerpc/include/asm/ |
H A D | pgtable.h | 252 void limit_zone_pfn(enum zone_type zone, unsigned long max_pfn);
|
/arch/x86/boot/compressed/ |
H A D | Makefile | 36 $(obj)/eboot.o: KBUILD_CFLAGS += -fshort-wchar -mno-red-zone
|
/arch/x86/ |
H A D | Makefile | 98 KBUILD_CFLAGS += -mno-red-zone
|
/arch/ia64/include/asm/ |
H A D | pgtable.h | 562 extern void memmap_init (unsigned long size, int nid, unsigned long zone,
|
/arch/sh/kernel/cpu/sh5/ |
H A D | entry.S | 684 bge SP, r6, tr0 ! ? below 1k from bottom of stack : danger zone
|