/arch/x86/mm/ |
H A D | numa_32.c | 67 unsigned long nr_pages = end_pfn - start_pfn; local 69 if (!nr_pages) 72 return (nr_pages + 1) * sizeof(struct page);
|
H A D | gup.c | 257 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 269 len = (unsigned long) nr_pages << PAGE_SHIFT; 312 * @nr_pages: number of pages from start to pin 315 * Should be at least nr_pages long. 322 * requested. If nr_pages is 0 or negative, returns 0. If no pages 325 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 336 len = (unsigned long) nr_pages << PAGE_SHIFT;
|
H A D | tlb.c | 122 unsigned long nr_pages = local 129 trace_tlb_flush(TLB_REMOTE_SHOOTDOWN, nr_pages);
|
H A D | init_64.c | 708 unsigned long nr_pages = size >> PAGE_SHIFT; local 713 ret = __add_pages(nid, zone, start_pfn, nr_pages); 728 unsigned int nr_pages = 1 << order; local 736 while (nr_pages--) 739 while (nr_pages--) 1032 unsigned long nr_pages = size >> PAGE_SHIFT; local 1038 ret = __remove_pages(zone, start_pfn, nr_pages); 1379 unsigned int nr_pages; local 1419 nr_pages = 1 << (get_order(PMD_SIZE)); 1421 while (nr_pages [all...] |
/arch/arm/mach-rpc/include/mach/ |
H A D | uncompress.h | 23 unsigned long nr_pages; member in struct:param_struct 119 unsigned int nr_pages = 0, page_size = PAGE_SIZE; local 138 nr_pages += (t->u.mem.size / PAGE_SIZE); 144 nr_pages = params->nr_pages; 189 if (nr_pages * page_size < 4096*1024) error("<4M of mem\n");
|
/arch/arm/xen/ |
H A D | p2m.c | 23 unsigned long nr_pages; member in struct:xen_p2m_entry 72 entry->pfn + entry->nr_pages > pfn) { 120 unsigned long mfn, unsigned long nr_pages) 132 p2m_entry->pfn + p2m_entry->nr_pages > pfn) { 153 p2m_entry->nr_pages = nr_pages; 119 __set_phys_to_machine_multi(unsigned long pfn, unsigned long mfn, unsigned long nr_pages) argument
|
/arch/powerpc/kvm/ |
H A D | book3s_hv_builtin.c | 134 struct page *kvm_alloc_hpt(unsigned long nr_pages) argument 138 VM_BUG_ON(order_base_2(nr_pages) < KVM_CMA_CHUNK_ORDER - PAGE_SHIFT); 142 align_pages = nr_pages; 143 return cma_alloc(kvm_cma, nr_pages, order_base_2(align_pages)); 147 void kvm_release_hpt(struct page *page, unsigned long nr_pages) argument 149 cma_release(kvm_cma, page, nr_pages);
|
/arch/powerpc/mm/ |
H A D | gup.c | 133 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 143 pr_devel("%s(%lx,%x,%s)\n", __func__, start, nr_pages, write ? "write" : "read"); 147 len = (unsigned long) nr_pages << PAGE_SHIFT; 201 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 208 nr = __get_user_pages_fast(start, nr_pages, write, pages); 211 if (nr < nr_pages) { 220 nr_pages - nr, write, 0, pages, NULL);
|
H A D | mem.c | 122 unsigned long nr_pages = size >> PAGE_SHIFT; local 134 return __add_pages(nid, zone, start_pfn, nr_pages); 141 unsigned long nr_pages = size >> PAGE_SHIFT; local 146 ret = __remove_pages(zone, start_pfn, nr_pages); 162 walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages, argument 166 unsigned long end_pfn = start_pfn + nr_pages;
|
/arch/s390/mm/ |
H A D | gup.c | 170 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 181 len = (unsigned long) nr_pages << PAGE_SHIFT; 211 * @nr_pages: number of pages from start to pin 214 * Should be at least nr_pages long. 221 * requested. If nr_pages is 0 or negative, returns 0. If no pages 224 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 231 nr = __get_user_pages_fast(start, nr_pages, write, pages); 232 if (nr == nr_pages) 240 nr_pages - nr, write, 0, pages, NULL);
|
H A D | init.c | 175 unsigned long zone_start_pfn, zone_end_pfn, nr_pages; local 197 nr_pages = (start_pfn + size_pages > zone_end_pfn) ? 199 rc = __add_pages(nid, zone, start_pfn, nr_pages); 202 start_pfn += nr_pages; 203 size_pages -= nr_pages;
|
/arch/s390/pci/ |
H A D | pci_dma.c | 145 unsigned int nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; local 151 if (!nr_pages) 158 for (i = 0; i < nr_pages; i++) { 177 nr_pages * PAGE_SIZE); 281 unsigned long nr_pages, iommu_page_index; local 287 nr_pages = iommu_num_pages(pa, size, PAGE_SIZE); 288 iommu_page_index = dma_alloc_iommu(zdev, nr_pages); 293 size = nr_pages * PAGE_SIZE; 303 atomic64_add(nr_pages, &zdev->fmb->mapped_pages); 308 dma_free_iommu(zdev, iommu_page_index, nr_pages); [all...] |
/arch/arm/kernel/ |
H A D | atags_compat.c | 46 unsigned long nr_pages; /* 4 */ member in struct:param_struct::__anon183::__anon184 108 if (params->u1.s.nr_pages != 0x02000 && 109 params->u1.s.nr_pages != 0x04000 && 110 params->u1.s.nr_pages != 0x08000 && 111 params->u1.s.nr_pages != 0x10000) { 115 params->u1.s.nr_pages = 0x1000; /* 16MB */ 163 tag = memtag(tag, PHYS_OFFSET, params->u1.s.nr_pages * PAGE_SIZE);
|
/arch/x86/xen/ |
H A D | setup.c | 206 unsigned long end_pfn, unsigned long nr_pages, unsigned long *identity, 212 *released += xen_do_chunk(start_pfn, min(end_pfn, nr_pages), true); 398 unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn, 413 if (cur_pfn >= nr_pages) { 419 if (cur_pfn + size > nr_pages) 420 size = nr_pages - cur_pfn; 427 cur_pfn + left, nr_pages, identity, released); 438 cur_pfn + left, nr_pages, identity, released); 462 const struct e820entry *list, size_t map_size, unsigned long nr_pages, 468 unsigned long last_pfn = nr_pages; 205 xen_set_identity_and_release_chunk(unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long *identity, unsigned long *released) argument 396 xen_set_identity_and_remap_chunk( const struct e820entry *list, size_t map_size, unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn, unsigned long *identity, unsigned long *remapped, unsigned long *released) argument 461 xen_set_identity_and_remap( const struct e820entry *list, size_t map_size, unsigned long nr_pages, unsigned long *released) argument [all...] |
/arch/mips/mm/ |
H A D | gup.c | 196 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 208 len = (unsigned long) nr_pages << PAGE_SHIFT; 250 * @nr_pages: number of pages from start to pin 253 * Should be at least nr_pages long. 260 * requested. If nr_pages is 0 or negative, returns 0. If no pages 263 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 274 len = (unsigned long) nr_pages << PAGE_SHIFT;
|
/arch/sh/mm/ |
H A D | gup.c | 163 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 175 len = (unsigned long) nr_pages << PAGE_SHIFT; 204 * @nr_pages: number of pages from start to pin 207 * Should be at least nr_pages long. 214 * requested. If nr_pages is 0 or negative, returns 0. If no pages 217 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 228 len = (unsigned long) nr_pages << PAGE_SHIFT;
|
H A D | init.c | 492 unsigned long nr_pages = size >> PAGE_SHIFT; local 500 start_pfn, nr_pages); 521 unsigned long nr_pages = size >> PAGE_SHIFT; local 526 ret = __remove_pages(zone, start_pfn, nr_pages);
|
/arch/sparc/mm/ |
H A D | gup.c | 163 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 174 len = (unsigned long) nr_pages << PAGE_SHIFT; 193 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument 204 len = (unsigned long) nr_pages << PAGE_SHIFT;
|
/arch/s390/kvm/ |
H A D | gaccess.c | 538 unsigned long *pages, unsigned long nr_pages, 551 while (nr_pages) { 576 nr_pages--; 585 unsigned long _len, nr_pages, gpa, idx; local 597 nr_pages = (((ga & ~PAGE_MASK) + len - 1) >> PAGE_SHIFT) + 1; 599 if (nr_pages > ARRAY_SIZE(pages_array)) 600 pages = vmalloc(nr_pages * sizeof(unsigned long)); 607 rc = guest_page_range(vcpu, ga, pages, nr_pages, write); 608 for (idx = 0; idx < nr_pages && !rc; idx++) { 621 if (nr_pages > ARRAY_SIZ 537 guest_page_range(struct kvm_vcpu *vcpu, unsigned long ga, unsigned long *pages, unsigned long nr_pages, int write) argument [all...] |
/arch/arm64/mm/ |
H A D | dma-mapping.c | 294 unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; local 301 if (off < nr_pages && nr_vma_pages <= (nr_pages - off)) { 368 unsigned long nr_pages = atomic_pool_size >> PAGE_SHIFT; local 374 page = dma_alloc_from_contiguous(NULL, nr_pages, 418 if (!dma_release_from_contiguous(NULL, page, nr_pages))
|
/arch/alpha/mm/ |
H A D | init.c | 192 unsigned long nr_pages = 0; local 198 nr_pages += crb->map[i].count; 202 console_remap_vm.size = nr_pages << PAGE_SHIFT;
|
/arch/arm/include/asm/xen/ |
H A D | page.h | 100 unsigned long nr_pages);
|
/arch/powerpc/platforms/pseries/ |
H A D | cmm.c | 476 unsigned long end = start + (marg->nr_pages << PAGE_SHIFT); 529 unsigned long end_page = start_page + (marg->nr_pages << PAGE_SHIFT); 535 start_page, marg->nr_pages);
|
/arch/ia64/mm/ |
H A D | init.c | 660 unsigned long nr_pages = size >> PAGE_SHIFT; local 667 ret = __add_pages(nid, zone, start_pfn, nr_pages); 680 unsigned long nr_pages = size >> PAGE_SHIFT; local 685 ret = __remove_pages(zone, start_pfn, nr_pages);
|
/arch/tile/kernel/ |
H A D | module.c | 58 area->nr_pages = npages;
|