Searched refs:nr_pages (Results 1 - 25 of 40) sorted by relevance

12

/arch/x86/mm/
H A Dnuma_32.c67 unsigned long nr_pages = end_pfn - start_pfn; local
69 if (!nr_pages)
72 return (nr_pages + 1) * sizeof(struct page);
H A Dgup.c257 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
269 len = (unsigned long) nr_pages << PAGE_SHIFT;
312 * @nr_pages: number of pages from start to pin
315 * Should be at least nr_pages long.
322 * requested. If nr_pages is 0 or negative, returns 0. If no pages
325 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
336 len = (unsigned long) nr_pages << PAGE_SHIFT;
H A Dtlb.c122 unsigned long nr_pages = local
129 trace_tlb_flush(TLB_REMOTE_SHOOTDOWN, nr_pages);
H A Dinit_64.c708 unsigned long nr_pages = size >> PAGE_SHIFT; local
713 ret = __add_pages(nid, zone, start_pfn, nr_pages);
728 unsigned int nr_pages = 1 << order; local
736 while (nr_pages--)
739 while (nr_pages--)
1032 unsigned long nr_pages = size >> PAGE_SHIFT; local
1038 ret = __remove_pages(zone, start_pfn, nr_pages);
1379 unsigned int nr_pages; local
1419 nr_pages = 1 << (get_order(PMD_SIZE));
1421 while (nr_pages
[all...]
/arch/arm/mach-rpc/include/mach/
H A Duncompress.h23 unsigned long nr_pages; member in struct:param_struct
119 unsigned int nr_pages = 0, page_size = PAGE_SIZE; local
138 nr_pages += (t->u.mem.size / PAGE_SIZE);
144 nr_pages = params->nr_pages;
189 if (nr_pages * page_size < 4096*1024) error("<4M of mem\n");
/arch/arm/xen/
H A Dp2m.c23 unsigned long nr_pages; member in struct:xen_p2m_entry
72 entry->pfn + entry->nr_pages > pfn) {
120 unsigned long mfn, unsigned long nr_pages)
132 p2m_entry->pfn + p2m_entry->nr_pages > pfn) {
153 p2m_entry->nr_pages = nr_pages;
119 __set_phys_to_machine_multi(unsigned long pfn, unsigned long mfn, unsigned long nr_pages) argument
/arch/powerpc/kvm/
H A Dbook3s_hv_builtin.c134 struct page *kvm_alloc_hpt(unsigned long nr_pages) argument
138 VM_BUG_ON(order_base_2(nr_pages) < KVM_CMA_CHUNK_ORDER - PAGE_SHIFT);
142 align_pages = nr_pages;
143 return cma_alloc(kvm_cma, nr_pages, order_base_2(align_pages));
147 void kvm_release_hpt(struct page *page, unsigned long nr_pages) argument
149 cma_release(kvm_cma, page, nr_pages);
/arch/powerpc/mm/
H A Dgup.c133 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
143 pr_devel("%s(%lx,%x,%s)\n", __func__, start, nr_pages, write ? "write" : "read");
147 len = (unsigned long) nr_pages << PAGE_SHIFT;
201 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
208 nr = __get_user_pages_fast(start, nr_pages, write, pages);
211 if (nr < nr_pages) {
220 nr_pages - nr, write, 0, pages, NULL);
H A Dmem.c122 unsigned long nr_pages = size >> PAGE_SHIFT; local
134 return __add_pages(nid, zone, start_pfn, nr_pages);
141 unsigned long nr_pages = size >> PAGE_SHIFT; local
146 ret = __remove_pages(zone, start_pfn, nr_pages);
162 walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages, argument
166 unsigned long end_pfn = start_pfn + nr_pages;
/arch/s390/mm/
H A Dgup.c170 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
181 len = (unsigned long) nr_pages << PAGE_SHIFT;
211 * @nr_pages: number of pages from start to pin
214 * Should be at least nr_pages long.
221 * requested. If nr_pages is 0 or negative, returns 0. If no pages
224 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
231 nr = __get_user_pages_fast(start, nr_pages, write, pages);
232 if (nr == nr_pages)
240 nr_pages - nr, write, 0, pages, NULL);
H A Dinit.c175 unsigned long zone_start_pfn, zone_end_pfn, nr_pages; local
197 nr_pages = (start_pfn + size_pages > zone_end_pfn) ?
199 rc = __add_pages(nid, zone, start_pfn, nr_pages);
202 start_pfn += nr_pages;
203 size_pages -= nr_pages;
/arch/s390/pci/
H A Dpci_dma.c145 unsigned int nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; local
151 if (!nr_pages)
158 for (i = 0; i < nr_pages; i++) {
177 nr_pages * PAGE_SIZE);
281 unsigned long nr_pages, iommu_page_index; local
287 nr_pages = iommu_num_pages(pa, size, PAGE_SIZE);
288 iommu_page_index = dma_alloc_iommu(zdev, nr_pages);
293 size = nr_pages * PAGE_SIZE;
303 atomic64_add(nr_pages, &zdev->fmb->mapped_pages);
308 dma_free_iommu(zdev, iommu_page_index, nr_pages);
[all...]
/arch/arm/kernel/
H A Datags_compat.c46 unsigned long nr_pages; /* 4 */ member in struct:param_struct::__anon183::__anon184
108 if (params->u1.s.nr_pages != 0x02000 &&
109 params->u1.s.nr_pages != 0x04000 &&
110 params->u1.s.nr_pages != 0x08000 &&
111 params->u1.s.nr_pages != 0x10000) {
115 params->u1.s.nr_pages = 0x1000; /* 16MB */
163 tag = memtag(tag, PHYS_OFFSET, params->u1.s.nr_pages * PAGE_SIZE);
/arch/x86/xen/
H A Dsetup.c206 unsigned long end_pfn, unsigned long nr_pages, unsigned long *identity,
212 *released += xen_do_chunk(start_pfn, min(end_pfn, nr_pages), true);
398 unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn,
413 if (cur_pfn >= nr_pages) {
419 if (cur_pfn + size > nr_pages)
420 size = nr_pages - cur_pfn;
427 cur_pfn + left, nr_pages, identity, released);
438 cur_pfn + left, nr_pages, identity, released);
462 const struct e820entry *list, size_t map_size, unsigned long nr_pages,
468 unsigned long last_pfn = nr_pages;
205 xen_set_identity_and_release_chunk(unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long *identity, unsigned long *released) argument
396 xen_set_identity_and_remap_chunk( const struct e820entry *list, size_t map_size, unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, unsigned long remap_pfn, unsigned long *identity, unsigned long *remapped, unsigned long *released) argument
461 xen_set_identity_and_remap( const struct e820entry *list, size_t map_size, unsigned long nr_pages, unsigned long *released) argument
[all...]
/arch/mips/mm/
H A Dgup.c196 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
208 len = (unsigned long) nr_pages << PAGE_SHIFT;
250 * @nr_pages: number of pages from start to pin
253 * Should be at least nr_pages long.
260 * requested. If nr_pages is 0 or negative, returns 0. If no pages
263 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
274 len = (unsigned long) nr_pages << PAGE_SHIFT;
/arch/sh/mm/
H A Dgup.c163 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
175 len = (unsigned long) nr_pages << PAGE_SHIFT;
204 * @nr_pages: number of pages from start to pin
207 * Should be at least nr_pages long.
214 * requested. If nr_pages is 0 or negative, returns 0. If no pages
217 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
228 len = (unsigned long) nr_pages << PAGE_SHIFT;
H A Dinit.c492 unsigned long nr_pages = size >> PAGE_SHIFT; local
500 start_pfn, nr_pages);
521 unsigned long nr_pages = size >> PAGE_SHIFT; local
526 ret = __remove_pages(zone, start_pfn, nr_pages);
/arch/sparc/mm/
H A Dgup.c163 int __get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
174 len = (unsigned long) nr_pages << PAGE_SHIFT;
193 int get_user_pages_fast(unsigned long start, int nr_pages, int write, argument
204 len = (unsigned long) nr_pages << PAGE_SHIFT;
/arch/s390/kvm/
H A Dgaccess.c538 unsigned long *pages, unsigned long nr_pages,
551 while (nr_pages) {
576 nr_pages--;
585 unsigned long _len, nr_pages, gpa, idx; local
597 nr_pages = (((ga & ~PAGE_MASK) + len - 1) >> PAGE_SHIFT) + 1;
599 if (nr_pages > ARRAY_SIZE(pages_array))
600 pages = vmalloc(nr_pages * sizeof(unsigned long));
607 rc = guest_page_range(vcpu, ga, pages, nr_pages, write);
608 for (idx = 0; idx < nr_pages && !rc; idx++) {
621 if (nr_pages > ARRAY_SIZ
537 guest_page_range(struct kvm_vcpu *vcpu, unsigned long ga, unsigned long *pages, unsigned long nr_pages, int write) argument
[all...]
/arch/arm64/mm/
H A Ddma-mapping.c294 unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; local
301 if (off < nr_pages && nr_vma_pages <= (nr_pages - off)) {
368 unsigned long nr_pages = atomic_pool_size >> PAGE_SHIFT; local
374 page = dma_alloc_from_contiguous(NULL, nr_pages,
418 if (!dma_release_from_contiguous(NULL, page, nr_pages))
/arch/alpha/mm/
H A Dinit.c192 unsigned long nr_pages = 0; local
198 nr_pages += crb->map[i].count;
202 console_remap_vm.size = nr_pages << PAGE_SHIFT;
/arch/arm/include/asm/xen/
H A Dpage.h100 unsigned long nr_pages);
/arch/powerpc/platforms/pseries/
H A Dcmm.c476 unsigned long end = start + (marg->nr_pages << PAGE_SHIFT);
529 unsigned long end_page = start_page + (marg->nr_pages << PAGE_SHIFT);
535 start_page, marg->nr_pages);
/arch/ia64/mm/
H A Dinit.c660 unsigned long nr_pages = size >> PAGE_SHIFT; local
667 ret = __add_pages(nid, zone, start_pfn, nr_pages);
680 unsigned long nr_pages = size >> PAGE_SHIFT; local
685 ret = __remove_pages(zone, start_pfn, nr_pages);
/arch/tile/kernel/
H A Dmodule.c58 area->nr_pages = npages;

Completed in 2450 milliseconds

12