/arch/powerpc/include/asm/ |
H A D | uaccess.h | 50 #define __access_ok(addr, size, segment) \ 51 (((addr) <= (segment).seg) && ((size) <= (segment).seg)) 55 #define __access_ok(addr, size, segment) \ 57 (((size) == 0) || (((size) - 1) <= ((segment).seg - (addr))))) 61 #define access_ok(type, addr, size) \ 63 __access_ok((__force unsigned long)(addr), (size), get_fs())) 85 * use the right size if we just have the right pointer type. 164 #define __put_user_size(x, ptr, size, retval) \ 167 switch (size) { \ 419 __copy_from_user(void *to, const void __user *from, unsigned long size) argument 426 __copy_to_user(void __user *to, const void *from, unsigned long size) argument 435 clear_user(void __user *addr, unsigned long size) argument [all...] |
/arch/arm/xen/ |
H A D | mm32.c | 76 size_t size, enum dma_data_direction dir, 80 size_t left = size; 129 size_t size, enum dma_data_direction dir) 135 outer_inv_range(handle, handle + size); 137 dma_cache_maint(handle & PAGE_MASK, handle & ~PAGE_MASK, size, dir, dmac_unmap_area); 141 size_t size, enum dma_data_direction dir) 144 dma_cache_maint(handle & PAGE_MASK, handle & ~PAGE_MASK, size, dir, dmac_map_area); 147 outer_inv_range(handle, handle + size); 149 outer_clean_range(handle, handle + size); 154 size_t size, enu 75 dma_cache_maint(dma_addr_t handle, unsigned long offset, size_t size, enum dma_data_direction dir, void (*op)(const void *, size_t, int)) argument 128 __xen_dma_page_dev_to_cpu(struct device *hwdev, dma_addr_t handle, size_t size, enum dma_data_direction dir) argument 140 __xen_dma_page_cpu_to_dev(struct device *hwdev, dma_addr_t handle, size_t size, enum dma_data_direction dir) argument 153 xen_dma_unmap_page(struct device *hwdev, dma_addr_t handle, size_t size, enum dma_data_direction dir, struct dma_attrs *attrs) argument 166 xen_dma_sync_single_for_cpu(struct device *hwdev, dma_addr_t handle, size_t size, enum dma_data_direction dir) argument 174 xen_dma_sync_single_for_device(struct device *hwdev, dma_addr_t handle, size_t size, enum dma_data_direction dir) argument [all...] |
/arch/cris/include/asm/ |
H A D | dma-mapping.h | 19 void *dma_alloc_coherent(struct device *dev, size_t size, 22 void dma_free_coherent(struct device *dev, size_t size, 26 dma_alloc_coherent(struct device *dev, size_t size, dma_addr_t *dma_handle, argument 34 dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, argument 41 dma_map_single(struct device *dev, void *ptr, size_t size, argument 49 dma_unmap_single(struct device *dev, dma_addr_t dma_addr, size_t size, argument 65 size_t size, enum dma_data_direction direction) 72 dma_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, argument 87 dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size, argument 93 dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size, argument 64 dma_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction direction) argument 99 dma_sync_single_range_for_cpu(struct device *dev, dma_addr_t dma_handle, unsigned long offset, size_t size, enum dma_data_direction direction) argument 106 dma_sync_single_range_for_device(struct device *dev, dma_addr_t dma_handle, unsigned long offset, size_t size, enum dma_data_direction direction) argument 156 dma_cache_sync(struct device *dev, void *vaddr, size_t size, enum dma_data_direction direction) argument [all...] |
/arch/metag/include/asm/ |
H A D | cacheflush.h | 128 static inline void flush_cache_sigtramp(unsigned long addr, int size) argument 140 metag_code_cache_flush((void *) addr, size); 157 static inline void cachew_region_op(void *start, unsigned long size, argument 163 size += offset; 166 i = (size - 1) >> 6; 198 static inline void flush_dcache_region(void *start, unsigned long size) argument 200 /* metag_data_cache_flush won't flush L2 cache lines if size >= 4096 */ 202 cachew_region_op(start, size, CACHEW_FLUSH_L1D_L2); 204 l2c_fence_flush(start + size - 1); 206 metag_data_cache_flush(start, size); 211 writeback_dcache_region(void *start, unsigned long size) argument 220 invalidate_dcache_region(void *start, unsigned long size) argument [all...] |
/arch/microblaze/kernel/ |
H A D | dma.c | 18 static void *dma_direct_alloc_coherent(struct device *dev, size_t size, argument 23 return consistent_alloc(flag, size, dma_handle); 32 page = alloc_pages_node(node, flag, get_order(size)); 36 memset(ret, 0, size); 43 static void dma_direct_free_coherent(struct device *dev, size_t size, argument 48 consistent_free(size, vaddr); 50 free_pages((unsigned long)vaddr, get_order(size)); 79 size_t size, 83 __dma_sync(page_to_phys(page) + offset, size, direction); 89 size_t size, 76 dma_direct_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction direction, struct dma_attrs *attrs) argument 87 dma_direct_unmap_page(struct device *dev, dma_addr_t dma_address, size_t size, enum dma_data_direction direction, struct dma_attrs *attrs) argument 102 dma_direct_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument 116 dma_direct_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument [all...] |
/arch/mips/sgi-ip22/ |
H A D | ip22-mc.c | 45 unsigned long size; member in struct:mem 63 bank[cnt].size = get_bank_size(tmp); 66 i, bank[cnt].size / 1024 / 1024, bank[cnt].addr); 72 unsigned long addr, size; local 78 size = bank[i].size; 80 bank[i].size = bank[i-1].size; 82 bank[i-1].size = size; [all...] |
/arch/openrisc/kernel/ |
H A D | dma.c | 84 or1k_dma_alloc(struct device *dev, size_t size, argument 95 page = alloc_pages_exact(size, gfp); 109 if (walk_page_range(va, va + size, &walk)) { 110 free_pages_exact(page, size); 119 or1k_dma_free(struct device *dev, size_t size, void *vaddr, argument 130 WARN_ON(walk_page_range(va, va + size, &walk)); 133 free_pages_exact(vaddr, size); 138 unsigned long offset, size_t size, 148 for (cl = addr; cl < addr + size; 154 for (cl = addr; cl < addr + size; 137 or1k_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction dir, struct dma_attrs *attrs) argument 171 or1k_unmap_page(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction dir, struct dma_attrs *attrs) argument 208 or1k_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction dir) argument 221 or1k_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction dir) argument [all...] |
/arch/powerpc/sysdev/ |
H A D | fsl_85xx_cache_sram.c | 38 void *mpc85xx_cache_sram_alloc(unsigned int size, argument 47 if (!size || (size > cache_sram->size) || (align > cache_sram->size)) { 48 pr_err("%s(): size(=%x) or align(=%x) zero or too big\n", 49 __func__, size, align); 60 offset = rh_alloc_align(cache_sram->rh, size, align, NULL); 100 cache_sram->size = sram_params.sram_size; 102 if (!request_mem_region(cache_sram->base_phys, cache_sram->size, [all...] |
/arch/arm/plat-orion/ |
H A D | pcie.c | 126 u32 size; local 151 * Setup windows for DDR banks. Count total DDR size on the fly. 153 size = 0; 159 writel(((cs->size - 1) & 0xffff0000) | 164 size += cs->size; 168 * Round up 'size' to the nearest power of two. 170 if ((size & (size - 1)) != 0) 171 size 208 orion_pcie_rd_conf(void __iomem *base, struct pci_bus *bus, u32 devfn, int where, int size, u32 *val) argument 227 orion_pcie_rd_conf_tlp(void __iomem *base, struct pci_bus *bus, u32 devfn, int where, int size, u32 *val) argument 250 orion_pcie_rd_conf_wa(void __iomem *wa_base, struct pci_bus *bus, u32 devfn, int where, int size, u32 *val) argument 266 orion_pcie_wr_conf(void __iomem *base, struct pci_bus *bus, u32 devfn, int where, int size, u32 val) argument [all...] |
/arch/x86/boot/compressed/ |
H A D | aslr.c | 38 size_t size) 43 for (i = 0; i < size / sizeof(hash); i++) { 109 unsigned long size; member in struct:mem_vector 121 if (item->start + item->size > region->start + region->size) 129 if (one->start + one->size <= two->start) 132 if (one->start >= two->start + two->size) 152 mem_avoid[0].size = unsafe_len; 160 mem_avoid[1].size = initrd_size; 165 /* Calculate size o 37 rotate_xor(unsigned long hash, const void *area, size_t size) argument 281 find_random_addr(unsigned long minimum, unsigned long size) argument [all...] |
/arch/xtensa/include/asm/ |
H A D | uaccess.h | 79 * kernel_ok determines whether we should bypass addr/size checking. 120 * <as> register containing memory size 127 * <at> destroyed (actually, (TASK_SIZE + 1 - size)) 149 * <as> register containing memory size 188 #define __user_ok(addr,size) \ 189 (((size) <= TASK_SIZE)&&((addr) <= TASK_SIZE-(size))) 190 #define __access_ok(addr,size) (__kernel_ok || __user_ok((addr),(size))) 191 #define access_ok(type,addr,size) __access_o [all...] |
/arch/parisc/kernel/ |
H A D | pci-dma.c | 84 unsigned long size, unsigned long *paddr_ptr) 90 end = vaddr + size; 111 unsigned long size, unsigned long *paddr_ptr) 117 end = vaddr + size; 133 static inline int map_uncached_pages(unsigned long vaddr, unsigned long size, argument 137 unsigned long end = vaddr + size; 155 unsigned long size) 170 end = vaddr + size; 191 unsigned long size) 206 end = vaddr + size; 82 map_pte_uncached(pte_t * pte, unsigned long vaddr, unsigned long size, unsigned long *paddr_ptr) argument 110 map_pmd_uncached(pmd_t * pmd, unsigned long vaddr, unsigned long size, unsigned long *paddr_ptr) argument 154 unmap_uncached_pte(pmd_t * pmd, unsigned long vaddr, unsigned long size) argument 190 unmap_uncached_pmd(pgd_t * dir, unsigned long vaddr, unsigned long size) argument 217 unmap_uncached_pages(unsigned long vaddr, unsigned long size) argument 250 pcxl_alloc_range(size_t size) argument 306 pcxl_free_range(unsigned long vaddr, size_t size) argument 416 pa11_dma_alloc_consistent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t flag) argument 442 pa11_dma_free_consistent(struct device *dev, size_t size, void *vaddr, dma_addr_t dma_handle) argument 453 pa11_dma_map_single(struct device *dev, void *addr, size_t size, enum dma_data_direction direction) argument 461 pa11_dma_unmap_single(struct device *dev, dma_addr_t dma_handle, size_t size, enum dma_data_direction direction) argument 509 pa11_dma_sync_single_for_cpu(struct device *dev, dma_addr_t dma_handle, unsigned long offset, size_t size, enum dma_data_direction direction) argument 516 pa11_dma_sync_single_for_device(struct device *dev, dma_addr_t dma_handle, unsigned long offset, size_t size, enum dma_data_direction direction) argument 558 fail_alloc_consistent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t flag) argument 564 pa11_dma_alloc_noncoherent(struct device *dev, size_t size, dma_addr_t *dma_handle, gfp_t flag) argument 576 pa11_dma_free_noncoherent(struct device *dev, size_t size, void *vaddr, dma_addr_t iova) argument [all...] |
/arch/x86/xen/ |
H A D | setup.c | 54 * The maximum amount of extra memory compared to the base size. The 55 * main scaling factor is the size of struct page. At extreme ratios 65 static void __init xen_add_extra_mem(u64 start, u64 size) argument 72 if (xen_extra_mem[i].size == 0) { 74 xen_extra_mem[i].size = size; 78 if (xen_extra_mem[i].start + xen_extra_mem[i].size == start) { 79 xen_extra_mem[i].size += size; 86 memblock_reserve(start, size); 254 xen_do_set_identity_and_remap_chunk( unsigned long start_pfn, unsigned long size, unsigned long remap_pfn) argument 409 unsigned long size = left; local 536 xen_align_and_add_e820_region(u64 start, u64 size, int type) argument 646 u64 size = map[i].size; local [all...] |
/arch/arm/mach-shmobile/ |
H A D | setup-rcar-gen2.c | 119 u64 base, size; member in struct:memory_reserve_config 143 u64 base, size; local 146 size = dt_mem_next_cell(dt_root_size_cells, ®); 151 if ((base + size) >= lpae_start) 152 size = lpae_start - base; 154 if (size < mrc->reserved) 161 mrc->base = base + size - mrc->reserved; 162 mrc->size = mrc->reserved; 180 if (mrc.size) 181 dma_contiguous_reserve_area(mrc.size, mr [all...] |
/arch/c6x/mm/ |
H A D | dma-coherent.c | 76 void *dma_alloc_coherent(struct device *dev, size_t size, argument 82 if (!dma_size || !size) 85 order = get_count_order(((size - 1) >> PAGE_SHIFT) + 1); 102 void dma_free_coherent(struct device *dev, size_t size, void *vaddr, argument 107 if (!dma_size || !size) 110 order = get_count_order(((size - 1) >> PAGE_SHIFT) + 1); 119 void __init coherent_mem_init(phys_addr_t start, u32 size) argument 123 if (!size) 127 "Coherent memory (DMA) region start=0x%x size=0x%x\n", 128 start, size); [all...] |
/arch/mips/pci/ |
H A D | ops-nile4.c | 77 int where, int size, u32 *val) 83 if ((size == 2) && (where & 1)) 85 else if ((size == 4) && (where & 3)) 96 if (size == 1) 98 else if (size == 2) 107 int where, int size, u32 val) 113 if ((size == 2) && (where & 1)) 115 else if ((size == 4) && (where & 3)) 126 if (size == 1) 129 else if (size 76 nile4_pcibios_read(struct pci_bus *bus, unsigned int devfn, int where, int size, u32 *val) argument 106 nile4_pcibios_write(struct pci_bus *bus, unsigned int devfn, int where, int size, u32 val) argument [all...] |
/arch/openrisc/include/asm/ |
H A D | dma-mapping.h | 43 static inline void *dma_alloc_attrs(struct device *dev, size_t size, argument 50 memory = ops->alloc(dev, size, dma_handle, gfp, attrs); 52 debug_dma_alloc_coherent(dev, size, *dma_handle, memory); 59 static inline void dma_free_attrs(struct device *dev, size_t size, argument 65 debug_dma_free_coherent(dev, size, cpu_addr, dma_handle); 67 ops->free(dev, size, cpu_addr, dma_handle, attrs); 70 static inline void *dma_alloc_noncoherent(struct device *dev, size_t size, argument 77 return dma_alloc_attrs(dev, size, dma_handle, gfp, &attrs); 80 static inline void dma_free_noncoherent(struct device *dev, size_t size, argument 87 dma_free_attrs(dev, size, cpu_add [all...] |
/arch/x86/kernel/ |
H A D | check.c | 24 u64 size; member in struct:scan_area 59 unsigned size; local 61 size = memparse(arg, &end); 64 corruption_check_size = size; 66 return (size == corruption_check_size) ? 0 : -EINVAL; 104 scan_areas[num_scan_areas].size = end - start; 128 unsigned long size = scan_areas[i].size; local 130 for (; size; addr++, size [all...] |
/arch/powerpc/boot/dts/ |
H A D | adder875-redboot.dts | 18 #size-cells = <1>; 28 #size-cells = <0>; 33 d-cache-line-size = <16>; 34 i-cache-line-size = <16>; 35 d-cache-size = <8192>; 36 i-cache-size = <8192>; 54 #size-cells = <1>; 73 #size-cells = <1>; 86 #size-cells = <0>; 130 #size [all...] |
H A D | adder875-uboot.dts | 18 #size-cells = <1>; 28 #size-cells = <0>; 33 d-cache-line-size = <16>; 34 i-cache-line-size = <16>; 35 d-cache-size = <8192>; 36 i-cache-size = <8192>; 54 #size-cells = <1>; 72 #size-cells = <1>; 85 #size-cells = <0>; 129 #size [all...] |
/arch/mips/mm/ |
H A D | ioremap.c | 20 phys_t size, phys_t phys_addr, unsigned long flags) 28 end = address + size; 46 phys_t size, phys_t phys_addr, unsigned long flags) 51 end = address + size; 68 phys_t size, unsigned long flags) 72 unsigned long end = address + size; 116 void __iomem * __ioremap(phys_t phys_addr, phys_t size, unsigned long flags) argument 123 phys_addr = fixup_bigphys_addr(phys_addr, size); 125 /* Don't allow wraparound or zero size */ 126 last_addr = phys_addr + size 19 remap_area_pte(pte_t * pte, unsigned long address, phys_t size, phys_t phys_addr, unsigned long flags) argument 45 remap_area_pmd(pmd_t * pmd, unsigned long address, phys_t size, phys_t phys_addr, unsigned long flags) argument 67 remap_area_pages(unsigned long address, phys_t phys_addr, phys_t size, unsigned long flags) argument [all...] |
/arch/s390/mm/ |
H A D | maccess.c | 25 static long probe_kernel_write_odd(void *dst, const void *src, size_t size) argument 33 count = min_t(unsigned long, 4 - offset, size); 51 long probe_kernel_write(void *dst, const void *src, size_t size) argument 55 while (size) { 56 copied = probe_kernel_write_odd(dst, src, size); 61 size -= copied; 133 int offs = 0, size, rc; local 141 size = min(PAGE_SIZE, count - offs); 142 if (memcpy_real(buf, src + offs, size)) 144 if (copy_to_user(dest + offs, buf, size)) 182 unsigned long size; local [all...] |
/arch/x86/kernel/cpu/mtrr/ |
H A D | cyrix.c | 14 unsigned long *size, mtrr_type * type) 41 *size = (reg < 7 ? 0x1UL : 0x40UL) << (shift - 1); 43 *size = 0; 85 * @size: the size (in bytes) of the region. 90 cyrix_get_free_region(unsigned long base, unsigned long size, int replace_reg) argument 98 if (size < 0x40) 111 if (size > 0x2000) { 124 * try ARR7 but its size must be at least 256K 127 if ((lsize == 0) && (size > 13 cyrix_get_arr(unsigned int reg, unsigned long *base, unsigned long *size, mtrr_type * type) argument 177 cyrix_set_arr(unsigned int reg, unsigned long base, unsigned long size, mtrr_type type) argument 237 unsigned long size; member in struct:__anon3191 [all...] |
/arch/frv/mb93090-mb00/ |
H A D | pci-dma.c | 21 void *dma_alloc_coherent(struct device *hwdev, size_t size, dma_addr_t *dma_handle, gfp_t gfp) argument 25 ret = consistent_alloc(gfp, size, dma_handle); 27 memset(ret, 0, size); 34 void dma_free_coherent(struct device *hwdev, size_t size, void *vaddr, dma_addr_t dma_handle) argument 41 dma_addr_t dma_map_single(struct device *dev, void *ptr, size_t size, argument 46 frv_cache_wback_inv((unsigned long) ptr, (unsigned long) ptr + size); 84 size_t size, enum dma_data_direction direction) 83 dma_map_page(struct device *dev, struct page *page, unsigned long offset, size_t size, enum dma_data_direction direction) argument
|
/arch/mips/bcm63xx/ |
H A D | dev-pcmcia.c | 70 u32 base, unsigned int size) 76 ret = bcm63xx_set_cs_base(cs, base, size); 85 unsigned int size; member in struct:__anon1949 90 .size = BCM_PCMCIA_COMMON_SIZE 95 .size = BCM_PCMCIA_ATTR_SIZE 100 .size = BCM_PCMCIA_IO_SIZE 134 pcmcia_cs[i].size); 69 config_pcmcia_cs(unsigned int cs, u32 base, unsigned int size) argument
|