/arch/sparc/mm/ |
H A D | tlb.c | 59 struct address_space *mapping; local 70 mapping = page_mapping(page); 71 if (!mapping)
|
H A D | init_64.c | 336 struct address_space *mapping; local 351 mapping = page_mapping(page); 352 if (mapping && !mapping_mapped(mapping)) { 1217 * a 1-to-1 mapping from CPU ID to NUMA node ID.
|
/arch/powerpc/include/asm/ |
H A D | abs_addr.h | 26 u32 *mapping; member in struct:mschunks_map
|
/arch/sh/mm/ |
H A D | cache-sh7705.c | 139 struct address_space *mapping = page_mapping(page); local 141 if (mapping && !mapping_mapped(mapping))
|
H A D | cache-sh4.c | 115 struct address_space *mapping = page_mapping(page); local 117 if (mapping && !mapping_mapped(mapping))
|
/arch/unicore32/mm/ |
H A D | flush.c | 61 void __flush_dcache_page(struct address_space *mapping, struct page *page) argument 64 * Writeback any data associated with the kernel mapping of this 66 * coherent with the kernels mapping. 72 * Ensure cache coherency between kernel mapping and userspace mapping 77 struct address_space *mapping; local 86 mapping = page_mapping(page); 88 if (mapping && !mapping_mapped(mapping)) 91 __flush_dcache_page(mapping, pag [all...] |
H A D | mmu.c | 181 * Try a section mapping - end, addr and phys must all be aligned 204 * page tables for the mapping specified by `md'. We 215 printk(KERN_WARNING "BUG: not creating mapping for " 223 printk(KERN_WARNING "BUG: mapping for 0x%08llx at 0x%08lx " 364 * Create a mapping for the machine vectors at the high-vectors 366 * create a mapping at the low-vectors virtual address. 375 * Create a mapping for the kuser page at the special 444 * In order to soft-boot, we need to insert a 1:1 mapping in place of 492 struct address_space *mapping; local 506 mapping [all...] |
/arch/xtensa/mm/ |
H A D | cache.c | 72 struct address_space *mapping = page_mapping(page); local 75 * If we have a mapping but the page is not mapped to user-space 80 if (mapping && !mapping_mapped(mapping)) { 96 * if we have a mapping. 99 if (!alias && !mapping) 109 if (mapping)
|
/arch/arm/mm/ |
H A D | fault-armv.c | 132 make_coherent(struct address_space *mapping, struct vm_area_struct *vma, argument 149 flush_dcache_mmap_lock(mapping); 150 vma_prio_tree_foreach(mpnt, &iter, &mapping->i_mmap, pgoff, pgoff) { 163 flush_dcache_mmap_unlock(mapping); 185 struct address_space *mapping; local 199 mapping = page_mapping(page); 201 __flush_dcache_page(mapping, page); 202 if (mapping) { 204 make_coherent(mapping, vma, addr, ptep, pfn);
|
H A D | flush.c | 163 void __flush_dcache_page(struct address_space *mapping, struct page *page) argument 166 * Writeback any data associated with the kernel mapping of this 168 * coherent with the kernels mapping. 190 if (mapping && cache_is_vipt_aliasing()) 195 static void __flush_dcache_aliases(struct address_space *mapping, struct page *page) argument 206 * - aliasing VIPT: we only need to find one mapping of this page. 210 flush_dcache_mmap_lock(mapping); 211 vma_prio_tree_foreach(mpnt, &iter, &mapping->i_mmap, pgoff, pgoff) { 224 flush_dcache_mmap_unlock(mapping); 232 struct address_space *mapping; local 278 struct address_space *mapping; local [all...] |
/arch/mips/mm/ |
H A D | cache.c | 82 struct address_space *mapping = page_mapping(page); local 87 if (mapping && !mapping_mapped(mapping)) {
|
/arch/score/mm/ |
H A D | cache.c | 57 struct address_space *mapping = page_mapping(page); local 62 if (mapping && !mapping_mapped(mapping)) {
|
/arch/parisc/kernel/ |
H A D | sys_parisc.c | 56 * existing mapping and use the same offset. New scheme is to use the 60 * The mapping is cacheline aligned, so there's no information in the bottom 64 static int get_offset(struct address_space *mapping) argument 66 int offset = (unsigned long) mapping << (PAGE_SHIFT - 8); 70 static unsigned long get_shared_area(struct address_space *mapping, argument 74 int offset = mapping ? get_offset(mapping) : 0;
|
H A D | cache.c | 277 struct address_space *mapping = page_mapping(page); local 284 if (mapping && !mapping_mapped(mapping)) { 291 if (!mapping) 301 flush_dcache_mmap_lock(mapping); 302 vma_prio_tree_foreach(mpnt, &iter, &mapping->i_mmap, pgoff, pgoff) { 308 * mapping, so here we kill the mapping then flush the 309 * page along a special flush only alias mapping. 323 flush_dcache_mmap_unlock(mapping); [all...] |
/arch/x86/mm/ |
H A D | hugetlbpage.c | 64 struct address_space *mapping = vma->vm_file->f_mapping; local 75 mutex_lock(&mapping->i_mmap_mutex); 76 vma_prio_tree_foreach(svma, &iter, &mapping->i_mmap, idx, idx) { 100 mutex_unlock(&mapping->i_mmap_mutex); 106 * Hugetlb pte page is ref counted at the time of mapping. If pte is shared 364 * mapping then try again:
|
/arch/xtensa/variants/s6000/ |
H A D | gpio.c | 186 const signed char *mapping = platform_irq_mappings[irq]; local 189 if (!mapping) 191 for(mask = 0; *mapping != -1; mapping++) 192 switch (*mapping) {
|
/arch/c6x/platforms/ |
H A D | megamod-pic.c | 55 /* hw mux mapping */ 165 * Parse the MUX mapping, if one exists. 178 int *mapping, int size) 194 mapping[i] = val; 204 int mapping[NR_MUX_OUTPUTS]; local 232 for (i = 0; i < ARRAY_SIZE(mapping); i++) 233 mapping[i] = IRQ_UNMAPPED; 235 parse_priority_map(pic, mapping, ARRAY_SIZE(mapping)); 262 /* record the mapping */ 177 parse_priority_map(struct megamod_pic *pic, int *mapping, int size) argument [all...] |
/arch/powerpc/kernel/ |
H A D | iommu.c | 8 * Dynamic DMA mapping support, bus-independent parts. 32 #include <linux/dma-mapping.h> 286 DBG("sg mapping %d elements:\n", nelems); 624 * to the dma address (mapping) of the first page. 631 dma_addr_t mapping; local 663 mapping = iommu_alloc(dev, tbl, ret, nio_pages, DMA_BIDIRECTIONAL, 665 if (mapping == DMA_ERROR_CODE) { 669 *dma_handle = mapping;
|
/arch/x86/include/asm/ |
H A D | x86_init.h | 138 struct x86_init_mapping mapping; member in struct:x86_init_ops
|
/arch/arm/kernel/ |
H A D | perf_event.c | 112 int mapping = (*event_map)[config]; local 113 return mapping == HW_OP_UNSUPPORTED ? -ENOENT : mapping; 306 /* Propagate our changes to the userspace mapping. */ 477 int mapping, err; local 479 mapping = armpmu->map_event(event); 481 if (mapping < 0) { 484 return mapping; 512 hwc->config_base |= (unsigned long)mapping;
|
/arch/x86/kernel/ |
H A D | pci-calgary_64.c | 33 #include <linux/dma-mapping.h> 436 dma_addr_t mapping; local 453 mapping = iommu_alloc(dev, tbl, ret, npages, DMA_BIDIRECTIONAL); 454 if (mapping == DMA_ERROR_CODE) 456 *dma_handle = mapping;
|
/arch/x86/xen/ |
H A D | mmu.c | 5 * The most important job they must perform is the mapping between the 27 * will disallow any pagetable update which will end up mapping a 96 /* l3 pud for userspace vsyscall mapping */ 474 * _PAGE_PSE, so avoid fiddling with the PAT mapping and hope 1593 * Like __va(), but returns address in the kernel mapping (which is 1594 * all we have until the physical memory mapping has been set up. 1685 struct xen_machphys_mapping mapping; local 1687 if (HYPERVISOR_memory_op(XENMEM_machphys_mapping, &mapping) == 0) { 1688 machine_to_phys_mapping = (unsigned long *)mapping.v_start; 1689 machine_to_phys_nr = mapping [all...] |
/arch/mips/include/asm/mach-rc32434/ |
H A D | pci.h | 38 u32 mapping; /* mapping. */ member in struct:pci_map
|
/arch/ia64/include/asm/ |
H A D | pal.h | 73 #define PAL_LOGICAL_TO_PHYSICAL 42 /* returns information on logical to physical processor mapping */ 1781 ia64_pal_logical_to_phys(u64 proc_number, pal_logical_to_physical_t *mapping) argument 1789 mapping->overview.overview_data = iprv.v0; 1790 mapping->ppli1.ppli1_data = iprv.v1; 1791 mapping->ppli2.ppli2_data = iprv.v2;
|