Searched defs:mapping (Results 1 - 24 of 24) sorted by relevance

/arch/sparc/mm/
H A Dtlb.c59 struct address_space *mapping; local
70 mapping = page_mapping(page);
71 if (!mapping)
H A Dinit_64.c336 struct address_space *mapping; local
351 mapping = page_mapping(page);
352 if (mapping && !mapping_mapped(mapping)) {
1217 * a 1-to-1 mapping from CPU ID to NUMA node ID.
/arch/powerpc/include/asm/
H A Dabs_addr.h26 u32 *mapping; member in struct:mschunks_map
/arch/sh/mm/
H A Dcache-sh7705.c139 struct address_space *mapping = page_mapping(page); local
141 if (mapping && !mapping_mapped(mapping))
H A Dcache-sh4.c115 struct address_space *mapping = page_mapping(page); local
117 if (mapping && !mapping_mapped(mapping))
/arch/unicore32/mm/
H A Dflush.c61 void __flush_dcache_page(struct address_space *mapping, struct page *page) argument
64 * Writeback any data associated with the kernel mapping of this
66 * coherent with the kernels mapping.
72 * Ensure cache coherency between kernel mapping and userspace mapping
77 struct address_space *mapping; local
86 mapping = page_mapping(page);
88 if (mapping && !mapping_mapped(mapping))
91 __flush_dcache_page(mapping, pag
[all...]
H A Dmmu.c181 * Try a section mapping - end, addr and phys must all be aligned
204 * page tables for the mapping specified by `md'. We
215 printk(KERN_WARNING "BUG: not creating mapping for "
223 printk(KERN_WARNING "BUG: mapping for 0x%08llx at 0x%08lx "
364 * Create a mapping for the machine vectors at the high-vectors
366 * create a mapping at the low-vectors virtual address.
375 * Create a mapping for the kuser page at the special
444 * In order to soft-boot, we need to insert a 1:1 mapping in place of
492 struct address_space *mapping; local
506 mapping
[all...]
/arch/xtensa/mm/
H A Dcache.c72 struct address_space *mapping = page_mapping(page); local
75 * If we have a mapping but the page is not mapped to user-space
80 if (mapping && !mapping_mapped(mapping)) {
96 * if we have a mapping.
99 if (!alias && !mapping)
109 if (mapping)
/arch/arm/mm/
H A Dfault-armv.c132 make_coherent(struct address_space *mapping, struct vm_area_struct *vma, argument
149 flush_dcache_mmap_lock(mapping);
150 vma_prio_tree_foreach(mpnt, &iter, &mapping->i_mmap, pgoff, pgoff) {
163 flush_dcache_mmap_unlock(mapping);
185 struct address_space *mapping; local
199 mapping = page_mapping(page);
201 __flush_dcache_page(mapping, page);
202 if (mapping) {
204 make_coherent(mapping, vma, addr, ptep, pfn);
H A Dflush.c163 void __flush_dcache_page(struct address_space *mapping, struct page *page) argument
166 * Writeback any data associated with the kernel mapping of this
168 * coherent with the kernels mapping.
190 if (mapping && cache_is_vipt_aliasing())
195 static void __flush_dcache_aliases(struct address_space *mapping, struct page *page) argument
206 * - aliasing VIPT: we only need to find one mapping of this page.
210 flush_dcache_mmap_lock(mapping);
211 vma_prio_tree_foreach(mpnt, &iter, &mapping->i_mmap, pgoff, pgoff) {
224 flush_dcache_mmap_unlock(mapping);
232 struct address_space *mapping; local
278 struct address_space *mapping; local
[all...]
/arch/mips/mm/
H A Dcache.c82 struct address_space *mapping = page_mapping(page); local
87 if (mapping && !mapping_mapped(mapping)) {
/arch/score/mm/
H A Dcache.c57 struct address_space *mapping = page_mapping(page); local
62 if (mapping && !mapping_mapped(mapping)) {
/arch/parisc/kernel/
H A Dsys_parisc.c56 * existing mapping and use the same offset. New scheme is to use the
60 * The mapping is cacheline aligned, so there's no information in the bottom
64 static int get_offset(struct address_space *mapping) argument
66 int offset = (unsigned long) mapping << (PAGE_SHIFT - 8);
70 static unsigned long get_shared_area(struct address_space *mapping, argument
74 int offset = mapping ? get_offset(mapping) : 0;
H A Dcache.c277 struct address_space *mapping = page_mapping(page); local
284 if (mapping && !mapping_mapped(mapping)) {
291 if (!mapping)
301 flush_dcache_mmap_lock(mapping);
302 vma_prio_tree_foreach(mpnt, &iter, &mapping->i_mmap, pgoff, pgoff) {
308 * mapping, so here we kill the mapping then flush the
309 * page along a special flush only alias mapping.
323 flush_dcache_mmap_unlock(mapping);
[all...]
/arch/x86/mm/
H A Dhugetlbpage.c64 struct address_space *mapping = vma->vm_file->f_mapping; local
75 mutex_lock(&mapping->i_mmap_mutex);
76 vma_prio_tree_foreach(svma, &iter, &mapping->i_mmap, idx, idx) {
100 mutex_unlock(&mapping->i_mmap_mutex);
106 * Hugetlb pte page is ref counted at the time of mapping. If pte is shared
364 * mapping then try again:
/arch/xtensa/variants/s6000/
H A Dgpio.c186 const signed char *mapping = platform_irq_mappings[irq]; local
189 if (!mapping)
191 for(mask = 0; *mapping != -1; mapping++)
192 switch (*mapping) {
/arch/c6x/platforms/
H A Dmegamod-pic.c55 /* hw mux mapping */
165 * Parse the MUX mapping, if one exists.
178 int *mapping, int size)
194 mapping[i] = val;
204 int mapping[NR_MUX_OUTPUTS]; local
232 for (i = 0; i < ARRAY_SIZE(mapping); i++)
233 mapping[i] = IRQ_UNMAPPED;
235 parse_priority_map(pic, mapping, ARRAY_SIZE(mapping));
262 /* record the mapping */
177 parse_priority_map(struct megamod_pic *pic, int *mapping, int size) argument
[all...]
/arch/powerpc/kernel/
H A Diommu.c8 * Dynamic DMA mapping support, bus-independent parts.
32 #include <linux/dma-mapping.h>
286 DBG("sg mapping %d elements:\n", nelems);
624 * to the dma address (mapping) of the first page.
631 dma_addr_t mapping; local
663 mapping = iommu_alloc(dev, tbl, ret, nio_pages, DMA_BIDIRECTIONAL,
665 if (mapping == DMA_ERROR_CODE) {
669 *dma_handle = mapping;
/arch/x86/include/asm/
H A Dx86_init.h138 struct x86_init_mapping mapping; member in struct:x86_init_ops
/arch/arm/kernel/
H A Dperf_event.c112 int mapping = (*event_map)[config]; local
113 return mapping == HW_OP_UNSUPPORTED ? -ENOENT : mapping;
306 /* Propagate our changes to the userspace mapping. */
477 int mapping, err; local
479 mapping = armpmu->map_event(event);
481 if (mapping < 0) {
484 return mapping;
512 hwc->config_base |= (unsigned long)mapping;
/arch/x86/kernel/
H A Dpci-calgary_64.c33 #include <linux/dma-mapping.h>
436 dma_addr_t mapping; local
453 mapping = iommu_alloc(dev, tbl, ret, npages, DMA_BIDIRECTIONAL);
454 if (mapping == DMA_ERROR_CODE)
456 *dma_handle = mapping;
/arch/x86/xen/
H A Dmmu.c5 * The most important job they must perform is the mapping between the
27 * will disallow any pagetable update which will end up mapping a
96 /* l3 pud for userspace vsyscall mapping */
474 * _PAGE_PSE, so avoid fiddling with the PAT mapping and hope
1593 * Like __va(), but returns address in the kernel mapping (which is
1594 * all we have until the physical memory mapping has been set up.
1685 struct xen_machphys_mapping mapping; local
1687 if (HYPERVISOR_memory_op(XENMEM_machphys_mapping, &mapping) == 0) {
1688 machine_to_phys_mapping = (unsigned long *)mapping.v_start;
1689 machine_to_phys_nr = mapping
[all...]
/arch/mips/include/asm/mach-rc32434/
H A Dpci.h38 u32 mapping; /* mapping. */ member in struct:pci_map
/arch/ia64/include/asm/
H A Dpal.h73 #define PAL_LOGICAL_TO_PHYSICAL 42 /* returns information on logical to physical processor mapping */
1781 ia64_pal_logical_to_phys(u64 proc_number, pal_logical_to_physical_t *mapping) argument
1789 mapping->overview.overview_data = iprv.v0;
1790 mapping->ppli1.ppli1_data = iprv.v1;
1791 mapping->ppli2.ppli2_data = iprv.v2;

Completed in 519 milliseconds