/arch/sh/kernel/ |
H A D | dumpstack.c | 89 struct thread_info *context; local 92 context = (struct thread_info *) 102 context, &graph);
|
/arch/sparc/kernel/ |
H A D | smp_64.c | 740 * Otherwise if we send an xcall from interrupt context it will 965 /* See if we need to allocate a new TLB context because 972 spin_lock_irqsave(&mm->context.lock, flags); 974 if (unlikely(!CTX_VALID(mm->context))) 977 spin_unlock_irqrestore(&mm->context.lock, flags); 980 __flush_tlb_mm(CTX_HWBITS(mm->context), 1024 * 2) TLB context numbers are shared globally across all processors 1030 * current cpu's bit set, that tlb context is flushed locally. 1037 * this task to flush the context from the local tlb if the process 1051 * bits of the context registe 1109 unsigned long context = CTX_HWBITS(mm->context); local [all...] |
/arch/sparc/mm/ |
H A D | srmmu.c | 74 #define FLUSH_BEGIN(mm) if ((mm)->context != NO_CONTEXT) { 99 /* The context table is a nocache user with the biggest alignment needs. */ 371 /* context handling - a dynamically sized pool is used */ 412 mm->context = ctxp->ctx_number; 425 ctxp->ctx_mm->context = NO_CONTEXT; 427 mm->context = ctxp->ctx_number; 430 static inline void free_context(int context) argument 434 ctx_old = ctx_list_pool + context; 463 if (mm->context == NO_CONTEXT) { 467 srmmu_ctxd_set(&srmmu_context_table[mm->context], m [all...] |
/arch/tile/gxio/ |
H A D | iorpc_mpipe.c | 24 int gxio_mpipe_alloc_buffer_stacks(gxio_mpipe_context_t *context, argument 35 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 48 int gxio_mpipe_init_buffer_stack_aux(gxio_mpipe_context_t *context, argument 69 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 83 int gxio_mpipe_alloc_notif_rings(gxio_mpipe_context_t *context, argument 94 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 105 int gxio_mpipe_init_notif_ring_aux(gxio_mpipe_context_t *context, void *mem_va, argument 124 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 136 int gxio_mpipe_request_notif_ring_interrupt(gxio_mpipe_context_t *context, argument 150 return hv_dev_pwrite(context 161 gxio_mpipe_enable_notif_ring_interrupt(gxio_mpipe_context_t *context, unsigned int ring) argument 182 gxio_mpipe_alloc_notif_groups(gxio_mpipe_context_t *context, unsigned int count, unsigned int first, unsigned int flags) argument 204 gxio_mpipe_init_notif_group(gxio_mpipe_context_t *context, unsigned int group, gxio_mpipe_notif_group_bits_t bits) argument 226 gxio_mpipe_alloc_buckets(gxio_mpipe_context_t *context, unsigned int count, unsigned int first, unsigned int flags) argument 247 gxio_mpipe_init_bucket(gxio_mpipe_context_t *context, unsigned int bucket, MPIPE_LBL_INIT_DAT_BSTS_TBL_t bucket_info) argument 268 gxio_mpipe_alloc_edma_rings(gxio_mpipe_context_t *context, unsigned int count, unsigned int first, unsigned int flags) argument 291 gxio_mpipe_init_edma_ring_aux(gxio_mpipe_context_t *context, void *mem_va, size_t mem_size, unsigned int mem_flags, unsigned int ring, unsigned int channel) argument 318 gxio_mpipe_commit_rules(gxio_mpipe_context_t *context, const void *blob, size_t blob_size) argument 335 gxio_mpipe_register_client_memory(gxio_mpipe_context_t *context, unsigned int iotlb, HV_PTE pte, unsigned int flags) argument 358 gxio_mpipe_link_open_aux(gxio_mpipe_context_t *context, _gxio_mpipe_link_name_t name, unsigned int flags) argument 377 gxio_mpipe_link_close_aux(gxio_mpipe_context_t *context, int mac) argument 396 gxio_mpipe_link_set_attr_aux(gxio_mpipe_context_t *context, int mac, uint32_t attr, int64_t val) argument 418 gxio_mpipe_get_timestamp_aux(gxio_mpipe_context_t *context, uint64_t *sec, uint64_t *nsec, uint64_t *cycles) argument 443 gxio_mpipe_set_timestamp_aux(gxio_mpipe_context_t *context, uint64_t sec, uint64_t nsec, uint64_t cycles) argument 463 gxio_mpipe_adjust_timestamp_aux(gxio_mpipe_context_t *context, int64_t nsec) argument 484 gxio_mpipe_config_edma_ring_blks(gxio_mpipe_context_t *context, unsigned int ering, unsigned int max_blks, unsigned int min_snf_blks, unsigned int db) argument 507 gxio_mpipe_adjust_timestamp_freq(gxio_mpipe_context_t *context, int32_t ppb) argument 525 gxio_mpipe_arm_pollfd(gxio_mpipe_context_t *context, int pollfd_cookie) argument 542 gxio_mpipe_close_pollfd(gxio_mpipe_context_t *context, int pollfd_cookie) argument 559 gxio_mpipe_get_mmio_base(gxio_mpipe_context_t *context, HV_PTE *base) argument 580 gxio_mpipe_check_mmio_offset(gxio_mpipe_context_t *context, unsigned long offset, unsigned long size) argument [all...] |
H A D | iorpc_mpipe_info.c | 22 int gxio_mpipe_info_instance_aux(gxio_mpipe_info_context_t *context, argument 30 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 41 int gxio_mpipe_info_enumerate_aux(gxio_mpipe_info_context_t *context, argument 51 hv_dev_pread(context->fd, 0, (HV_VirtAddr) params, sizeof(*params), 66 int gxio_mpipe_info_get_mmio_base(gxio_mpipe_info_context_t *context, argument 74 hv_dev_pread(context->fd, 0, (HV_VirtAddr) params, sizeof(*params), 88 int gxio_mpipe_info_check_mmio_offset(gxio_mpipe_info_context_t *context, argument 97 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params,
|
H A D | iorpc_trio.c | 24 int gxio_trio_alloc_asids(gxio_trio_context_t *context, unsigned int count, argument 34 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 47 int gxio_trio_alloc_memory_maps(gxio_trio_context_t *context, argument 58 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 70 int gxio_trio_alloc_scatter_queues(gxio_trio_context_t *context, argument 81 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 94 int gxio_trio_alloc_pio_regions(gxio_trio_context_t *context, argument 105 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 118 int gxio_trio_init_pio_region_aux(gxio_trio_context_t *context, argument 130 return hv_dev_pwrite(context 148 gxio_trio_init_memory_map_mmu_aux(gxio_trio_context_t *context, unsigned int map, unsigned long va, uint64_t size, unsigned int asid, unsigned int mac, uint64_t bus_address, unsigned int node, unsigned int order_mode) argument 178 gxio_trio_get_port_property(gxio_trio_context_t *context, struct pcie_trio_ports_property *trio_ports) argument 201 gxio_trio_config_legacy_intr(gxio_trio_context_t *context, int inter_x, int inter_y, int inter_ipi, int inter_event, unsigned int mac, unsigned int intx) argument 230 gxio_trio_config_msi_intr(gxio_trio_context_t *context, int inter_x, int inter_y, int inter_ipi, int inter_event, unsigned int mac, unsigned int mem_map, uint64_t mem_map_base, uint64_t mem_map_limit, unsigned int asid) argument 262 gxio_trio_set_mps_mrs(gxio_trio_context_t *context, uint16_t mps, uint16_t mrs, unsigned int mac) argument 282 gxio_trio_force_rc_link_up(gxio_trio_context_t *context, unsigned int mac) argument 299 gxio_trio_force_ep_link_up(gxio_trio_context_t *context, unsigned int mac) argument 316 gxio_trio_get_mmio_base(gxio_trio_context_t *context, HV_PTE *base) argument 337 gxio_trio_check_mmio_offset(gxio_trio_context_t *context, unsigned long offset, unsigned long size) argument [all...] |
H A D | iorpc_uart.c | 22 int gxio_uart_cfg_interrupt(gxio_uart_context_t *context, int inter_x, argument 33 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 43 int gxio_uart_get_mmio_base(gxio_uart_context_t *context, HV_PTE *base) argument 50 hv_dev_pread(context->fd, 0, (HV_VirtAddr) params, sizeof(*params), 64 int gxio_uart_check_mmio_offset(gxio_uart_context_t *context, argument 73 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params,
|
H A D | iorpc_usb_host.c | 22 int gxio_usb_host_cfg_interrupt(gxio_usb_host_context_t *context, int inter_x, argument 33 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 44 int gxio_usb_host_register_client_memory(gxio_usb_host_context_t *context, argument 53 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params, 64 int gxio_usb_host_get_mmio_base(gxio_usb_host_context_t *context, HV_PTE *base) argument 71 hv_dev_pread(context->fd, 0, (HV_VirtAddr) params, sizeof(*params), 85 int gxio_usb_host_check_mmio_offset(gxio_usb_host_context_t *context, argument 94 return hv_dev_pwrite(context->fd, 0, (HV_VirtAddr) params,
|
H A D | mpipe.c | 58 int gxio_mpipe_init(gxio_mpipe_context_t *context, unsigned int mpipe_index) argument 71 context->fd = fd; 81 context->mmio_cfg_base = (void __force *) 84 if (context->mmio_cfg_base == NULL) 87 context->mmio_fast_base = (void __force *) 90 if (context->mmio_fast_base == NULL) 95 context->__stacks.stacks[i] = 255; 97 context->instance = mpipe_index; 102 iounmap((void __force __iomem *)(context->mmio_cfg_base)); 104 hv_dev_close(context 111 gxio_mpipe_destroy(gxio_mpipe_context_t *context) argument 160 gxio_mpipe_init_buffer_stack(gxio_mpipe_context_t *context, unsigned int stack, gxio_mpipe_buffer_size_enum_t buffer_size_enum, void *mem, size_t mem_size, unsigned int mem_flags) argument 184 gxio_mpipe_init_notif_ring(gxio_mpipe_context_t *context, unsigned int ring, void *mem, size_t mem_size, unsigned int mem_flags) argument 195 gxio_mpipe_init_notif_group_and_buckets(gxio_mpipe_context_t *context, unsigned int group, unsigned int ring, unsigned int num_rings, unsigned int bucket, unsigned int num_buckets, gxio_mpipe_bucket_mode_t mode) argument 235 gxio_mpipe_init_edma_ring(gxio_mpipe_context_t *context, unsigned int ring, unsigned int channel, void *mem, size_t mem_size, unsigned int mem_flags) argument 248 gxio_mpipe_rules_init(gxio_mpipe_rules_t *rules, gxio_mpipe_context_t *context) argument 389 gxio_mpipe_iqueue_init(gxio_mpipe_iqueue_t *iqueue, gxio_mpipe_context_t *context, unsigned int ring, void *mem, size_t mem_size, unsigned int mem_flags) argument 417 gxio_mpipe_equeue_init(gxio_mpipe_equeue_t *equeue, gxio_mpipe_context_t *context, unsigned int ering, unsigned int channel, void *mem, unsigned int mem_size, unsigned int mem_flags) argument 458 gxio_mpipe_set_timestamp(gxio_mpipe_context_t *context, const struct timespec *ts) argument 467 gxio_mpipe_get_timestamp(gxio_mpipe_context_t *context, struct timespec *ts) argument 489 gxio_mpipe_adjust_timestamp(gxio_mpipe_context_t *context, int64_t delta) argument 499 static gxio_mpipe_context_t context; local 535 gxio_mpipe_context_t *context = _gxio_get_link_context(); local 552 gxio_mpipe_context_t *context = _gxio_get_link_context(); local 568 gxio_mpipe_link_open(gxio_mpipe_link_t *link, gxio_mpipe_context_t *context, const char *link_name, unsigned int flags) argument [all...] |
H A D | trio.c | 28 int gxio_trio_init(gxio_trio_context_t *context, unsigned int trio_index) argument 36 context->fd = -1; 44 context->fd = fd;
|
H A D | uart.c | 28 int gxio_uart_init(gxio_uart_context_t *context, int uart_index) argument 42 context->fd = fd; 45 context->mmio_base = (void __force *) 48 if (context->mmio_base == NULL) { 49 hv_dev_close(context->fd); 50 context->fd = -1; 59 int gxio_uart_destroy(gxio_uart_context_t *context) argument 61 iounmap((void __force __iomem *)(context->mmio_base)); 62 hv_dev_close(context->fd); 64 context 73 gxio_uart_write(gxio_uart_context_t *context, uint64_t offset, uint64_t word) argument 82 gxio_uart_read(gxio_uart_context_t *context, uint64_t offset) argument [all...] |
H A D | usb_host.c | 29 int gxio_usb_host_init(gxio_usb_host_context_t *context, int usb_index, argument 50 context->fd = fd; 53 context->mmio_base = 56 if (context->mmio_base == NULL) { 57 hv_dev_close(context->fd); 66 int gxio_usb_host_destroy(gxio_usb_host_context_t *context) argument 68 iounmap((void __force __iomem *)(context->mmio_base)); 69 hv_dev_close(context->fd); 71 context->mmio_base = NULL; 72 context 79 gxio_usb_host_get_reg_start(gxio_usb_host_context_t *context) argument 86 gxio_usb_host_get_reg_len(gxio_usb_host_context_t *context) argument [all...] |
/arch/tile/include/gxio/ |
H A D | mpipe.h | 90 * mPIPE context's set of open links; all VLANs; and all dMACs. 315 /* A context object used to manage mPIPE hardware resources. */ 338 /* Initialize an mPIPE context. 343 * @param context Context object to be initialized. 345 * context. 347 extern int gxio_mpipe_init(gxio_mpipe_context_t *context, 350 /* Destroy an mPIPE context. 356 * will destroy the mPIPE context as part of process teardown. 358 * @param context Context object to be destroyed. 360 extern int gxio_mpipe_destroy(gxio_mpipe_context_t *context); 457 gxio_mpipe_push_buffer(gxio_mpipe_context_t *context, unsigned int stack, void *buffer) argument 487 gxio_mpipe_pop_buffer(gxio_mpipe_context_t *context, unsigned int stack) argument 776 gxio_mpipe_credit(gxio_mpipe_context_t *context, int ring, int bucket, unsigned int count) argument 941 gxio_mpipe_context_t *context; member in struct:__anon2901 1119 gxio_mpipe_context_t *context; member in struct:__anon2902 1329 gxio_mpipe_context_t *context; member in struct:__anon2903 1720 gxio_mpipe_context_t *context; member in struct:__anon2904 [all...] |
/arch/tile/kernel/ |
H A D | pci_gx.c | 150 gxio_trio_context_t *context = &trio_contexts[trio_index]; local 155 ret = gxio_trio_init(context, trio_index); 160 ret = gxio_trio_alloc_asids(context, 1, 0, 0); 167 context->asid = ret; 175 ret = gxio_trio_alloc_pio_regions(context, 1, 0, 0); 182 context->pio_cfg_index = ret; 188 ret = gxio_trio_init_pio_region_aux(context, context->pio_cfg_index, 198 ret = gxio_trio_get_port_property(context, &pcie_ports[trio_index]); 205 context 349 gxio_trio_context_t *context = controller->trio; local 460 gxio_trio_context_t *context = &trio_contexts[i]; local [all...] |
/arch/um/drivers/ |
H A D | mconsole.h | 52 enum mc_context context; member in struct:mconsole_command
|
/arch/x86/include/asm/ |
H A D | init.h | 6 void *context; /* context for alloc_pgt_page */ member in struct:x86_mapping_info
|
/arch/x86/kernel/cpu/mcheck/ |
H A D | mce-inject.c | 113 int context = MCJ_CTX(m->inject_flags); local 119 switch (context) { 131 printk(KERN_INFO "Invalid MCE context\n"); 148 int context = MCJ_CTX(m->inject_flags); local 152 if (context == MCJ_CTX_RANDOM)
|
H A D | mce-severity.c | 32 enum context { IN_KERNEL = 1, IN_USER = 2 }; enum 42 unsigned char context; member in struct:severity 47 #define KERNEL .context = IN_KERNEL 48 #define USER .context = IN_USER 68 PANIC, "Processor context corrupt", 183 enum context ctx = error_context(m); 195 if (s->context && ctx != s->context)
|
/arch/x86/kernel/ |
H A D | dumpstack_32.c | 64 struct thread_info *context; local 71 context = task_thread_info(task); 72 bp = ops->walk_stack(context, stack, bp, ops, data,
|
H A D | ptrace.c | 185 unsigned long context = (unsigned long)regs & ~(THREAD_SIZE - 1); local 189 if (context == (sp & ~(THREAD_SIZE - 1))) 192 prev_esp = (u32 *)(context);
|
/arch/x86/kvm/ |
H A D | mmu.c | 3452 struct kvm_mmu *context) 3454 context->page_fault = nonpaging_page_fault; 3455 context->gva_to_gpa = nonpaging_gva_to_gpa; 3456 context->sync_page = nonpaging_sync_page; 3457 context->invlpg = nonpaging_invlpg; 3458 context->update_pte = nonpaging_update_pte; 3459 context->root_level = 0; 3460 context->shadow_root_level = PT32E_ROOT_LEVEL; 3461 context->root_hpa = INVALID_PAGE; 3462 context 3451 nonpaging_init_context(struct kvm_vcpu *vcpu, struct kvm_mmu *context) argument 3521 reset_rsvds_bits_mask(struct kvm_vcpu *vcpu, struct kvm_mmu *context) argument 3596 reset_rsvds_bits_mask_ept(struct kvm_vcpu *vcpu, struct kvm_mmu *context, bool execonly) argument 3711 paging64_init_context_common(struct kvm_vcpu *vcpu, struct kvm_mmu *context, int level) argument 3733 paging64_init_context(struct kvm_vcpu *vcpu, struct kvm_mmu *context) argument 3739 paging32_init_context(struct kvm_vcpu *vcpu, struct kvm_mmu *context) argument 3759 paging32E_init_context(struct kvm_vcpu *vcpu, struct kvm_mmu *context) argument 3767 struct kvm_mmu *context = vcpu->arch.walk_mmu; local 3807 kvm_init_shadow_mmu(struct kvm_vcpu *vcpu, struct kvm_mmu *context) argument 3830 kvm_init_shadow_ept_mmu(struct kvm_vcpu *vcpu, struct kvm_mmu *context, bool execonly) argument [all...] |
/arch/x86/pci/ |
H A D | mmconfig-shared.c | 413 void *context, void **rv) 415 struct resource *mcfg_res = context; 418 check_mcfg_resource, context); 412 find_mboard_resource(acpi_handle handle, u32 lvl, void *context, void **rv) argument
|
/arch/x86/platform/olpc/ |
H A D | olpc-xo15-sci.c | 136 static u32 xo15_sci_gpe_handler(acpi_handle gpe_device, u32 gpe, void *context) argument
|
/arch/x86/power/ |
H A D | hibernate_64.c | 42 static void *alloc_pgt_page(void *context) argument
|
/arch/blackfin/kernel/ |
H A D | trace.c | 736 unsigned int context = bfin_read_IPEND(); local 741 if (context & 0x0020 && (fp->seqstat & SEQSTAT_EXCAUSE) == VEC_HWERR) 742 pr_notice("HW Error context\n"); 743 else if (context & 0x0020) 744 pr_notice("Deferred Exception context\n"); 745 else if (context & 0x3FC0) 746 pr_notice("Interrupt context\n"); 747 else if (context & 0x4000) 748 pr_notice("Deferred Interrupt context\n"); 749 else if (context [all...] |