/arch/sparc/mm/ |
H A D | tlb.c | 26 struct tlb_batch *tb = &get_cpu_var(tlb_batch); local 28 if (tb->tlb_nr) { 29 flush_tsb_user(tb); 31 if (CTX_VALID(tb->mm->context)) { 33 smp_flush_tlb_pending(tb->mm, tb->tlb_nr, 34 &tb->vaddrs[0]); 36 __flush_tlb_pending(CTX_HWBITS(tb->mm->context), 37 tb->tlb_nr, &tb 49 struct tlb_batch *tb = &get_cpu_var(tlb_batch); local [all...] |
H A D | tsb.c | 49 static void __flush_tsb_one(struct tlb_batch *tb, unsigned long hash_shift, argument 54 for (i = 0; i < tb->tlb_nr; i++) { 55 unsigned long v = tb->vaddrs[i]; 68 void flush_tsb_user(struct tlb_batch *tb) argument 70 struct mm_struct *mm = tb->mm; 79 __flush_tsb_one(tb, PAGE_SHIFT, base, nentries); 87 __flush_tsb_one(tb, HPAGE_SHIFT, base, nentries);
|
/arch/m68k/platform/coldfire/ |
H A D | intc-2.c | 148 u16 pa, tb; local 152 tb = 0x1; 155 tb = 0x2; 158 tb = 0x3; 162 tb = 0; 166 if (tb) 171 pa = (pa & ~(0x3 << (irq * 2))) | (tb << (irq * 2));
|
H A D | intc-simr.c | 125 u16 pa, tb; local 129 tb = 0x1; 132 tb = 0x2; 135 tb = 0x3; 139 tb = 0; 143 if (tb) 148 pa = (pa & ~(0x3 << ebit)) | (tb << ebit);
|
/arch/powerpc/kernel/ |
H A D | smp-tbsync.c | 25 volatile u64 tb; member in struct:__anon2463 48 u64 tb; local 63 tb = tbsync->tb; 72 set_tb(tb >> 32, tb & 0xfffffffful); 81 u64 tb; local 88 tb = get_tb() + 400; 89 tbsync->tb = tb [all...] |
H A D | time.c | 184 static u64 read_spurr(u64 tb) argument 190 return tb;
|
/arch/x86/crypto/ |
H A D | fpu.c | 101 static struct crypto_instance *crypto_fpu_alloc(struct rtattr **tb) argument 107 err = crypto_check_attr_type(tb, CRYPTO_ALG_TYPE_BLKCIPHER); 111 alg = crypto_get_attr_alg(tb, CRYPTO_ALG_TYPE_BLKCIPHER,
|
/arch/ia64/include/asm/ |
H A D | kprobes.h | 49 unsigned long long tb : 1; member in struct:cmp_inst::__anon1555
|
H A D | processor.h | 108 __u64 tb : 1; member in struct:ia64_psr
|
/arch/powerpc/kvm/ |
H A D | emulate.c | 113 u32 kvmppc_get_dec(struct kvm_vcpu *vcpu, u64 tb) argument 115 u64 jd = tb - vcpu->arch.dec_jiffies;
|
H A D | booke.c | 741 u64 tb = get_tb(); local 752 sregs->u.e.dec = kvmppc_get_dec(vcpu, tb); 753 sregs->u.e.tb = tb;
|
/arch/powerpc/boot/ |
H A D | devtree.c | 62 void dt_fixup_cpu_clocks(u32 cpu, u32 tb, u32 bus) argument 67 printf("CPU timebase-frequency <- 0x%x (%dMHz)\n\r", tb, MHZ(tb)); 73 setprop_val(devp, "timebase-frequency", tb); 78 timebase_period_ns = 1000000000 / tb;
|
H A D | 4xx.c | 339 u32 cpu, plb, opb, ebc, tb, uart0, uart1, m; local 363 tb = sys_clk; 366 tb = cpu; 385 dt_fixup_cpu_clocks(cpu, tb, 0); 426 u32 ccr1, tb = tmr_clk; local 469 if (tb == 0) { 474 tb = cpu; 476 dt_fixup_cpu_clocks(cpu, tb, 0); 558 u32 cpu, plb, opb, ebc, tb, uart0, uart1, m; local 614 tb 632 u32 pllmr0_ccdv, tb, m; local 753 u32 cpu, plb, opb, ebc, vco, tb, uart0, uart1; local [all...] |
/arch/x86/kernel/cpu/mcheck/ |
H A D | mce_amd.c | 381 struct threshold_block *tb; member in struct:threshold_block_cross_cpu 388 struct threshold_block *b = tbcc->tb; 397 struct threshold_block_cross_cpu tbcc = { .tb = b, };
|
/arch/powerpc/include/asm/ |
H A D | kvm.h | 220 __u64 tb; member in struct:kvm_sregs::__anon2373::__anon2378
|
H A D | kvm_host.h | 314 bool tb : 1; /* 1TB segment */ member in struct:kvmppc_slb
|
/arch/sparc/kernel/ |
H A D | irq_64.c | 853 struct trap_per_cpu *tb = &trap_block[this_cpu]; local 855 register_one_mondo(tb->cpu_mondo_pa, HV_CPU_QUEUE_CPU_MONDO, 856 tb->cpu_mondo_qmask); 857 register_one_mondo(tb->dev_mondo_pa, HV_CPU_QUEUE_DEVICE_MONDO, 858 tb->dev_mondo_qmask); 859 register_one_mondo(tb->resum_mondo_pa, HV_CPU_QUEUE_RES_ERROR, 860 tb->resum_qmask); 861 register_one_mondo(tb->nonresum_mondo_pa, HV_CPU_QUEUE_NONRES_ERROR, 862 tb->nonresum_qmask); 884 static void __init init_cpu_send_mondo_info(struct trap_per_cpu *tb) argument 908 struct trap_per_cpu *tb = &trap_block[cpu]; local 925 struct trap_per_cpu *tb = &trap_block[cpu]; local [all...] |
H A D | mdesc.c | 746 struct trap_per_cpu *tb) 752 get_one_mondo_bits(val, &tb->cpu_mondo_qmask, 7, ilog2(max_cpus * 2)); 755 get_one_mondo_bits(val, &tb->dev_mondo_qmask, 7, 8); 758 get_one_mondo_bits(val, &tb->resum_qmask, 6, 7); 761 get_one_mondo_bits(val, &tb->nonresum_qmask, 2, 2); 765 tb->cpu_mondo_qmask + 1, 766 tb->dev_mondo_qmask + 1, 767 tb->resum_qmask + 1, 768 tb->nonresum_qmask + 1); 823 struct trap_per_cpu *tb; local 745 get_mondo_data(struct mdesc_handle *hp, u64 mp, struct trap_per_cpu *tb) argument [all...] |
H A D | smp_64.c | 291 struct trap_per_cpu *tb; local 310 tb = &trap_block[cpu]; 312 hdesc->fault_info_va = (unsigned long) &tb->fault_info; 313 hdesc->fault_info_pa = kimage_addr_to_ra(&tb->fault_info); 462 static void spitfire_xcall_deliver(struct trap_per_cpu *tb, int cnt) argument 470 cpu_list = __va(tb->cpu_list_pa); 471 mondo = __va(tb->cpu_mondo_block_pa); 483 static void cheetah_xcall_deliver(struct trap_per_cpu *tb, int cnt) argument 489 cpu_list = __va(tb->cpu_list_pa); 490 mondo = __va(tb 627 hypervisor_xcall_deliver(struct trap_per_cpu *tb, int cnt) argument 743 struct trap_per_cpu *tb; local 1260 struct trap_per_cpu *tb = &trap_block[cpu]; local [all...] |
H A D | traps_64.c | 1859 struct trap_per_cpu *tb; local 1865 tb = &trap_block[cpu]; 1866 paddr = tb->resum_kernel_buf_pa + offset; 1907 struct trap_per_cpu *tb; local 1913 tb = &trap_block[cpu]; 1914 paddr = tb->nonresum_kernel_buf_pa + offset;
|