1/* 2 * Copyright (C) 2012 - Virtual Open Systems and Columbia University 3 * Author: Christoffer Dall <c.dall@virtualopensystems.com> 4 * 5 * This program is free software; you can redistribute it and/or modify 6 * it under the terms of the GNU General Public License, version 2, as 7 * published by the Free Software Foundation. 8 * 9 * This program is distributed in the hope that it will be useful, 10 * but WITHOUT ANY WARRANTY; without even the implied warranty of 11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 12 * GNU General Public License for more details. 13 * 14 * You should have received a copy of the GNU General Public License 15 * along with this program; if not, write to the Free Software 16 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. 17 */ 18 19#ifndef __ARM_KVM_EMULATE_H__ 20#define __ARM_KVM_EMULATE_H__ 21 22#include <linux/kvm_host.h> 23#include <asm/kvm_asm.h> 24#include <asm/kvm_mmio.h> 25#include <asm/kvm_arm.h> 26 27unsigned long *vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num); 28unsigned long *vcpu_spsr(struct kvm_vcpu *vcpu); 29 30bool kvm_condition_valid(struct kvm_vcpu *vcpu); 31void kvm_skip_instr(struct kvm_vcpu *vcpu, bool is_wide_instr); 32void kvm_inject_undefined(struct kvm_vcpu *vcpu); 33void kvm_inject_dabt(struct kvm_vcpu *vcpu, unsigned long addr); 34void kvm_inject_pabt(struct kvm_vcpu *vcpu, unsigned long addr); 35 36static inline bool vcpu_mode_is_32bit(struct kvm_vcpu *vcpu) 37{ 38 return 1; 39} 40 41static inline unsigned long *vcpu_pc(struct kvm_vcpu *vcpu) 42{ 43 return &vcpu->arch.regs.usr_regs.ARM_pc; 44} 45 46static inline unsigned long *vcpu_cpsr(struct kvm_vcpu *vcpu) 47{ 48 return &vcpu->arch.regs.usr_regs.ARM_cpsr; 49} 50 51static inline void vcpu_set_thumb(struct kvm_vcpu *vcpu) 52{ 53 *vcpu_cpsr(vcpu) |= PSR_T_BIT; 54} 55 56static inline bool mode_has_spsr(struct kvm_vcpu *vcpu) 57{ 58 unsigned long cpsr_mode = vcpu->arch.regs.usr_regs.ARM_cpsr & MODE_MASK; 59 return (cpsr_mode > USR_MODE && cpsr_mode < SYSTEM_MODE); 60} 61 62static inline bool vcpu_mode_priv(struct kvm_vcpu *vcpu) 63{ 64 unsigned long cpsr_mode = vcpu->arch.regs.usr_regs.ARM_cpsr & MODE_MASK; 65 return cpsr_mode > USR_MODE;; 66} 67 68static inline u32 kvm_vcpu_get_hsr(struct kvm_vcpu *vcpu) 69{ 70 return vcpu->arch.fault.hsr; 71} 72 73static inline unsigned long kvm_vcpu_get_hfar(struct kvm_vcpu *vcpu) 74{ 75 return vcpu->arch.fault.hxfar; 76} 77 78static inline phys_addr_t kvm_vcpu_get_fault_ipa(struct kvm_vcpu *vcpu) 79{ 80 return ((phys_addr_t)vcpu->arch.fault.hpfar & HPFAR_MASK) << 8; 81} 82 83static inline unsigned long kvm_vcpu_get_hyp_pc(struct kvm_vcpu *vcpu) 84{ 85 return vcpu->arch.fault.hyp_pc; 86} 87 88static inline bool kvm_vcpu_dabt_isvalid(struct kvm_vcpu *vcpu) 89{ 90 return kvm_vcpu_get_hsr(vcpu) & HSR_ISV; 91} 92 93static inline bool kvm_vcpu_dabt_iswrite(struct kvm_vcpu *vcpu) 94{ 95 return kvm_vcpu_get_hsr(vcpu) & HSR_WNR; 96} 97 98static inline bool kvm_vcpu_dabt_issext(struct kvm_vcpu *vcpu) 99{ 100 return kvm_vcpu_get_hsr(vcpu) & HSR_SSE; 101} 102 103static inline int kvm_vcpu_dabt_get_rd(struct kvm_vcpu *vcpu) 104{ 105 return (kvm_vcpu_get_hsr(vcpu) & HSR_SRT_MASK) >> HSR_SRT_SHIFT; 106} 107 108static inline bool kvm_vcpu_dabt_isextabt(struct kvm_vcpu *vcpu) 109{ 110 return kvm_vcpu_get_hsr(vcpu) & HSR_DABT_EA; 111} 112 113static inline bool kvm_vcpu_dabt_iss1tw(struct kvm_vcpu *vcpu) 114{ 115 return kvm_vcpu_get_hsr(vcpu) & HSR_DABT_S1PTW; 116} 117 118/* Get Access Size from a data abort */ 119static inline int kvm_vcpu_dabt_get_as(struct kvm_vcpu *vcpu) 120{ 121 switch ((kvm_vcpu_get_hsr(vcpu) >> 22) & 0x3) { 122 case 0: 123 return 1; 124 case 1: 125 return 2; 126 case 2: 127 return 4; 128 default: 129 kvm_err("Hardware is weird: SAS 0b11 is reserved\n"); 130 return -EFAULT; 131 } 132} 133 134/* This one is not specific to Data Abort */ 135static inline bool kvm_vcpu_trap_il_is32bit(struct kvm_vcpu *vcpu) 136{ 137 return kvm_vcpu_get_hsr(vcpu) & HSR_IL; 138} 139 140static inline u8 kvm_vcpu_trap_get_class(struct kvm_vcpu *vcpu) 141{ 142 return kvm_vcpu_get_hsr(vcpu) >> HSR_EC_SHIFT; 143} 144 145static inline bool kvm_vcpu_trap_is_iabt(struct kvm_vcpu *vcpu) 146{ 147 return kvm_vcpu_trap_get_class(vcpu) == HSR_EC_IABT; 148} 149 150static inline u8 kvm_vcpu_trap_get_fault(struct kvm_vcpu *vcpu) 151{ 152 return kvm_vcpu_get_hsr(vcpu) & HSR_FSC; 153} 154 155static inline u8 kvm_vcpu_trap_get_fault_type(struct kvm_vcpu *vcpu) 156{ 157 return kvm_vcpu_get_hsr(vcpu) & HSR_FSC_TYPE; 158} 159 160static inline u32 kvm_vcpu_hvc_get_imm(struct kvm_vcpu *vcpu) 161{ 162 return kvm_vcpu_get_hsr(vcpu) & HSR_HVC_IMM_MASK; 163} 164 165static inline unsigned long kvm_vcpu_get_mpidr(struct kvm_vcpu *vcpu) 166{ 167 return vcpu->arch.cp15[c0_MPIDR]; 168} 169 170static inline void kvm_vcpu_set_be(struct kvm_vcpu *vcpu) 171{ 172 *vcpu_cpsr(vcpu) |= PSR_E_BIT; 173} 174 175static inline bool kvm_vcpu_is_be(struct kvm_vcpu *vcpu) 176{ 177 return !!(*vcpu_cpsr(vcpu) & PSR_E_BIT); 178} 179 180static inline unsigned long vcpu_data_guest_to_host(struct kvm_vcpu *vcpu, 181 unsigned long data, 182 unsigned int len) 183{ 184 if (kvm_vcpu_is_be(vcpu)) { 185 switch (len) { 186 case 1: 187 return data & 0xff; 188 case 2: 189 return be16_to_cpu(data & 0xffff); 190 default: 191 return be32_to_cpu(data); 192 } 193 } else { 194 switch (len) { 195 case 1: 196 return data & 0xff; 197 case 2: 198 return le16_to_cpu(data & 0xffff); 199 default: 200 return le32_to_cpu(data); 201 } 202 } 203} 204 205static inline unsigned long vcpu_data_host_to_guest(struct kvm_vcpu *vcpu, 206 unsigned long data, 207 unsigned int len) 208{ 209 if (kvm_vcpu_is_be(vcpu)) { 210 switch (len) { 211 case 1: 212 return data & 0xff; 213 case 2: 214 return cpu_to_be16(data & 0xffff); 215 default: 216 return cpu_to_be32(data); 217 } 218 } else { 219 switch (len) { 220 case 1: 221 return data & 0xff; 222 case 2: 223 return cpu_to_le16(data & 0xffff); 224 default: 225 return cpu_to_le32(data); 226 } 227 } 228} 229 230#endif /* __ARM_KVM_EMULATE_H__ */ 231