1/*
2 * common defines for all CPUs
3 *
4 * Copyright (c) 2003 Fabrice Bellard
5 *
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
10 *
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
14 * Lesser General Public License for more details.
15 *
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
18 */
19#ifndef CPU_DEFS_H
20#define CPU_DEFS_H
21
22#ifndef NEED_CPU_H
23#error cpu.h included from common code
24#endif
25
26#include "config.h"
27#include <setjmp.h>
28#include <inttypes.h>
29#include <signal.h>
30#include "osdep.h"
31#include "qemu-queue.h"
32#include "targphys.h"
33
34#ifndef TARGET_LONG_BITS
35#error TARGET_LONG_BITS must be defined before including this header
36#endif
37
38#define TARGET_LONG_SIZE (TARGET_LONG_BITS / 8)
39
40/* target_ulong is the type of a virtual address */
41#if TARGET_LONG_SIZE == 4
42typedef int32_t target_long;
43typedef uint32_t target_ulong;
44#define TARGET_FMT_lx "%08x"
45#define TARGET_FMT_ld "%d"
46#define TARGET_FMT_lu "%u"
47#elif TARGET_LONG_SIZE == 8
48typedef int64_t target_long;
49typedef uint64_t target_ulong;
50#define TARGET_FMT_lx "%016" PRIx64
51#define TARGET_FMT_ld "%" PRId64
52#define TARGET_FMT_lu "%" PRIu64
53#else
54#error TARGET_LONG_SIZE undefined
55#endif
56
57#define HOST_LONG_SIZE (HOST_LONG_BITS / 8)
58
59#define EXCP_INTERRUPT 	0x10000 /* async interruption */
60#define EXCP_HLT        0x10001 /* hlt instruction reached */
61#define EXCP_DEBUG      0x10002 /* cpu stopped after a breakpoint or singlestep */
62#define EXCP_HALTED     0x10003 /* cpu is halted (waiting for external event) */
63
64#define TB_JMP_CACHE_BITS 12
65#define TB_JMP_CACHE_SIZE (1 << TB_JMP_CACHE_BITS)
66
67/* Only the bottom TB_JMP_PAGE_BITS of the jump cache hash bits vary for
68   addresses on the same page.  The top bits are the same.  This allows
69   TLB invalidation to quickly clear a subset of the hash table.  */
70#define TB_JMP_PAGE_BITS (TB_JMP_CACHE_BITS / 2)
71#define TB_JMP_PAGE_SIZE (1 << TB_JMP_PAGE_BITS)
72#define TB_JMP_ADDR_MASK (TB_JMP_PAGE_SIZE - 1)
73#define TB_JMP_PAGE_MASK (TB_JMP_CACHE_SIZE - TB_JMP_PAGE_SIZE)
74
75#if !defined(CONFIG_USER_ONLY)
76#define CPU_TLB_BITS 8
77#define CPU_TLB_SIZE (1 << CPU_TLB_BITS)
78
79#if HOST_LONG_BITS == 32 && TARGET_LONG_BITS == 32
80#define CPU_TLB_ENTRY_BITS 4
81#else
82#define CPU_TLB_ENTRY_BITS 5
83#endif
84
85typedef struct CPUTLBEntry {
86    /* bit TARGET_LONG_BITS to TARGET_PAGE_BITS : virtual address
87       bit TARGET_PAGE_BITS-1..4  : Nonzero for accesses that should not
88                                    go directly to ram.
89       bit 3                      : indicates that the entry is invalid
90       bit 2..0                   : zero
91    */
92    target_ulong addr_read;
93    target_ulong addr_write;
94    target_ulong addr_code;
95    /* Addend to virtual address to get host address.  IO accesses
96       use the corresponding iotlb value.  */
97    size_t addend;
98    /* padding to get a power of two size */
99    uint8_t dummy[(1 << CPU_TLB_ENTRY_BITS) -
100                  (sizeof(target_ulong) * 3 +
101                   ((-sizeof(target_ulong) * 3) & (sizeof(size_t) - 1)) +
102                   sizeof(size_t))];
103} CPUTLBEntry;
104
105extern int CPUTLBEntry_wrong_size[sizeof(CPUTLBEntry) == (1 << CPU_TLB_ENTRY_BITS) ? 1 : -1];
106
107#define CPU_COMMON_TLB \
108    /* The meaning of the MMU modes is defined in the target code. */   \
109    CPUTLBEntry tlb_table[NB_MMU_MODES][CPU_TLB_SIZE];                  \
110    target_phys_addr_t iotlb[NB_MMU_MODES][CPU_TLB_SIZE];               \
111    target_ulong tlb_flush_addr;                                        \
112    target_ulong tlb_flush_mask;
113
114#else
115
116#define CPU_COMMON_TLB
117
118#endif
119
120
121#ifdef HOST_WORDS_BIGENDIAN
122typedef struct icount_decr_u16 {
123    uint16_t high;
124    uint16_t low;
125} icount_decr_u16;
126#else
127typedef struct icount_decr_u16 {
128    uint16_t low;
129    uint16_t high;
130} icount_decr_u16;
131#endif
132
133struct kvm_run;
134struct KVMState;
135struct qemu_work_item;
136
137typedef struct CPUBreakpoint {
138    target_ulong pc;
139    int flags; /* BP_* */
140    QTAILQ_ENTRY(CPUBreakpoint) entry;
141} CPUBreakpoint;
142
143typedef struct CPUWatchpoint {
144    target_ulong vaddr;
145    target_ulong len_mask;
146    int flags; /* BP_* */
147    QTAILQ_ENTRY(CPUWatchpoint) entry;
148} CPUWatchpoint;
149
150#define CPU_TEMP_BUF_NLONGS 128
151#define CPU_COMMON                                                      \
152    struct TranslationBlock *current_tb; /* currently executing TB  */  \
153    /* soft mmu support */                                              \
154    /* in order to avoid passing too many arguments to the MMIO         \
155       helpers, we store some rarely used information in the CPU        \
156       context) */                                                      \
157    unsigned long mem_io_pc; /* host pc at which the memory was         \
158                                accessed */                             \
159    target_ulong mem_io_vaddr; /* target virtual addr at which the      \
160                                     memory was accessed */             \
161    uint32_t halted; /* Nonzero if the CPU is in suspend state */       \
162    uint32_t interrupt_request;                                         \
163    volatile sig_atomic_t exit_request;                                 \
164    CPU_COMMON_TLB                                                      \
165    struct TranslationBlock *tb_jmp_cache[TB_JMP_CACHE_SIZE];           \
166    /* buffer for temporaries in the code generator */                  \
167    long temp_buf[CPU_TEMP_BUF_NLONGS];                                 \
168                                                                        \
169    int64_t icount_extra; /* Instructions until next timer event.  */   \
170    /* Number of cycles left, with interrupt flag in high bit.          \
171       This allows a single read-compare-cbranch-write sequence to test \
172       for both decrementer underflow and exceptions.  */               \
173    union {                                                             \
174        uint32_t u32;                                                   \
175        icount_decr_u16 u16;                                            \
176    } icount_decr;                                                      \
177    uint32_t can_do_io; /* nonzero if memory mapped IO is safe.  */     \
178                                                                        \
179    /* from this point: preserved by CPU reset */                       \
180    /* ice debug support */                                             \
181    QTAILQ_HEAD(breakpoints_head, CPUBreakpoint) breakpoints;            \
182    int singlestep_enabled;                                             \
183                                                                        \
184    QTAILQ_HEAD(watchpoints_head, CPUWatchpoint) watchpoints;            \
185    CPUWatchpoint *watchpoint_hit;                                      \
186                                                                        \
187    struct GDBRegisterState *gdb_regs;                                  \
188                                                                        \
189    /* Core interrupt code */                                           \
190    jmp_buf jmp_env;                                                    \
191    int exception_index;                                                \
192                                                                        \
193    CPUState *next_cpu; /* next CPU sharing TB cache */                 \
194    int cpu_index; /* CPU index (informative) */                        \
195    uint32_t host_tid; /* host thread ID */                             \
196    int numa_node; /* NUMA node this cpu is belonging to  */            \
197    int nr_cores;  /* number of cores within this CPU package */        \
198    int nr_threads;/* number of threads within this CPU */              \
199    int running; /* Nonzero if cpu is currently running(usermode).  */  \
200    /* user data */                                                     \
201    void *opaque;                                                       \
202                                                                        \
203    uint32_t created;                                                   \
204    uint32_t stop;   /* Stop request */                                 \
205    uint32_t stopped; /* Artificially stopped */                        \
206    struct QemuThread *thread;                                          \
207    struct QemuCond *halt_cond;                                         \
208    struct qemu_work_item *queued_work_first, *queued_work_last;        \
209    const char *cpu_model_str;                                          \
210    struct KVMState *kvm_state;                                         \
211    struct kvm_run *kvm_run;                                            \
212    int kvm_fd;                                                         \
213    int kvm_vcpu_dirty;                                                 \
214    struct hax_vcpu_state *hax_vcpu;
215
216#endif
217