1 | /*
|
---|
2 | * i386 virtual CPU header
|
---|
3 | *
|
---|
4 | * Copyright (c) 2003 Fabrice Bellard
|
---|
5 | *
|
---|
6 | * This library is free software; you can redistribute it and/or
|
---|
7 | * modify it under the terms of the GNU Lesser General Public
|
---|
8 | * License as published by the Free Software Foundation; either
|
---|
9 | * version 2 of the License, or (at your option) any later version.
|
---|
10 | *
|
---|
11 | * This library is distributed in the hope that it will be useful,
|
---|
12 | * but WITHOUT ANY WARRANTY; without even the implied warranty of
|
---|
13 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
---|
14 | * Lesser General Public License for more details.
|
---|
15 | *
|
---|
16 | * You should have received a copy of the GNU Lesser General Public
|
---|
17 | * License along with this library; if not, see <http://www.gnu.org/licenses/>.
|
---|
18 | */
|
---|
19 |
|
---|
20 | /*
|
---|
21 | * Oracle LGPL Disclaimer: For the avoidance of doubt, except that if any license choice
|
---|
22 | * other than GPL or LGPL is available it will apply instead, Oracle elects to use only
|
---|
23 | * the Lesser General Public License version 2.1 (LGPLv2) at this time for any software where
|
---|
24 | * a choice of LGPL license versions is made available with the language indicating
|
---|
25 | * that LGPLv2 or any later version may be used, or where a choice of which version
|
---|
26 | * of the LGPL is applied is otherwise unspecified.
|
---|
27 | */
|
---|
28 |
|
---|
29 | #ifndef CPU_I386_H
|
---|
30 | #define CPU_I386_H
|
---|
31 |
|
---|
32 | #include "config.h"
|
---|
33 |
|
---|
34 | #ifdef TARGET_X86_64
|
---|
35 | #define TARGET_LONG_BITS 64
|
---|
36 | #else
|
---|
37 | #define TARGET_LONG_BITS 32
|
---|
38 | #endif
|
---|
39 |
|
---|
40 | /* target supports implicit self modifying code */
|
---|
41 | #define TARGET_HAS_SMC
|
---|
42 | /* support for self modifying code even if the modified instruction is
|
---|
43 | close to the modifying instruction */
|
---|
44 | #define TARGET_HAS_PRECISE_SMC
|
---|
45 |
|
---|
46 | #define TARGET_HAS_ICE 1
|
---|
47 |
|
---|
48 | #ifdef TARGET_X86_64
|
---|
49 | #define ELF_MACHINE EM_X86_64
|
---|
50 | #else
|
---|
51 | #define ELF_MACHINE EM_386
|
---|
52 | #endif
|
---|
53 |
|
---|
54 | #define CPUState struct CPUX86State
|
---|
55 |
|
---|
56 | #include "cpu-defs.h"
|
---|
57 |
|
---|
58 | #include "softfloat.h"
|
---|
59 |
|
---|
60 | #if defined(VBOX)
|
---|
61 | # include <iprt/critsect.h>
|
---|
62 | # include <iprt/thread.h>
|
---|
63 | # include <iprt/assert.h>
|
---|
64 | # include <iprt/asm.h>
|
---|
65 | # include <VBox/vmm/vmm.h>
|
---|
66 | # include <VBox/vmm/stam.h>
|
---|
67 | #endif /* VBOX */
|
---|
68 |
|
---|
69 | #define R_EAX 0
|
---|
70 | #define R_ECX 1
|
---|
71 | #define R_EDX 2
|
---|
72 | #define R_EBX 3
|
---|
73 | #define R_ESP 4
|
---|
74 | #define R_EBP 5
|
---|
75 | #define R_ESI 6
|
---|
76 | #define R_EDI 7
|
---|
77 |
|
---|
78 | #define R_AL 0
|
---|
79 | #define R_CL 1
|
---|
80 | #define R_DL 2
|
---|
81 | #define R_BL 3
|
---|
82 | #define R_AH 4
|
---|
83 | #define R_CH 5
|
---|
84 | #define R_DH 6
|
---|
85 | #define R_BH 7
|
---|
86 |
|
---|
87 | #define R_ES 0
|
---|
88 | #define R_CS 1
|
---|
89 | #define R_SS 2
|
---|
90 | #define R_DS 3
|
---|
91 | #define R_FS 4
|
---|
92 | #define R_GS 5
|
---|
93 |
|
---|
94 | /* segment descriptor fields */
|
---|
95 | #define DESC_G_MASK (1 << 23)
|
---|
96 | #define DESC_B_SHIFT 22
|
---|
97 | #define DESC_B_MASK (1 << DESC_B_SHIFT)
|
---|
98 | #define DESC_L_SHIFT 21 /* x86_64 only : 64 bit code segment */
|
---|
99 | #define DESC_L_MASK (1 << DESC_L_SHIFT)
|
---|
100 | #define DESC_AVL_MASK (1 << 20)
|
---|
101 | #define DESC_P_MASK (1 << 15)
|
---|
102 | #define DESC_DPL_SHIFT 13
|
---|
103 | #define DESC_DPL_MASK (3 << DESC_DPL_SHIFT)
|
---|
104 | #define DESC_S_MASK (1 << 12)
|
---|
105 | #define DESC_TYPE_SHIFT 8
|
---|
106 | #define DESC_TYPE_MASK (15 << DESC_TYPE_SHIFT)
|
---|
107 | #define DESC_A_MASK (1 << 8)
|
---|
108 |
|
---|
109 | #define DESC_CS_MASK (1 << 11) /* 1=code segment 0=data segment */
|
---|
110 | #define DESC_C_MASK (1 << 10) /* code: conforming */
|
---|
111 | #define DESC_R_MASK (1 << 9) /* code: readable */
|
---|
112 |
|
---|
113 | #define DESC_E_MASK (1 << 10) /* data: expansion direction */
|
---|
114 | #define DESC_W_MASK (1 << 9) /* data: writable */
|
---|
115 |
|
---|
116 | #define DESC_TSS_BUSY_MASK (1 << 9)
|
---|
117 |
|
---|
118 | /* eflags masks */
|
---|
119 | #define CC_C 0x0001
|
---|
120 | #define CC_P 0x0004
|
---|
121 | #define CC_A 0x0010
|
---|
122 | #define CC_Z 0x0040
|
---|
123 | #define CC_S 0x0080
|
---|
124 | #define CC_O 0x0800
|
---|
125 |
|
---|
126 | #define TF_SHIFT 8
|
---|
127 | #define IOPL_SHIFT 12
|
---|
128 | #define VM_SHIFT 17
|
---|
129 |
|
---|
130 | #define TF_MASK 0x00000100
|
---|
131 | #define IF_MASK 0x00000200
|
---|
132 | #define DF_MASK 0x00000400
|
---|
133 | #define IOPL_MASK 0x00003000
|
---|
134 | #define NT_MASK 0x00004000
|
---|
135 | #define RF_MASK 0x00010000
|
---|
136 | #define VM_MASK 0x00020000
|
---|
137 | #define AC_MASK 0x00040000
|
---|
138 | #define VIF_MASK 0x00080000
|
---|
139 | #define VIP_MASK 0x00100000
|
---|
140 | #define ID_MASK 0x00200000
|
---|
141 |
|
---|
142 | /* hidden flags - used internally by qemu to represent additional cpu
|
---|
143 | states. Only the CPL, INHIBIT_IRQ, SMM and SVMI are not
|
---|
144 | redundant. We avoid using the IOPL_MASK, TF_MASK and VM_MASK bit
|
---|
145 | position to ease oring with eflags. */
|
---|
146 | /* current cpl */
|
---|
147 | #define HF_CPL_SHIFT 0
|
---|
148 | /* true if soft mmu is being used */
|
---|
149 | #define HF_SOFTMMU_SHIFT 2
|
---|
150 | /* true if hardware interrupts must be disabled for next instruction */
|
---|
151 | #define HF_INHIBIT_IRQ_SHIFT 3
|
---|
152 | /* 16 or 32 segments */
|
---|
153 | #define HF_CS32_SHIFT 4
|
---|
154 | #define HF_SS32_SHIFT 5
|
---|
155 | /* zero base for DS, ES and SS : can be '0' only in 32 bit CS segment */
|
---|
156 | #define HF_ADDSEG_SHIFT 6
|
---|
157 | /* copy of CR0.PE (protected mode) */
|
---|
158 | #define HF_PE_SHIFT 7
|
---|
159 | #define HF_TF_SHIFT 8 /* must be same as eflags */
|
---|
160 | #define HF_MP_SHIFT 9 /* the order must be MP, EM, TS */
|
---|
161 | #define HF_EM_SHIFT 10
|
---|
162 | #define HF_TS_SHIFT 11
|
---|
163 | #define HF_IOPL_SHIFT 12 /* must be same as eflags */
|
---|
164 | #define HF_LMA_SHIFT 14 /* only used on x86_64: long mode active */
|
---|
165 | #define HF_CS64_SHIFT 15 /* only used on x86_64: 64 bit code segment */
|
---|
166 | #define HF_RF_SHIFT 16 /* must be same as eflags */
|
---|
167 | #define HF_VM_SHIFT 17 /* must be same as eflags */
|
---|
168 | #define HF_SMM_SHIFT 19 /* CPU in SMM mode */
|
---|
169 | #define HF_SVME_SHIFT 20 /* SVME enabled (copy of EFER.SVME) */
|
---|
170 | #define HF_SVMI_SHIFT 21 /* SVM intercepts are active */
|
---|
171 | #define HF_OSFXSR_SHIFT 22 /* CR4.OSFXSR */
|
---|
172 |
|
---|
173 | #define HF_CPL_MASK (3 << HF_CPL_SHIFT)
|
---|
174 | #define HF_SOFTMMU_MASK (1 << HF_SOFTMMU_SHIFT)
|
---|
175 | #define HF_INHIBIT_IRQ_MASK (1 << HF_INHIBIT_IRQ_SHIFT)
|
---|
176 | #define HF_CS32_MASK (1 << HF_CS32_SHIFT)
|
---|
177 | #define HF_SS32_MASK (1 << HF_SS32_SHIFT)
|
---|
178 | #define HF_ADDSEG_MASK (1 << HF_ADDSEG_SHIFT)
|
---|
179 | #define HF_PE_MASK (1 << HF_PE_SHIFT)
|
---|
180 | #define HF_TF_MASK (1 << HF_TF_SHIFT)
|
---|
181 | #define HF_MP_MASK (1 << HF_MP_SHIFT)
|
---|
182 | #define HF_EM_MASK (1 << HF_EM_SHIFT)
|
---|
183 | #define HF_TS_MASK (1 << HF_TS_SHIFT)
|
---|
184 | #define HF_IOPL_MASK (3 << HF_IOPL_SHIFT)
|
---|
185 | #define HF_LMA_MASK (1 << HF_LMA_SHIFT)
|
---|
186 | #define HF_CS64_MASK (1 << HF_CS64_SHIFT)
|
---|
187 | #define HF_RF_MASK (1 << HF_RF_SHIFT)
|
---|
188 | #define HF_VM_MASK (1 << HF_VM_SHIFT)
|
---|
189 | #define HF_SMM_MASK (1 << HF_SMM_SHIFT)
|
---|
190 | #define HF_SVME_MASK (1 << HF_SVME_SHIFT)
|
---|
191 | #define HF_SVMI_MASK (1 << HF_SVMI_SHIFT)
|
---|
192 | #define HF_OSFXSR_MASK (1 << HF_OSFXSR_SHIFT)
|
---|
193 |
|
---|
194 | /* hflags2 */
|
---|
195 |
|
---|
196 | #define HF2_GIF_SHIFT 0 /* if set CPU takes interrupts */
|
---|
197 | #define HF2_HIF_SHIFT 1 /* value of IF_MASK when entering SVM */
|
---|
198 | #define HF2_NMI_SHIFT 2 /* CPU serving NMI */
|
---|
199 | #define HF2_VINTR_SHIFT 3 /* value of V_INTR_MASKING bit */
|
---|
200 |
|
---|
201 | #define HF2_GIF_MASK (1 << HF2_GIF_SHIFT)
|
---|
202 | #define HF2_HIF_MASK (1 << HF2_HIF_SHIFT)
|
---|
203 | #define HF2_NMI_MASK (1 << HF2_NMI_SHIFT)
|
---|
204 | #define HF2_VINTR_MASK (1 << HF2_VINTR_SHIFT)
|
---|
205 |
|
---|
206 | #define CR0_PE_SHIFT 0
|
---|
207 | #define CR0_MP_SHIFT 1
|
---|
208 |
|
---|
209 | #define CR0_PE_MASK (1 << 0)
|
---|
210 | #define CR0_MP_MASK (1 << 1)
|
---|
211 | #define CR0_EM_MASK (1 << 2)
|
---|
212 | #define CR0_TS_MASK (1 << 3)
|
---|
213 | #define CR0_ET_MASK (1 << 4)
|
---|
214 | #define CR0_NE_MASK (1 << 5)
|
---|
215 | #define CR0_WP_MASK (1 << 16)
|
---|
216 | #define CR0_AM_MASK (1 << 18)
|
---|
217 | #define CR0_PG_MASK (1 << 31)
|
---|
218 |
|
---|
219 | #define CR4_VME_MASK (1 << 0)
|
---|
220 | #define CR4_PVI_MASK (1 << 1)
|
---|
221 | #define CR4_TSD_MASK (1 << 2)
|
---|
222 | #define CR4_DE_MASK (1 << 3)
|
---|
223 | #define CR4_PSE_MASK (1 << 4)
|
---|
224 | #define CR4_PAE_MASK (1 << 5)
|
---|
225 | #define CR4_MCE_MASK (1 << 6)
|
---|
226 | #define CR4_PGE_MASK (1 << 7)
|
---|
227 | #define CR4_PCE_MASK (1 << 8)
|
---|
228 | #define CR4_OSFXSR_SHIFT 9
|
---|
229 | #define CR4_OSFXSR_MASK (1 << CR4_OSFXSR_SHIFT)
|
---|
230 | #define CR4_OSXMMEXCPT_MASK (1 << 10)
|
---|
231 |
|
---|
232 | #define DR6_BD (1 << 13)
|
---|
233 | #define DR6_BS (1 << 14)
|
---|
234 | #define DR6_BT (1 << 15)
|
---|
235 | #define DR6_FIXED_1 0xffff0ff0
|
---|
236 |
|
---|
237 | #define DR7_GD (1 << 13)
|
---|
238 | #define DR7_TYPE_SHIFT 16
|
---|
239 | #define DR7_LEN_SHIFT 18
|
---|
240 | #define DR7_FIXED_1 0x00000400
|
---|
241 |
|
---|
242 | #define PG_PRESENT_BIT 0
|
---|
243 | #define PG_RW_BIT 1
|
---|
244 | #define PG_USER_BIT 2
|
---|
245 | #define PG_PWT_BIT 3
|
---|
246 | #define PG_PCD_BIT 4
|
---|
247 | #define PG_ACCESSED_BIT 5
|
---|
248 | #define PG_DIRTY_BIT 6
|
---|
249 | #define PG_PSE_BIT 7
|
---|
250 | #define PG_GLOBAL_BIT 8
|
---|
251 | #define PG_NX_BIT 63
|
---|
252 |
|
---|
253 | #define PG_PRESENT_MASK (1 << PG_PRESENT_BIT)
|
---|
254 | #define PG_RW_MASK (1 << PG_RW_BIT)
|
---|
255 | #define PG_USER_MASK (1 << PG_USER_BIT)
|
---|
256 | #define PG_PWT_MASK (1 << PG_PWT_BIT)
|
---|
257 | #define PG_PCD_MASK (1 << PG_PCD_BIT)
|
---|
258 | #define PG_ACCESSED_MASK (1 << PG_ACCESSED_BIT)
|
---|
259 | #define PG_DIRTY_MASK (1 << PG_DIRTY_BIT)
|
---|
260 | #define PG_PSE_MASK (1 << PG_PSE_BIT)
|
---|
261 | #define PG_GLOBAL_MASK (1 << PG_GLOBAL_BIT)
|
---|
262 | #define PG_NX_MASK (1LL << PG_NX_BIT)
|
---|
263 |
|
---|
264 | #define PG_ERROR_W_BIT 1
|
---|
265 |
|
---|
266 | #define PG_ERROR_P_MASK 0x01
|
---|
267 | #define PG_ERROR_W_MASK (1 << PG_ERROR_W_BIT)
|
---|
268 | #define PG_ERROR_U_MASK 0x04
|
---|
269 | #define PG_ERROR_RSVD_MASK 0x08
|
---|
270 | #define PG_ERROR_I_D_MASK 0x10
|
---|
271 |
|
---|
272 | #define MCG_CTL_P (1UL<<8) /* MCG_CAP register available */
|
---|
273 |
|
---|
274 | #define MCE_CAP_DEF MCG_CTL_P
|
---|
275 | #define MCE_BANKS_DEF 10
|
---|
276 |
|
---|
277 | #define MCG_STATUS_MCIP (1ULL<<2) /* machine check in progress */
|
---|
278 |
|
---|
279 | #define MCI_STATUS_VAL (1ULL<<63) /* valid error */
|
---|
280 | #define MCI_STATUS_OVER (1ULL<<62) /* previous errors lost */
|
---|
281 | #define MCI_STATUS_UC (1ULL<<61) /* uncorrected error */
|
---|
282 |
|
---|
283 | #define MSR_IA32_TSC 0x10
|
---|
284 | #define MSR_IA32_APICBASE 0x1b
|
---|
285 | #define MSR_IA32_APICBASE_BSP (1<<8)
|
---|
286 | #define MSR_IA32_APICBASE_ENABLE (1<<11)
|
---|
287 | #define MSR_IA32_APICBASE_BASE (0xfffff<<12)
|
---|
288 |
|
---|
289 | #define MSR_MTRRcap 0xfe
|
---|
290 | #define MSR_MTRRcap_VCNT 8
|
---|
291 | #define MSR_MTRRcap_FIXRANGE_SUPPORT (1 << 8)
|
---|
292 | #define MSR_MTRRcap_WC_SUPPORTED (1 << 10)
|
---|
293 |
|
---|
294 | #define MSR_IA32_SYSENTER_CS 0x174
|
---|
295 | #define MSR_IA32_SYSENTER_ESP 0x175
|
---|
296 | #define MSR_IA32_SYSENTER_EIP 0x176
|
---|
297 |
|
---|
298 | #define MSR_MCG_CAP 0x179
|
---|
299 | #define MSR_MCG_STATUS 0x17a
|
---|
300 | #define MSR_MCG_CTL 0x17b
|
---|
301 |
|
---|
302 | #define MSR_IA32_PERF_STATUS 0x198
|
---|
303 |
|
---|
304 | #define MSR_MTRRphysBase(reg) (0x200 + 2 * (reg))
|
---|
305 | #define MSR_MTRRphysMask(reg) (0x200 + 2 * (reg) + 1)
|
---|
306 |
|
---|
307 | #define MSR_MTRRfix64K_00000 0x250
|
---|
308 | #define MSR_MTRRfix16K_80000 0x258
|
---|
309 | #define MSR_MTRRfix16K_A0000 0x259
|
---|
310 | #define MSR_MTRRfix4K_C0000 0x268
|
---|
311 | #define MSR_MTRRfix4K_C8000 0x269
|
---|
312 | #define MSR_MTRRfix4K_D0000 0x26a
|
---|
313 | #define MSR_MTRRfix4K_D8000 0x26b
|
---|
314 | #define MSR_MTRRfix4K_E0000 0x26c
|
---|
315 | #define MSR_MTRRfix4K_E8000 0x26d
|
---|
316 | #define MSR_MTRRfix4K_F0000 0x26e
|
---|
317 | #define MSR_MTRRfix4K_F8000 0x26f
|
---|
318 |
|
---|
319 | #define MSR_PAT 0x277
|
---|
320 |
|
---|
321 | #define MSR_MTRRdefType 0x2ff
|
---|
322 |
|
---|
323 | #define MSR_MC0_CTL 0x400
|
---|
324 | #define MSR_MC0_STATUS 0x401
|
---|
325 | #define MSR_MC0_ADDR 0x402
|
---|
326 | #define MSR_MC0_MISC 0x403
|
---|
327 |
|
---|
328 | #define MSR_EFER 0xc0000080
|
---|
329 |
|
---|
330 | #define MSR_EFER_SCE (1 << 0)
|
---|
331 | #define MSR_EFER_LME (1 << 8)
|
---|
332 | #define MSR_EFER_LMA (1 << 10)
|
---|
333 | #define MSR_EFER_NXE (1 << 11)
|
---|
334 | #define MSR_EFER_SVME (1 << 12)
|
---|
335 | #define MSR_EFER_FFXSR (1 << 14)
|
---|
336 |
|
---|
337 | #ifdef VBOX
|
---|
338 | # define MSR_APIC_RANGE_START 0x800
|
---|
339 | # define MSR_APIC_RANGE_END 0x900
|
---|
340 | #endif
|
---|
341 |
|
---|
342 | #define MSR_STAR 0xc0000081
|
---|
343 | #define MSR_LSTAR 0xc0000082
|
---|
344 | #define MSR_CSTAR 0xc0000083
|
---|
345 | #define MSR_FMASK 0xc0000084
|
---|
346 | #define MSR_FSBASE 0xc0000100
|
---|
347 | #define MSR_GSBASE 0xc0000101
|
---|
348 | #define MSR_KERNELGSBASE 0xc0000102
|
---|
349 |
|
---|
350 | #define MSR_VM_HSAVE_PA 0xc0010117
|
---|
351 |
|
---|
352 | /* cpuid_features bits */
|
---|
353 | #define CPUID_FP87 (1 << 0)
|
---|
354 | #define CPUID_VME (1 << 1)
|
---|
355 | #define CPUID_DE (1 << 2)
|
---|
356 | #define CPUID_PSE (1 << 3)
|
---|
357 | #define CPUID_TSC (1 << 4)
|
---|
358 | #define CPUID_MSR (1 << 5)
|
---|
359 | #define CPUID_PAE (1 << 6)
|
---|
360 | #define CPUID_MCE (1 << 7)
|
---|
361 | #define CPUID_CX8 (1 << 8)
|
---|
362 | #define CPUID_APIC (1 << 9)
|
---|
363 | #define CPUID_SEP (1 << 11) /* sysenter/sysexit */
|
---|
364 | #define CPUID_MTRR (1 << 12)
|
---|
365 | #define CPUID_PGE (1 << 13)
|
---|
366 | #define CPUID_MCA (1 << 14)
|
---|
367 | #define CPUID_CMOV (1 << 15)
|
---|
368 | #define CPUID_PAT (1 << 16)
|
---|
369 | #define CPUID_PSE36 (1 << 17)
|
---|
370 | #define CPUID_PN (1 << 18)
|
---|
371 | #define CPUID_CLFLUSH (1 << 19)
|
---|
372 | #define CPUID_DTS (1 << 21)
|
---|
373 | #define CPUID_ACPI (1 << 22)
|
---|
374 | #define CPUID_MMX (1 << 23)
|
---|
375 | #define CPUID_FXSR (1 << 24)
|
---|
376 | #define CPUID_SSE (1 << 25)
|
---|
377 | #define CPUID_SSE2 (1 << 26)
|
---|
378 | #define CPUID_SS (1 << 27)
|
---|
379 | #define CPUID_HT (1 << 28)
|
---|
380 | #define CPUID_TM (1 << 29)
|
---|
381 | #define CPUID_IA64 (1 << 30)
|
---|
382 | #define CPUID_PBE (1 << 31)
|
---|
383 |
|
---|
384 | #define CPUID_EXT_SSE3 (1 << 0)
|
---|
385 | #define CPUID_EXT_DTES64 (1 << 2)
|
---|
386 | #define CPUID_EXT_MONITOR (1 << 3)
|
---|
387 | #define CPUID_EXT_DSCPL (1 << 4)
|
---|
388 | #define CPUID_EXT_VMX (1 << 5)
|
---|
389 | #define CPUID_EXT_SMX (1 << 6)
|
---|
390 | #define CPUID_EXT_EST (1 << 7)
|
---|
391 | #define CPUID_EXT_TM2 (1 << 8)
|
---|
392 | #define CPUID_EXT_SSSE3 (1 << 9)
|
---|
393 | #define CPUID_EXT_CID (1 << 10)
|
---|
394 | #define CPUID_EXT_CX16 (1 << 13)
|
---|
395 | #define CPUID_EXT_XTPR (1 << 14)
|
---|
396 | #define CPUID_EXT_PDCM (1 << 15)
|
---|
397 | #define CPUID_EXT_DCA (1 << 18)
|
---|
398 | #define CPUID_EXT_SSE41 (1 << 19)
|
---|
399 | #define CPUID_EXT_SSE42 (1 << 20)
|
---|
400 | #define CPUID_EXT_X2APIC (1 << 21)
|
---|
401 | #define CPUID_EXT_MOVBE (1 << 22)
|
---|
402 | #define CPUID_EXT_POPCNT (1 << 23)
|
---|
403 | #define CPUID_EXT_XSAVE (1 << 26)
|
---|
404 | #define CPUID_EXT_OSXSAVE (1 << 27)
|
---|
405 | #define CPUID_EXT_HYPERVISOR (1 << 31)
|
---|
406 |
|
---|
407 | #define CPUID_EXT2_SYSCALL (1 << 11)
|
---|
408 | #define CPUID_EXT2_MP (1 << 19)
|
---|
409 | #define CPUID_EXT2_NX (1 << 20)
|
---|
410 | #define CPUID_EXT2_MMXEXT (1 << 22)
|
---|
411 | #define CPUID_EXT2_FFXSR (1 << 25)
|
---|
412 | #define CPUID_EXT2_PDPE1GB (1 << 26)
|
---|
413 | #define CPUID_EXT2_RDTSCP (1 << 27)
|
---|
414 | #define CPUID_EXT2_LM (1 << 29)
|
---|
415 | #define CPUID_EXT2_3DNOWEXT (1 << 30)
|
---|
416 | #define CPUID_EXT2_3DNOW (1 << 31)
|
---|
417 |
|
---|
418 | #define CPUID_EXT3_LAHF_LM (1 << 0)
|
---|
419 | #define CPUID_EXT3_CMP_LEG (1 << 1)
|
---|
420 | #define CPUID_EXT3_SVM (1 << 2)
|
---|
421 | #define CPUID_EXT3_EXTAPIC (1 << 3)
|
---|
422 | #define CPUID_EXT3_CR8LEG (1 << 4)
|
---|
423 | #define CPUID_EXT3_ABM (1 << 5)
|
---|
424 | #define CPUID_EXT3_SSE4A (1 << 6)
|
---|
425 | #define CPUID_EXT3_MISALIGNSSE (1 << 7)
|
---|
426 | #define CPUID_EXT3_3DNOWPREFETCH (1 << 8)
|
---|
427 | #define CPUID_EXT3_OSVW (1 << 9)
|
---|
428 | #define CPUID_EXT3_IBS (1 << 10)
|
---|
429 | #define CPUID_EXT3_SKINIT (1 << 12)
|
---|
430 |
|
---|
431 | #define CPUID_VENDOR_INTEL_1 0x756e6547 /* "Genu" */
|
---|
432 | #define CPUID_VENDOR_INTEL_2 0x49656e69 /* "ineI" */
|
---|
433 | #define CPUID_VENDOR_INTEL_3 0x6c65746e /* "ntel" */
|
---|
434 |
|
---|
435 | #define CPUID_VENDOR_AMD_1 0x68747541 /* "Auth" */
|
---|
436 | #define CPUID_VENDOR_AMD_2 0x69746e65 /* "enti" */
|
---|
437 | #define CPUID_VENDOR_AMD_3 0x444d4163 /* "cAMD" */
|
---|
438 |
|
---|
439 | #define CPUID_MWAIT_IBE (1 << 1) /* Interrupts can exit capability */
|
---|
440 | #define CPUID_MWAIT_EMX (1 << 0) /* enumeration supported */
|
---|
441 |
|
---|
442 | #define EXCP00_DIVZ 0
|
---|
443 | #define EXCP01_DB 1
|
---|
444 | #define EXCP02_NMI 2
|
---|
445 | #define EXCP03_INT3 3
|
---|
446 | #define EXCP04_INTO 4
|
---|
447 | #define EXCP05_BOUND 5
|
---|
448 | #define EXCP06_ILLOP 6
|
---|
449 | #define EXCP07_PREX 7
|
---|
450 | #define EXCP08_DBLE 8
|
---|
451 | #define EXCP09_XERR 9
|
---|
452 | #define EXCP0A_TSS 10
|
---|
453 | #define EXCP0B_NOSEG 11
|
---|
454 | #define EXCP0C_STACK 12
|
---|
455 | #define EXCP0D_GPF 13
|
---|
456 | #define EXCP0E_PAGE 14
|
---|
457 | #define EXCP10_COPR 16
|
---|
458 | #define EXCP11_ALGN 17
|
---|
459 | #define EXCP12_MCHK 18
|
---|
460 |
|
---|
461 | #define EXCP_SYSCALL 0x100 /* only happens in user only emulation
|
---|
462 | for syscall instruction */
|
---|
463 |
|
---|
464 | enum {
|
---|
465 | CC_OP_DYNAMIC, /* must use dynamic code to get cc_op */
|
---|
466 | CC_OP_EFLAGS, /* all cc are explicitly computed, CC_SRC = flags */
|
---|
467 |
|
---|
468 | CC_OP_MULB, /* modify all flags, C, O = (CC_SRC != 0) */
|
---|
469 | CC_OP_MULW,
|
---|
470 | CC_OP_MULL,
|
---|
471 | CC_OP_MULQ,
|
---|
472 |
|
---|
473 | CC_OP_ADDB, /* modify all flags, CC_DST = res, CC_SRC = src1 */
|
---|
474 | CC_OP_ADDW,
|
---|
475 | CC_OP_ADDL,
|
---|
476 | CC_OP_ADDQ,
|
---|
477 |
|
---|
478 | CC_OP_ADCB, /* modify all flags, CC_DST = res, CC_SRC = src1 */
|
---|
479 | CC_OP_ADCW,
|
---|
480 | CC_OP_ADCL,
|
---|
481 | CC_OP_ADCQ,
|
---|
482 |
|
---|
483 | CC_OP_SUBB, /* modify all flags, CC_DST = res, CC_SRC = src1 */
|
---|
484 | CC_OP_SUBW,
|
---|
485 | CC_OP_SUBL,
|
---|
486 | CC_OP_SUBQ,
|
---|
487 |
|
---|
488 | CC_OP_SBBB, /* modify all flags, CC_DST = res, CC_SRC = src1 */
|
---|
489 | CC_OP_SBBW,
|
---|
490 | CC_OP_SBBL,
|
---|
491 | CC_OP_SBBQ,
|
---|
492 |
|
---|
493 | CC_OP_LOGICB, /* modify all flags, CC_DST = res */
|
---|
494 | CC_OP_LOGICW,
|
---|
495 | CC_OP_LOGICL,
|
---|
496 | CC_OP_LOGICQ,
|
---|
497 |
|
---|
498 | CC_OP_INCB, /* modify all flags except, CC_DST = res, CC_SRC = C */
|
---|
499 | CC_OP_INCW,
|
---|
500 | CC_OP_INCL,
|
---|
501 | CC_OP_INCQ,
|
---|
502 |
|
---|
503 | CC_OP_DECB, /* modify all flags except, CC_DST = res, CC_SRC = C */
|
---|
504 | CC_OP_DECW,
|
---|
505 | CC_OP_DECL,
|
---|
506 | CC_OP_DECQ,
|
---|
507 |
|
---|
508 | CC_OP_SHLB, /* modify all flags, CC_DST = res, CC_SRC.msb = C */
|
---|
509 | CC_OP_SHLW,
|
---|
510 | CC_OP_SHLL,
|
---|
511 | CC_OP_SHLQ,
|
---|
512 |
|
---|
513 | CC_OP_SARB, /* modify all flags, CC_DST = res, CC_SRC.lsb = C */
|
---|
514 | CC_OP_SARW,
|
---|
515 | CC_OP_SARL,
|
---|
516 | CC_OP_SARQ,
|
---|
517 |
|
---|
518 | CC_OP_NB,
|
---|
519 | };
|
---|
520 |
|
---|
521 | #ifdef FLOATX80
|
---|
522 | #define USE_X86LDOUBLE
|
---|
523 | #endif
|
---|
524 |
|
---|
525 | #ifdef USE_X86LDOUBLE
|
---|
526 | typedef floatx80 CPU86_LDouble;
|
---|
527 | #else
|
---|
528 | typedef float64 CPU86_LDouble;
|
---|
529 | #endif
|
---|
530 |
|
---|
531 | typedef struct SegmentCache {
|
---|
532 | uint32_t selector;
|
---|
533 | target_ulong base;
|
---|
534 | uint32_t limit;
|
---|
535 | uint32_t flags;
|
---|
536 | #ifdef VBOX
|
---|
537 | /** The new selector is saved here when we are unable to sync it before invoking the recompiled code. */
|
---|
538 | uint32_t newselector;
|
---|
539 | #endif
|
---|
540 | } SegmentCache;
|
---|
541 |
|
---|
542 | typedef union {
|
---|
543 | uint8_t _b[16];
|
---|
544 | uint16_t _w[8];
|
---|
545 | uint32_t _l[4];
|
---|
546 | uint64_t _q[2];
|
---|
547 | float32 _s[4];
|
---|
548 | float64 _d[2];
|
---|
549 | } XMMReg;
|
---|
550 |
|
---|
551 | typedef union {
|
---|
552 | uint8_t _b[8];
|
---|
553 | uint16_t _w[4];
|
---|
554 | uint32_t _l[2];
|
---|
555 | float32 _s[2];
|
---|
556 | uint64_t q;
|
---|
557 | } MMXReg;
|
---|
558 |
|
---|
559 | #ifdef WORDS_BIGENDIAN
|
---|
560 | #define XMM_B(n) _b[15 - (n)]
|
---|
561 | #define XMM_W(n) _w[7 - (n)]
|
---|
562 | #define XMM_L(n) _l[3 - (n)]
|
---|
563 | #define XMM_S(n) _s[3 - (n)]
|
---|
564 | #define XMM_Q(n) _q[1 - (n)]
|
---|
565 | #define XMM_D(n) _d[1 - (n)]
|
---|
566 |
|
---|
567 | #define MMX_B(n) _b[7 - (n)]
|
---|
568 | #define MMX_W(n) _w[3 - (n)]
|
---|
569 | #define MMX_L(n) _l[1 - (n)]
|
---|
570 | #define MMX_S(n) _s[1 - (n)]
|
---|
571 | #else
|
---|
572 | #define XMM_B(n) _b[n]
|
---|
573 | #define XMM_W(n) _w[n]
|
---|
574 | #define XMM_L(n) _l[n]
|
---|
575 | #define XMM_S(n) _s[n]
|
---|
576 | #define XMM_Q(n) _q[n]
|
---|
577 | #define XMM_D(n) _d[n]
|
---|
578 |
|
---|
579 | #define MMX_B(n) _b[n]
|
---|
580 | #define MMX_W(n) _w[n]
|
---|
581 | #define MMX_L(n) _l[n]
|
---|
582 | #define MMX_S(n) _s[n]
|
---|
583 | #endif
|
---|
584 | #define MMX_Q(n) q
|
---|
585 |
|
---|
586 | #ifdef TARGET_X86_64
|
---|
587 | #define CPU_NB_REGS 16
|
---|
588 | #else
|
---|
589 | #define CPU_NB_REGS 8
|
---|
590 | #endif
|
---|
591 |
|
---|
592 | #define NB_MMU_MODES 2
|
---|
593 |
|
---|
594 | typedef struct CPUX86State {
|
---|
595 | /* standard registers */
|
---|
596 | target_ulong regs[CPU_NB_REGS];
|
---|
597 | target_ulong eip;
|
---|
598 | target_ulong eflags; /* eflags register. During CPU emulation, CC
|
---|
599 | flags and DF are set to zero because they are
|
---|
600 | stored elsewhere */
|
---|
601 |
|
---|
602 | /* emulator internal eflags handling */
|
---|
603 | target_ulong cc_src;
|
---|
604 | target_ulong cc_dst;
|
---|
605 | uint32_t cc_op;
|
---|
606 | int32_t df; /* D flag : 1 if D = 0, -1 if D = 1 */
|
---|
607 | uint32_t hflags; /* TB flags, see HF_xxx constants. These flags
|
---|
608 | are known at translation time. */
|
---|
609 | uint32_t hflags2; /* various other flags, see HF2_xxx constants. */
|
---|
610 |
|
---|
611 | /* segments */
|
---|
612 | SegmentCache segs[6]; /* selector values */
|
---|
613 | SegmentCache ldt;
|
---|
614 | SegmentCache tr;
|
---|
615 | SegmentCache gdt; /* only base and limit are used */
|
---|
616 | SegmentCache idt; /* only base and limit are used */
|
---|
617 |
|
---|
618 | target_ulong cr[5]; /* NOTE: cr1 is unused */
|
---|
619 | uint64_t a20_mask;
|
---|
620 |
|
---|
621 | /* FPU state */
|
---|
622 | unsigned int fpstt; /* top of stack index */
|
---|
623 | unsigned int fpus;
|
---|
624 | unsigned int fpuc;
|
---|
625 | uint8_t fptags[8]; /* 0 = valid, 1 = empty */
|
---|
626 | union {
|
---|
627 | #ifdef USE_X86LDOUBLE
|
---|
628 | CPU86_LDouble d __attribute__((aligned(16)));
|
---|
629 | #else
|
---|
630 | CPU86_LDouble d;
|
---|
631 | #endif
|
---|
632 | MMXReg mmx;
|
---|
633 | } fpregs[8];
|
---|
634 |
|
---|
635 | /* emulator internal variables */
|
---|
636 | float_status fp_status;
|
---|
637 | #ifdef VBOX
|
---|
638 | uint32_t alignment3[3]; /* force the long double to start a 16 byte line. */
|
---|
639 | #endif
|
---|
640 | CPU86_LDouble ft0;
|
---|
641 | #if defined(VBOX) && defined(RT_ARCH_X86) && !defined(RT_OS_DARWIN)
|
---|
642 | uint32_t alignment4; /* long double is 12 byte, pad it to 16. */
|
---|
643 | #endif
|
---|
644 |
|
---|
645 | float_status mmx_status; /* for 3DNow! float ops */
|
---|
646 | float_status sse_status;
|
---|
647 | uint32_t mxcsr;
|
---|
648 | XMMReg xmm_regs[CPU_NB_REGS];
|
---|
649 | XMMReg xmm_t0;
|
---|
650 | MMXReg mmx_t0;
|
---|
651 | target_ulong cc_tmp; /* temporary for rcr/rcl */
|
---|
652 |
|
---|
653 | /* sysenter registers */
|
---|
654 | uint32_t sysenter_cs;
|
---|
655 | #ifdef VBOX
|
---|
656 | uint32_t alignment0;
|
---|
657 | #endif
|
---|
658 | target_ulong sysenter_esp;
|
---|
659 | target_ulong sysenter_eip;
|
---|
660 | uint64_t efer;
|
---|
661 | uint64_t star;
|
---|
662 |
|
---|
663 | uint64_t vm_hsave;
|
---|
664 | uint64_t vm_vmcb;
|
---|
665 | uint64_t tsc_offset;
|
---|
666 | uint64_t intercept;
|
---|
667 | uint16_t intercept_cr_read;
|
---|
668 | uint16_t intercept_cr_write;
|
---|
669 | uint16_t intercept_dr_read;
|
---|
670 | uint16_t intercept_dr_write;
|
---|
671 | uint32_t intercept_exceptions;
|
---|
672 | uint8_t v_tpr;
|
---|
673 |
|
---|
674 | #ifdef TARGET_X86_64
|
---|
675 | target_ulong lstar;
|
---|
676 | target_ulong cstar;
|
---|
677 | target_ulong fmask;
|
---|
678 | target_ulong kernelgsbase;
|
---|
679 | #endif
|
---|
680 |
|
---|
681 | uint64_t tsc;
|
---|
682 |
|
---|
683 | uint64_t pat;
|
---|
684 |
|
---|
685 | /* exception/interrupt handling */
|
---|
686 | int error_code;
|
---|
687 | int exception_is_int;
|
---|
688 | target_ulong exception_next_eip;
|
---|
689 | target_ulong dr[8]; /* debug registers */
|
---|
690 | union {
|
---|
691 | CPUBreakpoint *cpu_breakpoint[4];
|
---|
692 | CPUWatchpoint *cpu_watchpoint[4];
|
---|
693 | }; /* break/watchpoints for dr[0..3] */
|
---|
694 | uint32_t smbase;
|
---|
695 | int old_exception; /* exception in flight */
|
---|
696 |
|
---|
697 | CPU_COMMON
|
---|
698 |
|
---|
699 | #ifdef VBOX
|
---|
700 | /** cpu state flags. (see defines below) */
|
---|
701 | uint32_t state;
|
---|
702 | /** The VM handle. */
|
---|
703 | PVM pVM;
|
---|
704 | /** The VMCPU handle. */
|
---|
705 | PVMCPU pVCpu;
|
---|
706 | /** code buffer for instruction emulation */
|
---|
707 | void *pvCodeBuffer;
|
---|
708 | /** code buffer size */
|
---|
709 | uint32_t cbCodeBuffer;
|
---|
710 | #endif /* VBOX */
|
---|
711 |
|
---|
712 | /* processor features (e.g. for CPUID insn) */
|
---|
713 | #ifndef VBOX /* remR3CpuId deals with these */
|
---|
714 | uint32_t cpuid_level;
|
---|
715 | uint32_t cpuid_vendor1;
|
---|
716 | uint32_t cpuid_vendor2;
|
---|
717 | uint32_t cpuid_vendor3;
|
---|
718 | uint32_t cpuid_version;
|
---|
719 | #endif /* !VBOX */
|
---|
720 | uint32_t cpuid_features;
|
---|
721 | uint32_t cpuid_ext_features;
|
---|
722 | #ifndef VBOX
|
---|
723 | uint32_t cpuid_xlevel;
|
---|
724 | uint32_t cpuid_model[12];
|
---|
725 | #endif /* !VBOX */
|
---|
726 | uint32_t cpuid_ext2_features;
|
---|
727 | uint32_t cpuid_ext3_features;
|
---|
728 | uint32_t cpuid_apic_id;
|
---|
729 | #ifndef VBOX
|
---|
730 | int cpuid_vendor_override;
|
---|
731 |
|
---|
732 | /* MTRRs */
|
---|
733 | uint64_t mtrr_fixed[11];
|
---|
734 | uint64_t mtrr_deftype;
|
---|
735 | struct {
|
---|
736 | uint64_t base;
|
---|
737 | uint64_t mask;
|
---|
738 | } mtrr_var[8];
|
---|
739 |
|
---|
740 | #ifdef CONFIG_KQEMU
|
---|
741 | int kqemu_enabled;
|
---|
742 | int last_io_time;
|
---|
743 | #endif
|
---|
744 |
|
---|
745 | /* For KVM */
|
---|
746 | uint64_t interrupt_bitmap[256 / 64];
|
---|
747 | uint32_t mp_state;
|
---|
748 |
|
---|
749 | /* in order to simplify APIC support, we leave this pointer to the
|
---|
750 | user */
|
---|
751 | struct APICState *apic_state;
|
---|
752 |
|
---|
753 | uint64 mcg_cap;
|
---|
754 | uint64 mcg_status;
|
---|
755 | uint64 mcg_ctl;
|
---|
756 | uint64 *mce_banks;
|
---|
757 | #else /* VBOX */
|
---|
758 |
|
---|
759 | /* see 641 line to consult current alignments for darwin 32-bit host. */
|
---|
760 | # if HC_ARCH_BITS == 64 || (defined(RT_OS_DARWIN) && HC_ARCH_BITS == 32)
|
---|
761 | uint32_t alignment2[3];
|
---|
762 | # else
|
---|
763 | uint32_t alignmnt2[2];
|
---|
764 | # endif
|
---|
765 | /** Profiling tb_flush. */
|
---|
766 | STAMPROFILE StatTbFlush;
|
---|
767 | #endif /* VBOX */
|
---|
768 | } CPUX86State;
|
---|
769 |
|
---|
770 | #ifdef VBOX
|
---|
771 |
|
---|
772 | /* Version 1.6 structure; just for loading the old saved state */
|
---|
773 | typedef struct SegmentCache_Ver16 {
|
---|
774 | uint32_t selector;
|
---|
775 | uint32_t base;
|
---|
776 | uint32_t limit;
|
---|
777 | uint32_t flags;
|
---|
778 | /** The new selector is saved here when we are unable to sync it before invoking the recompiled code. */
|
---|
779 | uint32_t newselector;
|
---|
780 | } SegmentCache_Ver16;
|
---|
781 |
|
---|
782 | # define CPU_NB_REGS_VER16 8
|
---|
783 |
|
---|
784 | /* Version 1.6 structure; just for loading the old saved state */
|
---|
785 | typedef struct CPUX86State_Ver16 {
|
---|
786 | # if TARGET_LONG_BITS > HOST_LONG_BITS
|
---|
787 | /* temporaries if we cannot store them in host registers */
|
---|
788 | uint32_t t0, t1, t2;
|
---|
789 | # endif
|
---|
790 |
|
---|
791 | /* standard registers */
|
---|
792 | uint32_t regs[CPU_NB_REGS_VER16];
|
---|
793 | uint32_t eip;
|
---|
794 | uint32_t eflags; /* eflags register. During CPU emulation, CC
|
---|
795 | flags and DF are set to zero because they are
|
---|
796 | stored elsewhere */
|
---|
797 |
|
---|
798 | /* emulator internal eflags handling */
|
---|
799 | uint32_t cc_src;
|
---|
800 | uint32_t cc_dst;
|
---|
801 | uint32_t cc_op;
|
---|
802 | int32_t df; /* D flag : 1 if D = 0, -1 if D = 1 */
|
---|
803 | uint32_t hflags; /* hidden flags, see HF_xxx constants */
|
---|
804 |
|
---|
805 | /* segments */
|
---|
806 | SegmentCache_Ver16 segs[6]; /* selector values */
|
---|
807 | SegmentCache_Ver16 ldt;
|
---|
808 | SegmentCache_Ver16 tr;
|
---|
809 | SegmentCache_Ver16 gdt; /* only base and limit are used */
|
---|
810 | SegmentCache_Ver16 idt; /* only base and limit are used */
|
---|
811 |
|
---|
812 | uint32_t cr[5]; /* NOTE: cr1 is unused */
|
---|
813 | uint32_t a20_mask;
|
---|
814 |
|
---|
815 | /* FPU state */
|
---|
816 | unsigned int fpstt; /* top of stack index */
|
---|
817 | unsigned int fpus;
|
---|
818 | unsigned int fpuc;
|
---|
819 | uint8_t fptags[8]; /* 0 = valid, 1 = empty */
|
---|
820 | union {
|
---|
821 | # ifdef USE_X86LDOUBLE
|
---|
822 | CPU86_LDouble d __attribute__((aligned(16)));
|
---|
823 | # else
|
---|
824 | CPU86_LDouble d;
|
---|
825 | # endif
|
---|
826 | MMXReg mmx;
|
---|
827 | } fpregs[8];
|
---|
828 |
|
---|
829 | /* emulator internal variables */
|
---|
830 | float_status fp_status;
|
---|
831 | # ifdef VBOX
|
---|
832 | uint32_t alignment3[3]; /* force the long double to start a 16 byte line. */
|
---|
833 | # endif
|
---|
834 | CPU86_LDouble ft0;
|
---|
835 | # if defined(VBOX) && defined(RT_ARCH_X86) && !defined(RT_OS_DARWIN)
|
---|
836 | uint32_t alignment4; /* long double is 12 byte, pad it to 16. */
|
---|
837 | # endif
|
---|
838 | union {
|
---|
839 | float f;
|
---|
840 | double d;
|
---|
841 | int i32;
|
---|
842 | int64_t i64;
|
---|
843 | } fp_convert;
|
---|
844 |
|
---|
845 | float_status sse_status;
|
---|
846 | uint32_t mxcsr;
|
---|
847 | XMMReg xmm_regs[CPU_NB_REGS_VER16];
|
---|
848 | XMMReg xmm_t0;
|
---|
849 | MMXReg mmx_t0;
|
---|
850 |
|
---|
851 | /* sysenter registers */
|
---|
852 | uint32_t sysenter_cs;
|
---|
853 | uint32_t sysenter_esp;
|
---|
854 | uint32_t sysenter_eip;
|
---|
855 | # ifdef VBOX
|
---|
856 | uint32_t alignment0;
|
---|
857 | # endif
|
---|
858 | uint64_t efer;
|
---|
859 | uint64_t star;
|
---|
860 |
|
---|
861 | uint64_t pat;
|
---|
862 |
|
---|
863 | /* temporary data for USE_CODE_COPY mode */
|
---|
864 | # ifdef USE_CODE_COPY
|
---|
865 | uint32_t tmp0;
|
---|
866 | uint32_t saved_esp;
|
---|
867 | int native_fp_regs; /* if true, the FPU state is in the native CPU regs */
|
---|
868 | # endif
|
---|
869 |
|
---|
870 | /* exception/interrupt handling */
|
---|
871 | jmp_buf jmp_env;
|
---|
872 | } CPUX86State_Ver16;
|
---|
873 |
|
---|
874 | /** CPUX86State state flags
|
---|
875 | * @{ */
|
---|
876 | # define CPU_RAW_RING0 0x0002 /* Set after first time RawR0 is executed, never cleared. */
|
---|
877 | # define CPU_EMULATE_SINGLE_INSTR 0x0040 /* Execute a single instruction in emulation mode */
|
---|
878 | # define CPU_EMULATE_SINGLE_STEP 0x0080 /* go into single step mode */
|
---|
879 | # define CPU_RAW_HWACC 0x0100 /* Set after first time HWACC is executed, never cleared. */
|
---|
880 | /** @} */
|
---|
881 | #endif /* !VBOX */
|
---|
882 |
|
---|
883 | #ifdef VBOX
|
---|
884 | CPUX86State *cpu_x86_init(CPUX86State *env, const char *cpu_model);
|
---|
885 | #else /* !VBOX */
|
---|
886 | CPUX86State *cpu_x86_init(const char *cpu_model);
|
---|
887 | #endif /* !VBOX */
|
---|
888 | int cpu_x86_exec(CPUX86State *s);
|
---|
889 | void cpu_x86_close(CPUX86State *s);
|
---|
890 | void x86_cpu_list (FILE *f, int (*cpu_fprintf)(FILE *f, const char *fmt,
|
---|
891 | ...));
|
---|
892 | int cpu_get_pic_interrupt(CPUX86State *s);
|
---|
893 | /* MSDOS compatibility mode FPU exception support */
|
---|
894 | void cpu_set_ferr(CPUX86State *s);
|
---|
895 |
|
---|
896 | /* this function must always be used to load data in the segment
|
---|
897 | cache: it synchronizes the hflags with the segment cache values */
|
---|
898 | static inline void cpu_x86_load_seg_cache(CPUX86State *env,
|
---|
899 | int seg_reg, unsigned int selector,
|
---|
900 | target_ulong base,
|
---|
901 | unsigned int limit,
|
---|
902 | unsigned int flags)
|
---|
903 | {
|
---|
904 | SegmentCache *sc;
|
---|
905 | unsigned int new_hflags;
|
---|
906 |
|
---|
907 | sc = &env->segs[seg_reg];
|
---|
908 | sc->selector = selector;
|
---|
909 | sc->base = base;
|
---|
910 | sc->limit = limit;
|
---|
911 | sc->flags = flags;
|
---|
912 | #ifdef VBOX
|
---|
913 | sc->newselector = 0;
|
---|
914 | #endif
|
---|
915 |
|
---|
916 | /* update the hidden flags */
|
---|
917 | {
|
---|
918 | if (seg_reg == R_CS) {
|
---|
919 | #ifdef TARGET_X86_64
|
---|
920 | if ((env->hflags & HF_LMA_MASK) && (flags & DESC_L_MASK)) {
|
---|
921 | /* long mode */
|
---|
922 | env->hflags |= HF_CS32_MASK | HF_SS32_MASK | HF_CS64_MASK;
|
---|
923 | env->hflags &= ~(HF_ADDSEG_MASK);
|
---|
924 | } else
|
---|
925 | #endif
|
---|
926 | {
|
---|
927 | /* legacy / compatibility case */
|
---|
928 | new_hflags = (env->segs[R_CS].flags & DESC_B_MASK)
|
---|
929 | >> (DESC_B_SHIFT - HF_CS32_SHIFT);
|
---|
930 | env->hflags = (env->hflags & ~(HF_CS32_MASK | HF_CS64_MASK)) |
|
---|
931 | new_hflags;
|
---|
932 | }
|
---|
933 | }
|
---|
934 | new_hflags = (env->segs[R_SS].flags & DESC_B_MASK)
|
---|
935 | >> (DESC_B_SHIFT - HF_SS32_SHIFT);
|
---|
936 | if (env->hflags & HF_CS64_MASK) {
|
---|
937 | /* zero base assumed for DS, ES and SS in long mode */
|
---|
938 | } else if (!(env->cr[0] & CR0_PE_MASK) ||
|
---|
939 | (env->eflags & VM_MASK) ||
|
---|
940 | !(env->hflags & HF_CS32_MASK)) {
|
---|
941 | /* XXX: try to avoid this test. The problem comes from the
|
---|
942 | fact that is real mode or vm86 mode we only modify the
|
---|
943 | 'base' and 'selector' fields of the segment cache to go
|
---|
944 | faster. A solution may be to force addseg to one in
|
---|
945 | translate-i386.c. */
|
---|
946 | new_hflags |= HF_ADDSEG_MASK;
|
---|
947 | } else {
|
---|
948 | new_hflags |= ((env->segs[R_DS].base |
|
---|
949 | env->segs[R_ES].base |
|
---|
950 | env->segs[R_SS].base) != 0) <<
|
---|
951 | HF_ADDSEG_SHIFT;
|
---|
952 | }
|
---|
953 | env->hflags = (env->hflags &
|
---|
954 | ~(HF_SS32_MASK | HF_ADDSEG_MASK)) | new_hflags;
|
---|
955 | }
|
---|
956 | }
|
---|
957 |
|
---|
958 | /* wrapper, just in case memory mappings must be changed */
|
---|
959 | static inline void cpu_x86_set_cpl(CPUX86State *s, int cpl)
|
---|
960 | {
|
---|
961 | #if HF_CPL_MASK == 3
|
---|
962 | s->hflags = (s->hflags & ~HF_CPL_MASK) | cpl;
|
---|
963 | #else
|
---|
964 | #error HF_CPL_MASK is hardcoded
|
---|
965 | #endif
|
---|
966 | }
|
---|
967 |
|
---|
968 | /* op_helper.c */
|
---|
969 | /* used for debug or cpu save/restore */
|
---|
970 | void cpu_get_fp80(uint64_t *pmant, uint16_t *pexp, CPU86_LDouble f);
|
---|
971 | CPU86_LDouble cpu_set_fp80(uint64_t mant, uint16_t upper);
|
---|
972 |
|
---|
973 | /* cpu-exec.c */
|
---|
974 | /* the following helpers are only usable in user mode simulation as
|
---|
975 | they can trigger unexpected exceptions */
|
---|
976 | void cpu_x86_load_seg(CPUX86State *s, int seg_reg, int selector);
|
---|
977 | void cpu_x86_fsave(CPUX86State *s, target_ulong ptr, int data32);
|
---|
978 | void cpu_x86_frstor(CPUX86State *s, target_ulong ptr, int data32);
|
---|
979 |
|
---|
980 | /* you can call this signal handler from your SIGBUS and SIGSEGV
|
---|
981 | signal handlers to inform the virtual CPU of exceptions. non zero
|
---|
982 | is returned if the signal was handled by the virtual CPU. */
|
---|
983 | int cpu_x86_signal_handler(int host_signum, void *pinfo,
|
---|
984 | void *puc);
|
---|
985 |
|
---|
986 | /* helper.c */
|
---|
987 | int cpu_x86_handle_mmu_fault(CPUX86State *env, target_ulong addr,
|
---|
988 | int is_write, int mmu_idx, int is_softmmu);
|
---|
989 | void cpu_x86_set_a20(CPUX86State *env, int a20_state);
|
---|
990 | void cpu_x86_cpuid(CPUX86State *env, uint32_t index, uint32_t count,
|
---|
991 | uint32_t *eax, uint32_t *ebx,
|
---|
992 | uint32_t *ecx, uint32_t *edx);
|
---|
993 |
|
---|
994 | static inline int hw_breakpoint_enabled(unsigned long dr7, int index)
|
---|
995 | {
|
---|
996 | return (dr7 >> (index * 2)) & 3;
|
---|
997 | }
|
---|
998 |
|
---|
999 | static inline int hw_breakpoint_type(unsigned long dr7, int index)
|
---|
1000 | {
|
---|
1001 | return (dr7 >> (DR7_TYPE_SHIFT + (index * 2))) & 3;
|
---|
1002 | }
|
---|
1003 |
|
---|
1004 | static inline int hw_breakpoint_len(unsigned long dr7, int index)
|
---|
1005 | {
|
---|
1006 | int len = ((dr7 >> (DR7_LEN_SHIFT + (index * 2))) & 3);
|
---|
1007 | return (len == 2) ? 8 : len + 1;
|
---|
1008 | }
|
---|
1009 |
|
---|
1010 | void hw_breakpoint_insert(CPUX86State *env, int index);
|
---|
1011 | void hw_breakpoint_remove(CPUX86State *env, int index);
|
---|
1012 | int check_hw_breakpoints(CPUX86State *env, int force_dr6_update);
|
---|
1013 |
|
---|
1014 | /* will be suppressed */
|
---|
1015 | void cpu_x86_update_cr0(CPUX86State *env, uint32_t new_cr0);
|
---|
1016 | void cpu_x86_update_cr3(CPUX86State *env, target_ulong new_cr3);
|
---|
1017 | void cpu_x86_update_cr4(CPUX86State *env, uint32_t new_cr4);
|
---|
1018 |
|
---|
1019 | /* hw/apic.c */
|
---|
1020 | void cpu_set_apic_base(CPUX86State *env, uint64_t val);
|
---|
1021 | uint64_t cpu_get_apic_base(CPUX86State *env);
|
---|
1022 | void cpu_set_apic_tpr(CPUX86State *env, uint8_t val);
|
---|
1023 | #ifndef NO_CPU_IO_DEFS
|
---|
1024 | uint8_t cpu_get_apic_tpr(CPUX86State *env);
|
---|
1025 | #endif
|
---|
1026 |
|
---|
1027 | /* hw/pc.c */
|
---|
1028 | void cpu_smm_update(CPUX86State *env);
|
---|
1029 | uint64_t cpu_get_tsc(CPUX86State *env);
|
---|
1030 |
|
---|
1031 | /* used to debug */
|
---|
1032 | #define X86_DUMP_FPU 0x0001 /* dump FPU state too */
|
---|
1033 | #define X86_DUMP_CCOP 0x0002 /* dump qemu flag cache */
|
---|
1034 |
|
---|
1035 | #ifdef CONFIG_KQEMU
|
---|
1036 | static inline int cpu_get_time_fast(void)
|
---|
1037 | {
|
---|
1038 | int low, high;
|
---|
1039 | asm volatile("rdtsc" : "=a" (low), "=d" (high));
|
---|
1040 | return low;
|
---|
1041 | }
|
---|
1042 | #endif
|
---|
1043 |
|
---|
1044 | #ifdef VBOX
|
---|
1045 | int cpu_rdmsr(CPUX86State *env, uint32_t idMsr, uint64_t *puValue);
|
---|
1046 | int cpu_wrmsr(CPUX86State *env, uint32_t idMsr, uint64_t uValue);
|
---|
1047 | void cpu_trap_raw(CPUX86State *env1);
|
---|
1048 |
|
---|
1049 | /* in helper.c */
|
---|
1050 | uint8_t read_byte(CPUX86State *env1, target_ulong addr);
|
---|
1051 | uint16_t read_word(CPUX86State *env1, target_ulong addr);
|
---|
1052 | void write_byte(CPUX86State *env1, target_ulong addr, uint8_t val);
|
---|
1053 | uint32_t read_dword(CPUX86State *env1, target_ulong addr);
|
---|
1054 | void write_word(CPUX86State *env1, target_ulong addr, uint16_t val);
|
---|
1055 | void write_dword(CPUX86State *env1, target_ulong addr, uint32_t val);
|
---|
1056 | /* in helper.c */
|
---|
1057 | int emulate_single_instr(CPUX86State *env1);
|
---|
1058 | int get_ss_esp_from_tss_raw(CPUX86State *env1, uint32_t *ss_ptr, uint32_t *esp_ptr, int dpl);
|
---|
1059 |
|
---|
1060 | void restore_raw_fp_state(CPUX86State *env, uint8_t *ptr);
|
---|
1061 | void save_raw_fp_state(CPUX86State *env, uint8_t *ptr);
|
---|
1062 | #endif /* VBOX */
|
---|
1063 |
|
---|
1064 | #define TARGET_PAGE_BITS 12
|
---|
1065 |
|
---|
1066 | #define cpu_init cpu_x86_init
|
---|
1067 | #define cpu_exec cpu_x86_exec
|
---|
1068 | #define cpu_gen_code cpu_x86_gen_code
|
---|
1069 | #define cpu_signal_handler cpu_x86_signal_handler
|
---|
1070 | #define cpu_list x86_cpu_list
|
---|
1071 |
|
---|
1072 | #define CPU_SAVE_VERSION 10
|
---|
1073 |
|
---|
1074 | /* MMU modes definitions */
|
---|
1075 | #define MMU_MODE0_SUFFIX _kernel
|
---|
1076 | #define MMU_MODE1_SUFFIX _user
|
---|
1077 | #define MMU_USER_IDX 1
|
---|
1078 | static inline int cpu_mmu_index (CPUState *env)
|
---|
1079 | {
|
---|
1080 | return (env->hflags & HF_CPL_MASK) == 3 ? 1 : 0;
|
---|
1081 | }
|
---|
1082 |
|
---|
1083 | /* translate.c */
|
---|
1084 | void optimize_flags_init(void);
|
---|
1085 |
|
---|
1086 | typedef struct CCTable {
|
---|
1087 | int (*compute_all)(void); /* return all the flags */
|
---|
1088 | int (*compute_c)(void); /* return the C flag */
|
---|
1089 | } CCTable;
|
---|
1090 |
|
---|
1091 | #if defined(CONFIG_USER_ONLY)
|
---|
1092 | static inline void cpu_clone_regs(CPUState *env, target_ulong newsp)
|
---|
1093 | {
|
---|
1094 | if (newsp)
|
---|
1095 | env->regs[R_ESP] = newsp;
|
---|
1096 | env->regs[R_EAX] = 0;
|
---|
1097 | }
|
---|
1098 | #endif
|
---|
1099 |
|
---|
1100 | #include "cpu-all.h"
|
---|
1101 | #include "exec-all.h"
|
---|
1102 |
|
---|
1103 | #include "svm.h"
|
---|
1104 |
|
---|
1105 | static inline void cpu_pc_from_tb(CPUState *env, TranslationBlock *tb)
|
---|
1106 | {
|
---|
1107 | env->eip = tb->pc - tb->cs_base;
|
---|
1108 | }
|
---|
1109 |
|
---|
1110 | static inline void cpu_get_tb_cpu_state(CPUState *env, target_ulong *pc,
|
---|
1111 | target_ulong *cs_base, int *flags)
|
---|
1112 | {
|
---|
1113 | *cs_base = env->segs[R_CS].base;
|
---|
1114 | *pc = *cs_base + env->eip;
|
---|
1115 | *flags = env->hflags |
|
---|
1116 | (env->eflags & (IOPL_MASK | TF_MASK | RF_MASK | VM_MASK));
|
---|
1117 | }
|
---|
1118 |
|
---|
1119 | #ifndef VBOX
|
---|
1120 | void apic_init_reset(CPUState *env);
|
---|
1121 | void apic_sipi(CPUState *env);
|
---|
1122 | void do_cpu_init(CPUState *env);
|
---|
1123 | void do_cpu_sipi(CPUState *env);
|
---|
1124 | #endif /* !VBOX */
|
---|
1125 | #endif /* CPU_I386_H */
|
---|