1
2/*---------------------------------------------------------------*/
3/*--- begin                                guest_amd64_defs.h ---*/
4/*---------------------------------------------------------------*/
5
6/*
7   This file is part of Valgrind, a dynamic binary instrumentation
8   framework.
9
10   Copyright (C) 2004-2011 OpenWorks LLP
11      info@open-works.net
12
13   This program is free software; you can redistribute it and/or
14   modify it under the terms of the GNU General Public License as
15   published by the Free Software Foundation; either version 2 of the
16   License, or (at your option) any later version.
17
18   This program is distributed in the hope that it will be useful, but
19   WITHOUT ANY WARRANTY; without even the implied warranty of
20   MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
21   General Public License for more details.
22
23   You should have received a copy of the GNU General Public License
24   along with this program; if not, write to the Free Software
25   Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA
26   02110-1301, USA.
27
28   The GNU General Public License is contained in the file COPYING.
29
30   Neither the names of the U.S. Department of Energy nor the
31   University of California nor the names of its contributors may be
32   used to endorse or promote products derived from this software
33   without prior written permission.
34*/
35
36/* Only to be used within the guest-amd64 directory. */
37
38#ifndef __VEX_GUEST_AMD64_DEFS_H
39#define __VEX_GUEST_AMD64_DEFS_H
40
41
42/*---------------------------------------------------------*/
43/*--- amd64 to IR conversion                            ---*/
44/*---------------------------------------------------------*/
45
46/* Convert one amd64 insn to IR.  See the type DisOneInstrFn in
47   bb_to_IR.h. */
48extern
49DisResult disInstr_AMD64 ( IRSB*        irbb,
50                           Bool         put_IP,
51                           Bool         (*resteerOkFn) ( void*, Addr64 ),
52                           Bool         resteerCisOk,
53                           void*        callback_opaque,
54                           UChar*       guest_code,
55                           Long         delta,
56                           Addr64       guest_IP,
57                           VexArch      guest_arch,
58                           VexArchInfo* archinfo,
59                           VexAbiInfo*  abiinfo,
60                           Bool         host_bigendian );
61
62/* Used by the optimiser to specialise calls to helpers. */
63extern
64IRExpr* guest_amd64_spechelper ( HChar*   function_name,
65                                 IRExpr** args,
66                                 IRStmt** precedingStmts,
67                                 Int      n_precedingStmts );
68
69/* Describes to the optimiser which part of the guest state require
70   precise memory exceptions.  This is logically part of the guest
71   state description. */
72extern
73Bool guest_amd64_state_requires_precise_mem_exns ( Int, Int );
74
75extern
76VexGuestLayout amd64guest_layout;
77
78
79/*---------------------------------------------------------*/
80/*--- amd64 guest helpers                               ---*/
81/*---------------------------------------------------------*/
82
83/* --- CLEAN HELPERS --- */
84
85extern ULong amd64g_calculate_rflags_all (
86                ULong cc_op,
87                ULong cc_dep1, ULong cc_dep2, ULong cc_ndep
88             );
89
90extern ULong amd64g_calculate_rflags_c (
91                ULong cc_op,
92                ULong cc_dep1, ULong cc_dep2, ULong cc_ndep
93             );
94
95extern ULong amd64g_calculate_condition (
96                ULong/*AMD64Condcode*/ cond,
97                ULong cc_op,
98                ULong cc_dep1, ULong cc_dep2, ULong cc_ndep
99             );
100
101extern ULong amd64g_calculate_FXAM ( ULong tag, ULong dbl );
102
103extern ULong amd64g_calculate_RCR  (
104                ULong arg, ULong rot_amt, ULong rflags_in, Long sz
105             );
106
107extern ULong amd64g_calculate_RCL  (
108                ULong arg, ULong rot_amt, ULong rflags_in, Long sz
109             );
110
111extern ULong amd64g_calculate_pclmul(ULong s1, ULong s2, ULong which);
112
113extern ULong amd64g_check_fldcw ( ULong fpucw );
114
115extern ULong amd64g_create_fpucw ( ULong fpround );
116
117extern ULong amd64g_check_ldmxcsr ( ULong mxcsr );
118
119extern ULong amd64g_create_mxcsr ( ULong sseround );
120
121extern VexEmWarn amd64g_dirtyhelper_FLDENV ( VexGuestAMD64State*, HWord );
122
123extern void amd64g_dirtyhelper_FSTENV ( VexGuestAMD64State*, HWord );
124
125/* Translate a guest virtual_addr into a guest linear address by
126   consulting the supplied LDT/GDT structures.  Their representation
127   must be as specified in pub/libvex_guest_amd64.h.  To indicate a
128   translation failure, 1<<32 is returned.  On success, the lower 32
129   bits of the returned result indicate the linear address.
130*/
131//extern
132//ULong amd64g_use_seg_selector ( HWord ldt, HWord gdt,
133//                              UInt seg_selector, UInt virtual_addr );
134
135extern ULong amd64g_calculate_mmx_pmaddwd  ( ULong, ULong );
136extern ULong amd64g_calculate_mmx_psadbw   ( ULong, ULong );
137extern ULong amd64g_calculate_mmx_pmovmskb ( ULong );
138extern ULong amd64g_calculate_sse_pmovmskb ( ULong w64hi, ULong w64lo );
139
140extern ULong amd64g_calc_crc32b ( ULong crcIn, ULong b );
141extern ULong amd64g_calc_crc32w ( ULong crcIn, ULong w );
142extern ULong amd64g_calc_crc32l ( ULong crcIn, ULong l );
143extern ULong amd64g_calc_crc32q ( ULong crcIn, ULong q );
144
145/* --- DIRTY HELPERS --- */
146
147extern ULong amd64g_dirtyhelper_loadF80le  ( ULong/*addr*/ );
148
149extern void  amd64g_dirtyhelper_storeF80le ( ULong/*addr*/, ULong/*data*/ );
150
151extern void  amd64g_dirtyhelper_CPUID_baseline ( VexGuestAMD64State* st );
152extern void  amd64g_dirtyhelper_CPUID_sse3_and_cx16 ( VexGuestAMD64State* st );
153extern void  amd64g_dirtyhelper_CPUID_sse42_and_cx16 ( VexGuestAMD64State* st );
154
155extern void  amd64g_dirtyhelper_FINIT ( VexGuestAMD64State* );
156
157extern void      amd64g_dirtyhelper_FXSAVE  ( VexGuestAMD64State*, HWord );
158extern VexEmWarn amd64g_dirtyhelper_FXRSTOR ( VexGuestAMD64State*, HWord );
159
160extern ULong amd64g_dirtyhelper_RDTSC ( void );
161
162extern ULong amd64g_dirtyhelper_IN  ( ULong portno, ULong sz/*1,2 or 4*/ );
163extern void  amd64g_dirtyhelper_OUT ( ULong portno, ULong data,
164                                      ULong sz/*1,2 or 4*/ );
165
166extern void amd64g_dirtyhelper_SxDT ( void* address,
167                                      ULong op /* 0 or 1 */ );
168
169/* Helps with PCMP{I,E}STR{I,M}.
170
171   CALLED FROM GENERATED CODE: DIRTY HELPER(s).  (But not really,
172   actually it could be a clean helper, but for the fact that we can't
173   pass by value 2 x V128 to a clean helper, nor have one returned.)
174   Reads guest state, writes to guest state for the xSTRM cases, no
175   accesses of memory, is a pure function.
176
177   opc_and_imm contains (4th byte of opcode << 8) | the-imm8-byte so
178   the callee knows which I/E and I/M variant it is dealing with and
179   what the specific operation is.  4th byte of opcode is in the range
180   0x60 to 0x63:
181       istri  66 0F 3A 63
182       istrm  66 0F 3A 62
183       estri  66 0F 3A 61
184       estrm  66 0F 3A 60
185
186   gstOffL and gstOffR are the guest state offsets for the two XMM
187   register inputs.  We never have to deal with the memory case since
188   that is handled by pre-loading the relevant value into the fake
189   XMM16 register.
190
191   For ESTRx variants, edxIN and eaxIN hold the values of those two
192   registers.
193
194   In all cases, the bottom 16 bits of the result contain the new
195   OSZACP %rflags values.  For xSTRI variants, bits[31:16] of the
196   result hold the new %ecx value.  For xSTRM variants, the helper
197   writes the result directly to the guest XMM0.
198
199   Declarable side effects: in all cases, reads guest state at
200   [gstOffL, +16) and [gstOffR, +16).  For xSTRM variants, also writes
201   guest_XMM0.
202
203   Is expected to be called with opc_and_imm combinations which have
204   actually been validated, and will assert if otherwise.  The front
205   end should ensure we're only called with verified values.
206*/
207extern ULong amd64g_dirtyhelper_PCMPxSTRx (
208          VexGuestAMD64State*,
209          HWord opc4_and_imm,
210          HWord gstOffL, HWord gstOffR,
211          HWord edxIN, HWord eaxIN
212       );
213
214
215//extern void  amd64g_dirtyhelper_CPUID_sse0 ( VexGuestAMD64State* );
216//extern void  amd64g_dirtyhelper_CPUID_sse1 ( VexGuestAMD64State* );
217//extern void  amd64g_dirtyhelper_CPUID_sse2 ( VexGuestAMD64State* );
218
219//extern void  amd64g_dirtyhelper_FSAVE ( VexGuestAMD64State*, HWord );
220
221//extern VexEmWarn
222//            amd64g_dirtyhelper_FRSTOR ( VexGuestAMD64State*, HWord );
223
224//extern void amd64g_dirtyhelper_FSTENV ( VexGuestAMD64State*, HWord );
225
226//extern VexEmWarn
227//            amd64g_dirtyhelper_FLDENV ( VexGuestAMD64State*, HWord );
228
229
230
231/*---------------------------------------------------------*/
232/*--- Condition code stuff                              ---*/
233/*---------------------------------------------------------*/
234
235/* rflags masks */
236#define AMD64G_CC_SHIFT_O   11
237#define AMD64G_CC_SHIFT_S   7
238#define AMD64G_CC_SHIFT_Z   6
239#define AMD64G_CC_SHIFT_A   4
240#define AMD64G_CC_SHIFT_C   0
241#define AMD64G_CC_SHIFT_P   2
242
243#define AMD64G_CC_MASK_O    (1ULL << AMD64G_CC_SHIFT_O)
244#define AMD64G_CC_MASK_S    (1ULL << AMD64G_CC_SHIFT_S)
245#define AMD64G_CC_MASK_Z    (1ULL << AMD64G_CC_SHIFT_Z)
246#define AMD64G_CC_MASK_A    (1ULL << AMD64G_CC_SHIFT_A)
247#define AMD64G_CC_MASK_C    (1ULL << AMD64G_CC_SHIFT_C)
248#define AMD64G_CC_MASK_P    (1ULL << AMD64G_CC_SHIFT_P)
249
250/* FPU flag masks */
251#define AMD64G_FC_SHIFT_C3   14
252#define AMD64G_FC_SHIFT_C2   10
253#define AMD64G_FC_SHIFT_C1   9
254#define AMD64G_FC_SHIFT_C0   8
255
256#define AMD64G_FC_MASK_C3    (1ULL << AMD64G_FC_SHIFT_C3)
257#define AMD64G_FC_MASK_C2    (1ULL << AMD64G_FC_SHIFT_C2)
258#define AMD64G_FC_MASK_C1    (1ULL << AMD64G_FC_SHIFT_C1)
259#define AMD64G_FC_MASK_C0    (1ULL << AMD64G_FC_SHIFT_C0)
260
261
262/* %RFLAGS thunk descriptors.  A four-word thunk is used to record
263   details of the most recent flag-setting operation, so the flags can
264   be computed later if needed.  It is possible to do this a little
265   more efficiently using a 3-word thunk, but that makes it impossible
266   to describe the flag data dependencies sufficiently accurately for
267   Memcheck.  Hence 4 words are used, with minimal loss of efficiency.
268
269   The four words are:
270
271      CC_OP, which describes the operation.
272
273      CC_DEP1 and CC_DEP2.  These are arguments to the operation.
274         We want Memcheck to believe that the resulting flags are
275         data-dependent on both CC_DEP1 and CC_DEP2, hence the
276         name DEP.
277
278      CC_NDEP.  This is a 3rd argument to the operation which is
279         sometimes needed.  We arrange things so that Memcheck does
280         not believe the resulting flags are data-dependent on CC_NDEP
281         ("not dependent").
282
283   To make Memcheck believe that (the definedness of) the encoded
284   flags depends only on (the definedness of) CC_DEP1 and CC_DEP2
285   requires two things:
286
287   (1) In the guest state layout info (amd64guest_layout), CC_OP and
288       CC_NDEP are marked as always defined.
289
290   (2) When passing the thunk components to an evaluation function
291       (calculate_condition, calculate_eflags, calculate_eflags_c) the
292       IRCallee's mcx_mask must be set so as to exclude from
293       consideration all passed args except CC_DEP1 and CC_DEP2.
294
295   Strictly speaking only (2) is necessary for correctness.  However,
296   (1) helps efficiency in that since (2) means we never ask about the
297   definedness of CC_OP or CC_NDEP, we may as well not even bother to
298   track their definedness.
299
300   When building the thunk, it is always necessary to write words into
301   CC_DEP1 and CC_DEP2, even if those args are not used given the
302   CC_OP field (eg, CC_DEP2 is not used if CC_OP is CC_LOGIC1/2/4).
303   This is important because otherwise Memcheck could give false
304   positives as it does not understand the relationship between the
305   CC_OP field and CC_DEP1 and CC_DEP2, and so believes that the
306   definedness of the stored flags always depends on both CC_DEP1 and
307   CC_DEP2.
308
309   However, it is only necessary to set CC_NDEP when the CC_OP value
310   requires it, because Memcheck ignores CC_NDEP, and the evaluation
311   functions do understand the CC_OP fields and will only examine
312   CC_NDEP for suitable values of CC_OP.
313
314   A summary of the field usages is:
315
316   Operation          DEP1               DEP2               NDEP
317   ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
318
319   add/sub/mul        first arg          second arg         unused
320
321   adc/sbb            first arg          (second arg)
322                                         XOR old_carry      old_carry
323
324   and/or/xor         result             zero               unused
325
326   inc/dec            result             zero               old_carry
327
328   shl/shr/sar        result             subshifted-        unused
329                                         result
330
331   rol/ror            result             zero               old_flags
332
333   copy               old_flags          zero               unused.
334
335
336   Therefore Memcheck will believe the following:
337
338   * add/sub/mul -- definedness of result flags depends on definedness
339     of both args.
340
341   * adc/sbb -- definedness of result flags depends on definedness of
342     both args and definedness of the old C flag.  Because only two
343     DEP fields are available, the old C flag is XOR'd into the second
344     arg so that Memcheck sees the data dependency on it.  That means
345     the NDEP field must contain a second copy of the old C flag
346     so that the evaluation functions can correctly recover the second
347     arg.
348
349   * and/or/xor are straightforward -- definedness of result flags
350     depends on definedness of result value.
351
352   * inc/dec -- definedness of result flags depends only on
353     definedness of result.  This isn't really true -- it also depends
354     on the old C flag.  However, we don't want Memcheck to see that,
355     and so the old C flag must be passed in NDEP and not in DEP2.
356     It's inconceivable that a compiler would generate code that puts
357     the C flag in an undefined state, then does an inc/dec, which
358     leaves C unchanged, and then makes a conditional jump/move based
359     on C.  So our fiction seems a good approximation.
360
361   * shl/shr/sar -- straightforward, again, definedness of result
362     flags depends on definedness of result value.  The subshifted
363     value (value shifted one less) is also needed, but its
364     definedness is the same as the definedness of the shifted value.
365
366   * rol/ror -- these only set O and C, and leave A Z C P alone.
367     However it seems prudent (as per inc/dec) to say the definedness
368     of all resulting flags depends on the definedness of the result,
369     hence the old flags must go in as NDEP and not DEP2.
370
371   * rcl/rcr are too difficult to do in-line, and so are done by a
372     helper function.  They are not part of this scheme.  The helper
373     function takes the value to be rotated, the rotate amount and the
374     old flags, and returns the new flags and the rotated value.
375     Since the helper's mcx_mask does not have any set bits, Memcheck
376     will lazily propagate undefinedness from any of the 3 args into
377     both results (flags and actual value).
378*/
379enum {
380    AMD64G_CC_OP_COPY=0,  /* DEP1 = current flags, DEP2 = 0, NDEP = unused */
381                          /* just copy DEP1 to output */
382
383    AMD64G_CC_OP_ADDB,    /* 1 */
384    AMD64G_CC_OP_ADDW,    /* 2 DEP1 = argL, DEP2 = argR, NDEP = unused */
385    AMD64G_CC_OP_ADDL,    /* 3 */
386    AMD64G_CC_OP_ADDQ,    /* 4 */
387
388    AMD64G_CC_OP_SUBB,    /* 5 */
389    AMD64G_CC_OP_SUBW,    /* 6 DEP1 = argL, DEP2 = argR, NDEP = unused */
390    AMD64G_CC_OP_SUBL,    /* 7 */
391    AMD64G_CC_OP_SUBQ,    /* 8 */
392
393    AMD64G_CC_OP_ADCB,    /* 9 */
394    AMD64G_CC_OP_ADCW,    /* 10 DEP1 = argL, DEP2 = argR ^ oldCarry, NDEP = oldCarry */
395    AMD64G_CC_OP_ADCL,    /* 11 */
396    AMD64G_CC_OP_ADCQ,    /* 12 */
397
398    AMD64G_CC_OP_SBBB,    /* 13 */
399    AMD64G_CC_OP_SBBW,    /* 14 DEP1 = argL, DEP2 = argR ^ oldCarry, NDEP = oldCarry */
400    AMD64G_CC_OP_SBBL,    /* 15 */
401    AMD64G_CC_OP_SBBQ,    /* 16 */
402
403    AMD64G_CC_OP_LOGICB,  /* 17 */
404    AMD64G_CC_OP_LOGICW,  /* 18 DEP1 = result, DEP2 = 0, NDEP = unused */
405    AMD64G_CC_OP_LOGICL,  /* 19 */
406    AMD64G_CC_OP_LOGICQ,  /* 20 */
407
408    AMD64G_CC_OP_INCB,    /* 21 */
409    AMD64G_CC_OP_INCW,    /* 22 DEP1 = result, DEP2 = 0, NDEP = oldCarry (0 or 1) */
410    AMD64G_CC_OP_INCL,    /* 23 */
411    AMD64G_CC_OP_INCQ,    /* 24 */
412
413    AMD64G_CC_OP_DECB,    /* 25 */
414    AMD64G_CC_OP_DECW,    /* 26 DEP1 = result, DEP2 = 0, NDEP = oldCarry (0 or 1) */
415    AMD64G_CC_OP_DECL,    /* 27 */
416    AMD64G_CC_OP_DECQ,    /* 28 */
417
418    AMD64G_CC_OP_SHLB,    /* 29 DEP1 = res, DEP2 = res', NDEP = unused */
419    AMD64G_CC_OP_SHLW,    /* 30 where res' is like res but shifted one bit less */
420    AMD64G_CC_OP_SHLL,    /* 31 */
421    AMD64G_CC_OP_SHLQ,    /* 32 */
422
423    AMD64G_CC_OP_SHRB,    /* 33 DEP1 = res, DEP2 = res', NDEP = unused */
424    AMD64G_CC_OP_SHRW,    /* 34 where res' is like res but shifted one bit less */
425    AMD64G_CC_OP_SHRL,    /* 35 */
426    AMD64G_CC_OP_SHRQ,    /* 36 */
427
428    AMD64G_CC_OP_ROLB,    /* 37 */
429    AMD64G_CC_OP_ROLW,    /* 38 DEP1 = res, DEP2 = 0, NDEP = old flags */
430    AMD64G_CC_OP_ROLL,    /* 39 */
431    AMD64G_CC_OP_ROLQ,    /* 40 */
432
433    AMD64G_CC_OP_RORB,    /* 41 */
434    AMD64G_CC_OP_RORW,    /* 42 DEP1 = res, DEP2 = 0, NDEP = old flags */
435    AMD64G_CC_OP_RORL,    /* 43 */
436    AMD64G_CC_OP_RORQ,    /* 44 */
437
438    AMD64G_CC_OP_UMULB,   /* 45 */
439    AMD64G_CC_OP_UMULW,   /* 46 DEP1 = argL, DEP2 = argR, NDEP = unused */
440    AMD64G_CC_OP_UMULL,   /* 47 */
441    AMD64G_CC_OP_UMULQ,   /* 48 */
442
443    AMD64G_CC_OP_SMULB,   /* 49 */
444    AMD64G_CC_OP_SMULW,   /* 50 DEP1 = argL, DEP2 = argR, NDEP = unused */
445    AMD64G_CC_OP_SMULL,   /* 51 */
446    AMD64G_CC_OP_SMULQ,   /* 52 */
447
448    AMD64G_CC_OP_NUMBER
449};
450
451typedef
452   enum {
453      AMD64CondO      = 0,  /* overflow           */
454      AMD64CondNO     = 1,  /* no overflow        */
455
456      AMD64CondB      = 2,  /* below              */
457      AMD64CondNB     = 3,  /* not below          */
458
459      AMD64CondZ      = 4,  /* zero               */
460      AMD64CondNZ     = 5,  /* not zero           */
461
462      AMD64CondBE     = 6,  /* below or equal     */
463      AMD64CondNBE    = 7,  /* not below or equal */
464
465      AMD64CondS      = 8,  /* negative           */
466      AMD64CondNS     = 9,  /* not negative       */
467
468      AMD64CondP      = 10, /* parity even        */
469      AMD64CondNP     = 11, /* not parity even    */
470
471      AMD64CondL      = 12, /* jump less          */
472      AMD64CondNL     = 13, /* not less           */
473
474      AMD64CondLE     = 14, /* less or equal      */
475      AMD64CondNLE    = 15, /* not less or equal  */
476
477      AMD64CondAlways = 16  /* HACK */
478   }
479   AMD64Condcode;
480
481#endif /* ndef __VEX_GUEST_AMD64_DEFS_H */
482
483/*---------------------------------------------------------------*/
484/*--- end                                  guest_amd64_defs.h ---*/
485/*---------------------------------------------------------------*/
486