code_generator.h revision 3416601a9e9be81bb7494864287fd3602d18ef13
1/*
2 * Copyright (C) 2014 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 *      http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17#ifndef ART_COMPILER_OPTIMIZING_CODE_GENERATOR_H_
18#define ART_COMPILER_OPTIMIZING_CODE_GENERATOR_H_
19
20#include "arch/instruction_set.h"
21#include "arch/instruction_set_features.h"
22#include "base/bit_field.h"
23#include "globals.h"
24#include "locations.h"
25#include "memory_region.h"
26#include "nodes.h"
27#include "stack_map_stream.h"
28
29namespace art {
30
31static size_t constexpr kVRegSize = 4;
32static size_t constexpr kUninitializedFrameSize = 0;
33
34// Binary encoding of 2^32 for type double.
35static int64_t constexpr k2Pow32EncodingForDouble = INT64_C(0x41F0000000000000);
36// Binary encoding of 2^31 for type double.
37static int64_t constexpr k2Pow31EncodingForDouble = INT64_C(0x41E0000000000000);
38
39// Maximum value for a primitive integer.
40static int32_t constexpr kPrimIntMax = 0x7fffffff;
41// Maximum value for a primitive long.
42static int64_t constexpr kPrimLongMax = 0x7fffffffffffffff;
43
44class Assembler;
45class CodeGenerator;
46class DexCompilationUnit;
47class ParallelMoveResolver;
48class SrcMapElem;
49template <class Alloc>
50class SrcMap;
51using DefaultSrcMap = SrcMap<std::allocator<SrcMapElem>>;
52
53class CodeAllocator {
54 public:
55  CodeAllocator() {}
56  virtual ~CodeAllocator() {}
57
58  virtual uint8_t* Allocate(size_t size) = 0;
59
60 private:
61  DISALLOW_COPY_AND_ASSIGN(CodeAllocator);
62};
63
64struct PcInfo {
65  uint32_t dex_pc;
66  uintptr_t native_pc;
67};
68
69class SlowPathCode : public ArenaObject<kArenaAllocSlowPaths> {
70 public:
71  SlowPathCode() {}
72  virtual ~SlowPathCode() {}
73
74  virtual void EmitNativeCode(CodeGenerator* codegen) = 0;
75
76 private:
77  DISALLOW_COPY_AND_ASSIGN(SlowPathCode);
78};
79
80class CodeGenerator : public ArenaObject<kArenaAllocMisc> {
81 public:
82  // Compiles the graph to executable instructions. Returns whether the compilation
83  // succeeded.
84  void CompileBaseline(CodeAllocator* allocator, bool is_leaf = false);
85  void CompileOptimized(CodeAllocator* allocator);
86  static CodeGenerator* Create(ArenaAllocator* allocator,
87                               HGraph* graph,
88                               InstructionSet instruction_set,
89                               const InstructionSetFeatures& isa_features);
90
91  HGraph* GetGraph() const { return graph_; }
92
93  bool GoesToNextBlock(HBasicBlock* current, HBasicBlock* next) const;
94
95  size_t GetStackSlotOfParameter(HParameterValue* parameter) const {
96    // Note that this follows the current calling convention.
97    return GetFrameSize()
98        + kVRegSize  // Art method
99        + parameter->GetIndex() * kVRegSize;
100  }
101
102  virtual void Initialize() = 0;
103  virtual void Finalize(CodeAllocator* allocator);
104  virtual void GenerateFrameEntry() = 0;
105  virtual void GenerateFrameExit() = 0;
106  virtual void Bind(HBasicBlock* block) = 0;
107  virtual void Move(HInstruction* instruction, Location location, HInstruction* move_for) = 0;
108  virtual HGraphVisitor* GetLocationBuilder() = 0;
109  virtual HGraphVisitor* GetInstructionVisitor() = 0;
110  virtual Assembler* GetAssembler() = 0;
111  virtual size_t GetWordSize() const = 0;
112  virtual uintptr_t GetAddressOf(HBasicBlock* block) const = 0;
113  void ComputeFrameSize(size_t number_of_spill_slots,
114                        size_t maximum_number_of_live_registers,
115                        size_t number_of_out_slots);
116  virtual size_t FrameEntrySpillSize() const = 0;
117  int32_t GetStackSlot(HLocal* local) const;
118  Location GetTemporaryLocation(HTemporary* temp) const;
119
120  uint32_t GetFrameSize() const { return frame_size_; }
121  void SetFrameSize(uint32_t size) { frame_size_ = size; }
122  uint32_t GetCoreSpillMask() const { return core_spill_mask_; }
123
124  size_t GetNumberOfCoreRegisters() const { return number_of_core_registers_; }
125  size_t GetNumberOfFloatingPointRegisters() const { return number_of_fpu_registers_; }
126  virtual void SetupBlockedRegisters() const = 0;
127
128  virtual void DumpCoreRegister(std::ostream& stream, int reg) const = 0;
129  virtual void DumpFloatingPointRegister(std::ostream& stream, int reg) const = 0;
130  virtual InstructionSet GetInstructionSet() const = 0;
131  // Saves the register in the stack. Returns the size taken on stack.
132  virtual size_t SaveCoreRegister(size_t stack_index, uint32_t reg_id) = 0;
133  // Restores the register from the stack. Returns the size taken on stack.
134  virtual size_t RestoreCoreRegister(size_t stack_index, uint32_t reg_id) = 0;
135  virtual size_t SaveFloatingPointRegister(size_t stack_index, uint32_t reg_id) {
136    UNUSED(stack_index, reg_id);
137    UNIMPLEMENTED(FATAL);
138    UNREACHABLE();
139  }
140  virtual size_t RestoreFloatingPointRegister(size_t stack_index, uint32_t reg_id) {
141    UNUSED(stack_index, reg_id);
142    UNIMPLEMENTED(FATAL);
143    UNREACHABLE();
144  }
145
146  void RecordPcInfo(HInstruction* instruction, uint32_t dex_pc);
147
148  void AddSlowPath(SlowPathCode* slow_path) {
149    slow_paths_.Add(slow_path);
150  }
151
152  void GenerateSlowPaths();
153
154  void BuildMappingTable(std::vector<uint8_t>* vector, DefaultSrcMap* src_map) const;
155  void BuildVMapTable(std::vector<uint8_t>* vector) const;
156  void BuildNativeGCMap(
157      std::vector<uint8_t>* vector, const DexCompilationUnit& dex_compilation_unit) const;
158  void BuildStackMaps(std::vector<uint8_t>* vector);
159  void SaveLiveRegisters(LocationSummary* locations);
160  void RestoreLiveRegisters(LocationSummary* locations);
161
162  bool IsLeafMethod() const {
163    return is_leaf_;
164  }
165
166  void MarkNotLeaf() {
167    is_leaf_ = false;
168  }
169
170  // Clears the spill slots taken by loop phis in the `LocationSummary` of the
171  // suspend check. This is called when the code generator generates code
172  // for the suspend check at the back edge (instead of where the suspend check
173  // is, which is the loop entry). At this point, the spill slots for the phis
174  // have not been written to.
175  void ClearSpillSlotsFromLoopPhisInStackMap(HSuspendCheck* suspend_check) const;
176
177  bool* GetBlockedCoreRegisters() const { return blocked_core_registers_; }
178  bool* GetBlockedFloatingPointRegisters() const { return blocked_fpu_registers_; }
179
180  // Helper that returns the pointer offset of an index in an object array.
181  // Note: this method assumes we always have the same pointer size, regardless
182  // of the architecture.
183  static size_t GetCacheOffset(uint32_t index);
184
185  void EmitParallelMoves(Location from1, Location to1, Location from2, Location to2);
186
187  static bool StoreNeedsWriteBarrier(Primitive::Type type, HInstruction* value) {
188    if (kIsDebugBuild) {
189      if (type == Primitive::kPrimNot && value->IsIntConstant()) {
190        CHECK_EQ(value->AsIntConstant()->GetValue(), 0);
191      }
192    }
193    return type == Primitive::kPrimNot && !value->IsIntConstant();
194  }
195
196 protected:
197  CodeGenerator(HGraph* graph,
198                size_t number_of_core_registers,
199                size_t number_of_fpu_registers,
200                size_t number_of_register_pairs)
201      : frame_size_(kUninitializedFrameSize),
202        core_spill_mask_(0),
203        first_register_slot_in_slow_path_(0),
204        blocked_core_registers_(graph->GetArena()->AllocArray<bool>(number_of_core_registers)),
205        blocked_fpu_registers_(graph->GetArena()->AllocArray<bool>(number_of_fpu_registers)),
206        blocked_register_pairs_(graph->GetArena()->AllocArray<bool>(number_of_register_pairs)),
207        number_of_core_registers_(number_of_core_registers),
208        number_of_fpu_registers_(number_of_fpu_registers),
209        number_of_register_pairs_(number_of_register_pairs),
210        graph_(graph),
211        pc_infos_(graph->GetArena(), 32),
212        slow_paths_(graph->GetArena(), 8),
213        is_leaf_(true),
214        stack_map_stream_(graph->GetArena()) {}
215  ~CodeGenerator() {}
216
217  // Register allocation logic.
218  void AllocateRegistersLocally(HInstruction* instruction) const;
219
220  // Backend specific implementation for allocating a register.
221  virtual Location AllocateFreeRegister(Primitive::Type type) const = 0;
222
223  static size_t FindFreeEntry(bool* array, size_t length);
224  static size_t FindTwoFreeConsecutiveAlignedEntries(bool* array, size_t length);
225
226  virtual Location GetStackLocation(HLoadLocal* load) const = 0;
227
228  virtual ParallelMoveResolver* GetMoveResolver() = 0;
229
230  // Frame size required for this method.
231  uint32_t frame_size_;
232  uint32_t core_spill_mask_;
233  uint32_t first_register_slot_in_slow_path_;
234
235  // Arrays used when doing register allocation to know which
236  // registers we can allocate. `SetupBlockedRegisters` updates the
237  // arrays.
238  bool* const blocked_core_registers_;
239  bool* const blocked_fpu_registers_;
240  bool* const blocked_register_pairs_;
241  size_t number_of_core_registers_;
242  size_t number_of_fpu_registers_;
243  size_t number_of_register_pairs_;
244
245 private:
246  void InitLocations(HInstruction* instruction);
247  size_t GetStackOffsetOfSavedRegister(size_t index);
248
249  HGraph* const graph_;
250
251  GrowableArray<PcInfo> pc_infos_;
252  GrowableArray<SlowPathCode*> slow_paths_;
253
254  bool is_leaf_;
255
256  StackMapStream stack_map_stream_;
257
258  DISALLOW_COPY_AND_ASSIGN(CodeGenerator);
259};
260
261template <typename C, typename F>
262class CallingConvention {
263 public:
264  CallingConvention(const C* registers,
265                    size_t number_of_registers,
266                    const F* fpu_registers,
267                    size_t number_of_fpu_registers)
268      : registers_(registers),
269        number_of_registers_(number_of_registers),
270        fpu_registers_(fpu_registers),
271        number_of_fpu_registers_(number_of_fpu_registers) {}
272
273  size_t GetNumberOfRegisters() const { return number_of_registers_; }
274  size_t GetNumberOfFpuRegisters() const { return number_of_fpu_registers_; }
275
276  C GetRegisterAt(size_t index) const {
277    DCHECK_LT(index, number_of_registers_);
278    return registers_[index];
279  }
280
281  F GetFpuRegisterAt(size_t index) const {
282    DCHECK_LT(index, number_of_fpu_registers_);
283    return fpu_registers_[index];
284  }
285
286  size_t GetStackOffsetOf(size_t index) const {
287    // We still reserve the space for parameters passed by registers.
288    // Add one for the method pointer.
289    return (index + 1) * kVRegSize;
290  }
291
292 private:
293  const C* registers_;
294  const size_t number_of_registers_;
295  const F* fpu_registers_;
296  const size_t number_of_fpu_registers_;
297
298  DISALLOW_COPY_AND_ASSIGN(CallingConvention);
299};
300
301}  // namespace art
302
303#endif  // ART_COMPILER_OPTIMIZING_CODE_GENERATOR_H_
304