code_generator.h revision 32f5b4d2c8c9b52e9522941c159577b21752d0fa
1/* 2 * Copyright (C) 2014 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17#ifndef ART_COMPILER_OPTIMIZING_CODE_GENERATOR_H_ 18#define ART_COMPILER_OPTIMIZING_CODE_GENERATOR_H_ 19 20#include "arch/instruction_set.h" 21#include "base/bit_field.h" 22#include "globals.h" 23#include "locations.h" 24#include "memory_region.h" 25#include "nodes.h" 26#include "stack_map_stream.h" 27 28namespace art { 29 30static size_t constexpr kVRegSize = 4; 31static size_t constexpr kUninitializedFrameSize = 0; 32 33// Binary encoding of 2^32 for type double. 34static int64_t constexpr k2Pow32EncodingForDouble = INT64_C(0x41F0000000000000); 35// Binary encoding of 2^31 for type double. 36static int64_t constexpr k2Pow31EncodingForDouble = INT64_C(0x41E0000000000000); 37 38class Assembler; 39class CodeGenerator; 40class DexCompilationUnit; 41class ParallelMoveResolver; 42class SrcMap; 43 44class CodeAllocator { 45 public: 46 CodeAllocator() {} 47 virtual ~CodeAllocator() {} 48 49 virtual uint8_t* Allocate(size_t size) = 0; 50 51 private: 52 DISALLOW_COPY_AND_ASSIGN(CodeAllocator); 53}; 54 55struct PcInfo { 56 uint32_t dex_pc; 57 uintptr_t native_pc; 58}; 59 60class SlowPathCode : public ArenaObject<kArenaAllocSlowPaths> { 61 public: 62 SlowPathCode() {} 63 virtual ~SlowPathCode() {} 64 65 virtual void EmitNativeCode(CodeGenerator* codegen) = 0; 66 67 private: 68 DISALLOW_COPY_AND_ASSIGN(SlowPathCode); 69}; 70 71class CodeGenerator : public ArenaObject<kArenaAllocMisc> { 72 public: 73 // Compiles the graph to executable instructions. Returns whether the compilation 74 // succeeded. 75 void CompileBaseline(CodeAllocator* allocator, bool is_leaf = false); 76 void CompileOptimized(CodeAllocator* allocator); 77 static CodeGenerator* Create(ArenaAllocator* allocator, 78 HGraph* graph, 79 InstructionSet instruction_set); 80 81 HGraph* GetGraph() const { return graph_; } 82 83 bool GoesToNextBlock(HBasicBlock* current, HBasicBlock* next) const; 84 85 size_t GetStackSlotOfParameter(HParameterValue* parameter) const { 86 // Note that this follows the current calling convention. 87 return GetFrameSize() 88 + kVRegSize // Art method 89 + parameter->GetIndex() * kVRegSize; 90 } 91 92 virtual void Initialize() = 0; 93 virtual void Finalize(CodeAllocator* allocator); 94 virtual void GenerateFrameEntry() = 0; 95 virtual void GenerateFrameExit() = 0; 96 virtual void Bind(HBasicBlock* block) = 0; 97 virtual void Move(HInstruction* instruction, Location location, HInstruction* move_for) = 0; 98 virtual HGraphVisitor* GetLocationBuilder() = 0; 99 virtual HGraphVisitor* GetInstructionVisitor() = 0; 100 virtual Assembler* GetAssembler() = 0; 101 virtual size_t GetWordSize() const = 0; 102 virtual uintptr_t GetAddressOf(HBasicBlock* block) const = 0; 103 void ComputeFrameSize(size_t number_of_spill_slots, 104 size_t maximum_number_of_live_registers, 105 size_t number_of_out_slots); 106 virtual size_t FrameEntrySpillSize() const = 0; 107 int32_t GetStackSlot(HLocal* local) const; 108 Location GetTemporaryLocation(HTemporary* temp) const; 109 110 uint32_t GetFrameSize() const { return frame_size_; } 111 void SetFrameSize(uint32_t size) { frame_size_ = size; } 112 uint32_t GetCoreSpillMask() const { return core_spill_mask_; } 113 114 size_t GetNumberOfCoreRegisters() const { return number_of_core_registers_; } 115 size_t GetNumberOfFloatingPointRegisters() const { return number_of_fpu_registers_; } 116 virtual void SetupBlockedRegisters() const = 0; 117 118 virtual void DumpCoreRegister(std::ostream& stream, int reg) const = 0; 119 virtual void DumpFloatingPointRegister(std::ostream& stream, int reg) const = 0; 120 virtual InstructionSet GetInstructionSet() const = 0; 121 // Saves the register in the stack. Returns the size taken on stack. 122 virtual size_t SaveCoreRegister(size_t stack_index, uint32_t reg_id) = 0; 123 // Restores the register from the stack. Returns the size taken on stack. 124 virtual size_t RestoreCoreRegister(size_t stack_index, uint32_t reg_id) = 0; 125 virtual size_t SaveFloatingPointRegister(size_t stack_index, uint32_t reg_id) { 126 UNUSED(stack_index, reg_id); 127 UNIMPLEMENTED(FATAL); 128 UNREACHABLE(); 129 } 130 virtual size_t RestoreFloatingPointRegister(size_t stack_index, uint32_t reg_id) { 131 UNUSED(stack_index, reg_id); 132 UNIMPLEMENTED(FATAL); 133 UNREACHABLE(); 134 } 135 136 void RecordPcInfo(HInstruction* instruction, uint32_t dex_pc); 137 138 void AddSlowPath(SlowPathCode* slow_path) { 139 slow_paths_.Add(slow_path); 140 } 141 142 void GenerateSlowPaths(); 143 144 void BuildMappingTable(std::vector<uint8_t>* vector, SrcMap* src_map) const; 145 void BuildVMapTable(std::vector<uint8_t>* vector) const; 146 void BuildNativeGCMap( 147 std::vector<uint8_t>* vector, const DexCompilationUnit& dex_compilation_unit) const; 148 void BuildStackMaps(std::vector<uint8_t>* vector); 149 void SaveLiveRegisters(LocationSummary* locations); 150 void RestoreLiveRegisters(LocationSummary* locations); 151 152 bool IsLeafMethod() const { 153 return is_leaf_; 154 } 155 156 void MarkNotLeaf() { 157 is_leaf_ = false; 158 } 159 160 // Clears the spill slots taken by loop phis in the `LocationSummary` of the 161 // suspend check. This is called when the code generator generates code 162 // for the suspend check at the back edge (instead of where the suspend check 163 // is, which is the loop entry). At this point, the spill slots for the phis 164 // have not been written to. 165 void ClearSpillSlotsFromLoopPhisInStackMap(HSuspendCheck* suspend_check) const; 166 167 bool* GetBlockedCoreRegisters() const { return blocked_core_registers_; } 168 bool* GetBlockedFloatingPointRegisters() const { return blocked_fpu_registers_; } 169 170 // Helper that returns the pointer offset of an index in an object array. 171 // Note: this method assumes we always have the same pointer size, regardless 172 // of the architecture. 173 static size_t GetCacheOffset(uint32_t index); 174 175 void EmitParallelMoves(Location from1, Location to1, Location from2, Location to2); 176 177 static bool StoreNeedsWriteBarrier(Primitive::Type type, HInstruction* value) { 178 if (kIsDebugBuild) { 179 if (type == Primitive::kPrimNot && value->IsIntConstant()) { 180 CHECK_EQ(value->AsIntConstant()->GetValue(), 0); 181 } 182 } 183 return type == Primitive::kPrimNot && !value->IsIntConstant(); 184 } 185 186 protected: 187 CodeGenerator(HGraph* graph, 188 size_t number_of_core_registers, 189 size_t number_of_fpu_registers, 190 size_t number_of_register_pairs) 191 : frame_size_(kUninitializedFrameSize), 192 core_spill_mask_(0), 193 first_register_slot_in_slow_path_(0), 194 blocked_core_registers_(graph->GetArena()->AllocArray<bool>(number_of_core_registers)), 195 blocked_fpu_registers_(graph->GetArena()->AllocArray<bool>(number_of_fpu_registers)), 196 blocked_register_pairs_(graph->GetArena()->AllocArray<bool>(number_of_register_pairs)), 197 number_of_core_registers_(number_of_core_registers), 198 number_of_fpu_registers_(number_of_fpu_registers), 199 number_of_register_pairs_(number_of_register_pairs), 200 graph_(graph), 201 pc_infos_(graph->GetArena(), 32), 202 slow_paths_(graph->GetArena(), 8), 203 is_leaf_(true), 204 stack_map_stream_(graph->GetArena()) {} 205 ~CodeGenerator() {} 206 207 // Register allocation logic. 208 void AllocateRegistersLocally(HInstruction* instruction) const; 209 210 // Backend specific implementation for allocating a register. 211 virtual Location AllocateFreeRegister(Primitive::Type type) const = 0; 212 213 static size_t FindFreeEntry(bool* array, size_t length); 214 static size_t FindTwoFreeConsecutiveAlignedEntries(bool* array, size_t length); 215 216 virtual Location GetStackLocation(HLoadLocal* load) const = 0; 217 218 virtual ParallelMoveResolver* GetMoveResolver() = 0; 219 220 // Frame size required for this method. 221 uint32_t frame_size_; 222 uint32_t core_spill_mask_; 223 uint32_t first_register_slot_in_slow_path_; 224 225 // Arrays used when doing register allocation to know which 226 // registers we can allocate. `SetupBlockedRegisters` updates the 227 // arrays. 228 bool* const blocked_core_registers_; 229 bool* const blocked_fpu_registers_; 230 bool* const blocked_register_pairs_; 231 size_t number_of_core_registers_; 232 size_t number_of_fpu_registers_; 233 size_t number_of_register_pairs_; 234 235 private: 236 void InitLocations(HInstruction* instruction); 237 size_t GetStackOffsetOfSavedRegister(size_t index); 238 239 HGraph* const graph_; 240 241 GrowableArray<PcInfo> pc_infos_; 242 GrowableArray<SlowPathCode*> slow_paths_; 243 244 bool is_leaf_; 245 246 StackMapStream stack_map_stream_; 247 248 DISALLOW_COPY_AND_ASSIGN(CodeGenerator); 249}; 250 251template <typename C, typename F> 252class CallingConvention { 253 public: 254 CallingConvention(const C* registers, 255 size_t number_of_registers, 256 const F* fpu_registers, 257 size_t number_of_fpu_registers) 258 : registers_(registers), 259 number_of_registers_(number_of_registers), 260 fpu_registers_(fpu_registers), 261 number_of_fpu_registers_(number_of_fpu_registers) {} 262 263 size_t GetNumberOfRegisters() const { return number_of_registers_; } 264 size_t GetNumberOfFpuRegisters() const { return number_of_fpu_registers_; } 265 266 C GetRegisterAt(size_t index) const { 267 DCHECK_LT(index, number_of_registers_); 268 return registers_[index]; 269 } 270 271 F GetFpuRegisterAt(size_t index) const { 272 DCHECK_LT(index, number_of_fpu_registers_); 273 return fpu_registers_[index]; 274 } 275 276 size_t GetStackOffsetOf(size_t index) const { 277 // We still reserve the space for parameters passed by registers. 278 // Add one for the method pointer. 279 return (index + 1) * kVRegSize; 280 } 281 282 private: 283 const C* registers_; 284 const size_t number_of_registers_; 285 const F* fpu_registers_; 286 const size_t number_of_fpu_registers_; 287 288 DISALLOW_COPY_AND_ASSIGN(CallingConvention); 289}; 290 291} // namespace art 292 293#endif // ART_COMPILER_OPTIMIZING_CODE_GENERATOR_H_ 294