calling_convention_arm64.cc revision b551fdcda9eb128c80de37c4fb978968bec6d4b3
1/* 2 * Copyright (C) 2014 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17#include "base/logging.h" 18#include "calling_convention_arm64.h" 19#include "utils/arm64/managed_register_arm64.h" 20 21namespace art { 22namespace arm64 { 23 24static const Register kCoreArgumentRegisters[] = { 25 X0, X1, X2, X3, X4, X5, X6, X7 26}; 27 28static const WRegister kWArgumentRegisters[] = { 29 W0, W1, W2, W3, W4, W5, W6, W7 30}; 31 32static const DRegister kDArgumentRegisters[] = { 33 D0, D1, D2, D3, D4, D5, D6, D7 34}; 35 36static const SRegister kSArgumentRegisters[] = { 37 S0, S1, S2, S3, S4, S5, S6, S7 38}; 39 40// Calling convention 41ManagedRegister Arm64ManagedRuntimeCallingConvention::InterproceduralScratchRegister() { 42 return Arm64ManagedRegister::FromCoreRegister(X20); // saved on entry restored on exit 43} 44 45ManagedRegister Arm64JniCallingConvention::InterproceduralScratchRegister() { 46 return Arm64ManagedRegister::FromCoreRegister(X20); // saved on entry restored on exit 47} 48 49static ManagedRegister ReturnRegisterForShorty(const char* shorty) { 50 if (shorty[0] == 'F') { 51 return Arm64ManagedRegister::FromSRegister(S0); 52 } else if (shorty[0] == 'D') { 53 return Arm64ManagedRegister::FromDRegister(D0); 54 } else if (shorty[0] == 'J') { 55 return Arm64ManagedRegister::FromCoreRegister(X0); 56 } else if (shorty[0] == 'V') { 57 return Arm64ManagedRegister::NoRegister(); 58 } else { 59 return Arm64ManagedRegister::FromWRegister(W0); 60 } 61} 62 63ManagedRegister Arm64ManagedRuntimeCallingConvention::ReturnRegister() { 64 return ReturnRegisterForShorty(GetShorty()); 65} 66 67ManagedRegister Arm64JniCallingConvention::ReturnRegister() { 68 return ReturnRegisterForShorty(GetShorty()); 69} 70 71ManagedRegister Arm64JniCallingConvention::IntReturnRegister() { 72 return Arm64ManagedRegister::FromWRegister(W0); 73} 74 75// Managed runtime calling convention 76 77ManagedRegister Arm64ManagedRuntimeCallingConvention::MethodRegister() { 78 return Arm64ManagedRegister::FromCoreRegister(X0); 79} 80 81bool Arm64ManagedRuntimeCallingConvention::IsCurrentParamInRegister() { 82 return false; // Everything moved to stack on entry. 83} 84 85bool Arm64ManagedRuntimeCallingConvention::IsCurrentParamOnStack() { 86 return true; 87} 88 89ManagedRegister Arm64ManagedRuntimeCallingConvention::CurrentParamRegister() { 90 LOG(FATAL) << "Should not reach here"; 91 return ManagedRegister::NoRegister(); 92} 93 94FrameOffset Arm64ManagedRuntimeCallingConvention::CurrentParamStackOffset() { 95 CHECK(IsCurrentParamOnStack()); 96 FrameOffset result = 97 FrameOffset(displacement_.Int32Value() + // displacement 98 sizeof(StackReference<mirror::ArtMethod>) + // Method ref 99 (itr_slots_ * sizeof(uint32_t))); // offset into in args 100 return result; 101} 102 103const ManagedRegisterEntrySpills& Arm64ManagedRuntimeCallingConvention::EntrySpills() { 104 // We spill the argument registers on ARM64 to free them up for scratch use, we then assume 105 // all arguments are on the stack. 106 if ((entry_spills_.size() == 0) && (NumArgs() > 0)) { 107 int gp_reg_index = 1; // we start from X1/W1, X0 holds ArtMethod*. 108 int fp_reg_index = 0; // D0/S0. 109 110 // We need to choose the correct register (D/S or X/W) since the managed 111 // stack uses 32bit stack slots. 112 ResetIterator(FrameOffset(0)); 113 while (HasNext()) { 114 if (IsCurrentParamAFloatOrDouble()) { // FP regs. 115 if (fp_reg_index < 8) { 116 if (!IsCurrentParamADouble()) { 117 entry_spills_.push_back(Arm64ManagedRegister::FromSRegister(kSArgumentRegisters[fp_reg_index])); 118 } else { 119 entry_spills_.push_back(Arm64ManagedRegister::FromDRegister(kDArgumentRegisters[fp_reg_index])); 120 } 121 fp_reg_index++; 122 } else { // just increase the stack offset. 123 if (!IsCurrentParamADouble()) { 124 entry_spills_.push_back(ManagedRegister::NoRegister(), 4); 125 } else { 126 entry_spills_.push_back(ManagedRegister::NoRegister(), 8); 127 } 128 } 129 } else { // GP regs. 130 if (gp_reg_index < 8) { 131 if (IsCurrentParamALong() && (!IsCurrentParamAReference())) { 132 entry_spills_.push_back(Arm64ManagedRegister::FromCoreRegister(kCoreArgumentRegisters[gp_reg_index])); 133 } else { 134 entry_spills_.push_back(Arm64ManagedRegister::FromWRegister(kWArgumentRegisters[gp_reg_index])); 135 } 136 gp_reg_index++; 137 } else { // just increase the stack offset. 138 if (IsCurrentParamALong() && (!IsCurrentParamAReference())) { 139 entry_spills_.push_back(ManagedRegister::NoRegister(), 8); 140 } else { 141 entry_spills_.push_back(ManagedRegister::NoRegister(), 4); 142 } 143 } 144 } 145 Next(); 146 } 147 } 148 return entry_spills_; 149} 150 151// JNI calling convention 152Arm64JniCallingConvention::Arm64JniCallingConvention(bool is_static, bool is_synchronized, 153 const char* shorty) 154 : JniCallingConvention(is_static, is_synchronized, shorty, kFramePointerSize) { 155 // TODO: Ugly hard code... 156 // Should generate these according to the spill mask automatically. 157 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X20)); 158 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X21)); 159 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X22)); 160 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X23)); 161 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X24)); 162 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X25)); 163 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X26)); 164 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X27)); 165 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X28)); 166 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X29)); 167 callee_save_regs_.push_back(Arm64ManagedRegister::FromCoreRegister(X30)); 168} 169 170uint32_t Arm64JniCallingConvention::CoreSpillMask() const { 171 // Compute spill mask to agree with callee saves initialized in the constructor 172 // Note: The native jni function may call to some VM runtime functions which may suspend 173 // or trigger GC. And the jni method frame will become top quick frame in those cases. 174 // So we need to satisfy GC to save LR and callee-save registers which is similar to 175 // CalleeSaveMethod(RefOnly) frame. 176 // Jni function is the native function which the java code wants to call. 177 // Jni method is the method that compiled by jni compiler. 178 // Call chain: managed code(java) --> jni method --> jni function. 179 // Thread register(X18, scratched by aapcs64) is not saved on stack, it is saved in ETR(X21). 180 // Suspend register(x19) is preserved by aapcs64 and it is not used in Jni method. 181 return 1 << X20 | 1 << X21 | 1 << X22 | 1 << X23 | 1 << X24 | 1 << X25 | 182 1 << X26 | 1 << X27 | 1 << X28 | 1 << X29 | 1 << LR; 183} 184 185uint32_t Arm64JniCallingConvention::FpSpillMask() const { 186 // Compute spill mask to agree with callee saves initialized in the constructor 187 // Note: All callee-save fp registers will be preserved by aapcs64. And they are not used 188 // in the jni method. 189 return 0; 190} 191 192ManagedRegister Arm64JniCallingConvention::ReturnScratchRegister() const { 193 return ManagedRegister::NoRegister(); 194} 195 196size_t Arm64JniCallingConvention::FrameSize() { 197 // Method*, callee save area size, local reference segment state 198 size_t frame_data_size = sizeof(StackReference<mirror::ArtMethod>) + 199 CalleeSaveRegisters().size() * kFramePointerSize + sizeof(uint32_t); 200 // References plus 2 words for HandleScope header 201 size_t handle_scope_size = HandleScope::SizeOf(kFramePointerSize, ReferenceCount()); 202 // Plus return value spill area size 203 return RoundUp(frame_data_size + handle_scope_size + SizeOfReturnValue(), kStackAlignment); 204} 205 206size_t Arm64JniCallingConvention::OutArgSize() { 207 return RoundUp(NumberOfOutgoingStackArgs() * kFramePointerSize, kStackAlignment); 208} 209 210bool Arm64JniCallingConvention::IsCurrentParamInRegister() { 211 if (IsCurrentParamAFloatOrDouble()) { 212 return (itr_float_and_doubles_ < 8); 213 } else { 214 return ((itr_args_ - itr_float_and_doubles_) < 8); 215 } 216} 217 218bool Arm64JniCallingConvention::IsCurrentParamOnStack() { 219 return !IsCurrentParamInRegister(); 220} 221 222ManagedRegister Arm64JniCallingConvention::CurrentParamRegister() { 223 CHECK(IsCurrentParamInRegister()); 224 if (IsCurrentParamAFloatOrDouble()) { 225 CHECK_LT(itr_float_and_doubles_, 8u); 226 if (IsCurrentParamADouble()) { 227 return Arm64ManagedRegister::FromDRegister(kDArgumentRegisters[itr_float_and_doubles_]); 228 } else { 229 return Arm64ManagedRegister::FromSRegister(kSArgumentRegisters[itr_float_and_doubles_]); 230 } 231 } else { 232 int gp_reg = itr_args_ - itr_float_and_doubles_; 233 CHECK_LT(static_cast<unsigned int>(gp_reg), 8u); 234 if (IsCurrentParamALong() || IsCurrentParamAReference() || IsCurrentParamJniEnv()) { 235 return Arm64ManagedRegister::FromCoreRegister(kCoreArgumentRegisters[gp_reg]); 236 } else { 237 return Arm64ManagedRegister::FromWRegister(kWArgumentRegisters[gp_reg]); 238 } 239 } 240} 241 242FrameOffset Arm64JniCallingConvention::CurrentParamStackOffset() { 243 CHECK(IsCurrentParamOnStack()); 244 size_t args_on_stack = itr_args_ 245 - std::min(8u, itr_float_and_doubles_) 246 - std::min(8u, (itr_args_ - itr_float_and_doubles_)); 247 size_t offset = displacement_.Int32Value() - OutArgSize() + (args_on_stack * kFramePointerSize); 248 CHECK_LT(offset, OutArgSize()); 249 return FrameOffset(offset); 250} 251 252size_t Arm64JniCallingConvention::NumberOfOutgoingStackArgs() { 253 // all arguments including JNI args 254 size_t all_args = NumArgs() + NumberOfExtraArgumentsForJni(); 255 256 size_t all_stack_args = all_args - 257 std::min(8u, static_cast<unsigned int>(NumFloatOrDoubleArgs())) - 258 std::min(8u, static_cast<unsigned int>((all_args - NumFloatOrDoubleArgs()))); 259 260 return all_stack_args; 261} 262 263} // namespace arm64 264} // namespace art 265