InterpC-x86.cpp revision 60fc806b679a3655c228b4093058c59941a49cfe
1/* 2 * This file was generated automatically by gen-mterp.py for 'x86'. 3 * 4 * --> DO NOT EDIT <-- 5 */ 6 7/* File: c/header.cpp */ 8/* 9 * Copyright (C) 2008 The Android Open Source Project 10 * 11 * Licensed under the Apache License, Version 2.0 (the "License"); 12 * you may not use this file except in compliance with the License. 13 * You may obtain a copy of the License at 14 * 15 * http://www.apache.org/licenses/LICENSE-2.0 16 * 17 * Unless required by applicable law or agreed to in writing, software 18 * distributed under the License is distributed on an "AS IS" BASIS, 19 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 20 * See the License for the specific language governing permissions and 21 * limitations under the License. 22 */ 23 24/* common includes */ 25#include "Dalvik.h" 26#include "interp/InterpDefs.h" 27#include "mterp/Mterp.h" 28#include <math.h> // needed for fmod, fmodf 29#include "mterp/common/FindInterface.h" 30 31/* 32 * Configuration defines. These affect the C implementations, i.e. the 33 * portable interpreter(s) and C stubs. 34 * 35 * Some defines are controlled by the Makefile, e.g.: 36 * WITH_INSTR_CHECKS 37 * WITH_TRACKREF_CHECKS 38 * EASY_GDB 39 * NDEBUG 40 */ 41 42#ifdef WITH_INSTR_CHECKS /* instruction-level paranoia (slow!) */ 43# define CHECK_BRANCH_OFFSETS 44# define CHECK_REGISTER_INDICES 45#endif 46 47/* 48 * Some architectures require 64-bit alignment for access to 64-bit data 49 * types. We can't just use pointers to copy 64-bit values out of our 50 * interpreted register set, because gcc may assume the pointer target is 51 * aligned and generate invalid code. 52 * 53 * There are two common approaches: 54 * (1) Use a union that defines a 32-bit pair and a 64-bit value. 55 * (2) Call memcpy(). 56 * 57 * Depending upon what compiler you're using and what options are specified, 58 * one may be faster than the other. For example, the compiler might 59 * convert a memcpy() of 8 bytes into a series of instructions and omit 60 * the call. The union version could cause some strange side-effects, 61 * e.g. for a while ARM gcc thought it needed separate storage for each 62 * inlined instance, and generated instructions to zero out ~700 bytes of 63 * stack space at the top of the interpreter. 64 * 65 * The default is to use memcpy(). The current gcc for ARM seems to do 66 * better with the union. 67 */ 68#if defined(__ARM_EABI__) 69# define NO_UNALIGN_64__UNION 70#endif 71 72 73//#define LOG_INSTR /* verbose debugging */ 74/* set and adjust ANDROID_LOG_TAGS='*:i jdwp:i dalvikvm:i dalvikvmi:i' */ 75 76/* 77 * Export another copy of the PC on every instruction; this is largely 78 * redundant with EXPORT_PC and the debugger code. This value can be 79 * compared against what we have stored on the stack with EXPORT_PC to 80 * help ensure that we aren't missing any export calls. 81 */ 82#if WITH_EXTRA_GC_CHECKS > 1 83# define EXPORT_EXTRA_PC() (self->currentPc2 = pc) 84#else 85# define EXPORT_EXTRA_PC() 86#endif 87 88/* 89 * Adjust the program counter. "_offset" is a signed int, in 16-bit units. 90 * 91 * Assumes the existence of "const u2* pc" and "const u2* curMethod->insns". 92 * 93 * We don't advance the program counter until we finish an instruction or 94 * branch, because we do want to have to unroll the PC if there's an 95 * exception. 96 */ 97#ifdef CHECK_BRANCH_OFFSETS 98# define ADJUST_PC(_offset) do { \ 99 int myoff = _offset; /* deref only once */ \ 100 if (pc + myoff < curMethod->insns || \ 101 pc + myoff >= curMethod->insns + dvmGetMethodInsnsSize(curMethod)) \ 102 { \ 103 char* desc; \ 104 desc = dexProtoCopyMethodDescriptor(&curMethod->prototype); \ 105 LOGE("Invalid branch %d at 0x%04x in %s.%s %s", \ 106 myoff, (int) (pc - curMethod->insns), \ 107 curMethod->clazz->descriptor, curMethod->name, desc); \ 108 free(desc); \ 109 dvmAbort(); \ 110 } \ 111 pc += myoff; \ 112 EXPORT_EXTRA_PC(); \ 113 } while (false) 114#else 115# define ADJUST_PC(_offset) do { \ 116 pc += _offset; \ 117 EXPORT_EXTRA_PC(); \ 118 } while (false) 119#endif 120 121/* 122 * If enabled, log instructions as we execute them. 123 */ 124#ifdef LOG_INSTR 125# define ILOGD(...) ILOG(LOG_DEBUG, __VA_ARGS__) 126# define ILOGV(...) ILOG(LOG_VERBOSE, __VA_ARGS__) 127# define ILOG(_level, ...) do { \ 128 char debugStrBuf[128]; \ 129 snprintf(debugStrBuf, sizeof(debugStrBuf), __VA_ARGS__); \ 130 if (curMethod != NULL) \ 131 LOG(_level, LOG_TAG"i", "%-2d|%04x%s", \ 132 self->threadId, (int)(pc - curMethod->insns), debugStrBuf); \ 133 else \ 134 LOG(_level, LOG_TAG"i", "%-2d|####%s", \ 135 self->threadId, debugStrBuf); \ 136 } while(false) 137void dvmDumpRegs(const Method* method, const u4* framePtr, bool inOnly); 138# define DUMP_REGS(_meth, _frame, _inOnly) dvmDumpRegs(_meth, _frame, _inOnly) 139static const char kSpacing[] = " "; 140#else 141# define ILOGD(...) ((void)0) 142# define ILOGV(...) ((void)0) 143# define DUMP_REGS(_meth, _frame, _inOnly) ((void)0) 144#endif 145 146/* get a long from an array of u4 */ 147static inline s8 getLongFromArray(const u4* ptr, int idx) 148{ 149#if defined(NO_UNALIGN_64__UNION) 150 union { s8 ll; u4 parts[2]; } conv; 151 152 ptr += idx; 153 conv.parts[0] = ptr[0]; 154 conv.parts[1] = ptr[1]; 155 return conv.ll; 156#else 157 s8 val; 158 memcpy(&val, &ptr[idx], 8); 159 return val; 160#endif 161} 162 163/* store a long into an array of u4 */ 164static inline void putLongToArray(u4* ptr, int idx, s8 val) 165{ 166#if defined(NO_UNALIGN_64__UNION) 167 union { s8 ll; u4 parts[2]; } conv; 168 169 ptr += idx; 170 conv.ll = val; 171 ptr[0] = conv.parts[0]; 172 ptr[1] = conv.parts[1]; 173#else 174 memcpy(&ptr[idx], &val, 8); 175#endif 176} 177 178/* get a double from an array of u4 */ 179static inline double getDoubleFromArray(const u4* ptr, int idx) 180{ 181#if defined(NO_UNALIGN_64__UNION) 182 union { double d; u4 parts[2]; } conv; 183 184 ptr += idx; 185 conv.parts[0] = ptr[0]; 186 conv.parts[1] = ptr[1]; 187 return conv.d; 188#else 189 double dval; 190 memcpy(&dval, &ptr[idx], 8); 191 return dval; 192#endif 193} 194 195/* store a double into an array of u4 */ 196static inline void putDoubleToArray(u4* ptr, int idx, double dval) 197{ 198#if defined(NO_UNALIGN_64__UNION) 199 union { double d; u4 parts[2]; } conv; 200 201 ptr += idx; 202 conv.d = dval; 203 ptr[0] = conv.parts[0]; 204 ptr[1] = conv.parts[1]; 205#else 206 memcpy(&ptr[idx], &dval, 8); 207#endif 208} 209 210/* 211 * If enabled, validate the register number on every access. Otherwise, 212 * just do an array access. 213 * 214 * Assumes the existence of "u4* fp". 215 * 216 * "_idx" may be referenced more than once. 217 */ 218#ifdef CHECK_REGISTER_INDICES 219# define GET_REGISTER(_idx) \ 220 ( (_idx) < curMethod->registersSize ? \ 221 (fp[(_idx)]) : (assert(!"bad reg"),1969) ) 222# define SET_REGISTER(_idx, _val) \ 223 ( (_idx) < curMethod->registersSize ? \ 224 (fp[(_idx)] = (u4)(_val)) : (assert(!"bad reg"),1969) ) 225# define GET_REGISTER_AS_OBJECT(_idx) ((Object *)GET_REGISTER(_idx)) 226# define SET_REGISTER_AS_OBJECT(_idx, _val) SET_REGISTER(_idx, (s4)_val) 227# define GET_REGISTER_INT(_idx) ((s4) GET_REGISTER(_idx)) 228# define SET_REGISTER_INT(_idx, _val) SET_REGISTER(_idx, (s4)_val) 229# define GET_REGISTER_WIDE(_idx) \ 230 ( (_idx) < curMethod->registersSize-1 ? \ 231 getLongFromArray(fp, (_idx)) : (assert(!"bad reg"),1969) ) 232# define SET_REGISTER_WIDE(_idx, _val) \ 233 ( (_idx) < curMethod->registersSize-1 ? \ 234 (void)putLongToArray(fp, (_idx), (_val)) : assert(!"bad reg") ) 235# define GET_REGISTER_FLOAT(_idx) \ 236 ( (_idx) < curMethod->registersSize ? \ 237 (*((float*) &fp[(_idx)])) : (assert(!"bad reg"),1969.0f) ) 238# define SET_REGISTER_FLOAT(_idx, _val) \ 239 ( (_idx) < curMethod->registersSize ? \ 240 (*((float*) &fp[(_idx)]) = (_val)) : (assert(!"bad reg"),1969.0f) ) 241# define GET_REGISTER_DOUBLE(_idx) \ 242 ( (_idx) < curMethod->registersSize-1 ? \ 243 getDoubleFromArray(fp, (_idx)) : (assert(!"bad reg"),1969.0) ) 244# define SET_REGISTER_DOUBLE(_idx, _val) \ 245 ( (_idx) < curMethod->registersSize-1 ? \ 246 (void)putDoubleToArray(fp, (_idx), (_val)) : assert(!"bad reg") ) 247#else 248# define GET_REGISTER(_idx) (fp[(_idx)]) 249# define SET_REGISTER(_idx, _val) (fp[(_idx)] = (_val)) 250# define GET_REGISTER_AS_OBJECT(_idx) ((Object*) fp[(_idx)]) 251# define SET_REGISTER_AS_OBJECT(_idx, _val) (fp[(_idx)] = (u4)(_val)) 252# define GET_REGISTER_INT(_idx) ((s4)GET_REGISTER(_idx)) 253# define SET_REGISTER_INT(_idx, _val) SET_REGISTER(_idx, (s4)_val) 254# define GET_REGISTER_WIDE(_idx) getLongFromArray(fp, (_idx)) 255# define SET_REGISTER_WIDE(_idx, _val) putLongToArray(fp, (_idx), (_val)) 256# define GET_REGISTER_FLOAT(_idx) (*((float*) &fp[(_idx)])) 257# define SET_REGISTER_FLOAT(_idx, _val) (*((float*) &fp[(_idx)]) = (_val)) 258# define GET_REGISTER_DOUBLE(_idx) getDoubleFromArray(fp, (_idx)) 259# define SET_REGISTER_DOUBLE(_idx, _val) putDoubleToArray(fp, (_idx), (_val)) 260#endif 261 262/* 263 * Get 16 bits from the specified offset of the program counter. We always 264 * want to load 16 bits at a time from the instruction stream -- it's more 265 * efficient than 8 and won't have the alignment problems that 32 might. 266 * 267 * Assumes existence of "const u2* pc". 268 */ 269#define FETCH(_offset) (pc[(_offset)]) 270 271/* 272 * Extract instruction byte from 16-bit fetch (_inst is a u2). 273 */ 274#define INST_INST(_inst) ((_inst) & 0xff) 275 276/* 277 * Replace the opcode (used when handling breakpoints). _opcode is a u1. 278 */ 279#define INST_REPLACE_OP(_inst, _opcode) (((_inst) & 0xff00) | _opcode) 280 281/* 282 * Extract the "vA, vB" 4-bit registers from the instruction word (_inst is u2). 283 */ 284#define INST_A(_inst) (((_inst) >> 8) & 0x0f) 285#define INST_B(_inst) ((_inst) >> 12) 286 287/* 288 * Get the 8-bit "vAA" 8-bit register index from the instruction word. 289 * (_inst is u2) 290 */ 291#define INST_AA(_inst) ((_inst) >> 8) 292 293/* 294 * The current PC must be available to Throwable constructors, e.g. 295 * those created by the various exception throw routines, so that the 296 * exception stack trace can be generated correctly. If we don't do this, 297 * the offset within the current method won't be shown correctly. See the 298 * notes in Exception.c. 299 * 300 * This is also used to determine the address for precise GC. 301 * 302 * Assumes existence of "u4* fp" and "const u2* pc". 303 */ 304#define EXPORT_PC() (SAVEAREA_FROM_FP(fp)->xtra.currentPc = pc) 305 306/* 307 * Check to see if "obj" is NULL. If so, throw an exception. Assumes the 308 * pc has already been exported to the stack. 309 * 310 * Perform additional checks on debug builds. 311 * 312 * Use this to check for NULL when the instruction handler calls into 313 * something that could throw an exception (so we have already called 314 * EXPORT_PC at the top). 315 */ 316static inline bool checkForNull(Object* obj) 317{ 318 if (obj == NULL) { 319 dvmThrowNullPointerException(NULL); 320 return false; 321 } 322#ifdef WITH_EXTRA_OBJECT_VALIDATION 323 if (!dvmIsValidObject(obj)) { 324 LOGE("Invalid object %p", obj); 325 dvmAbort(); 326 } 327#endif 328#ifndef NDEBUG 329 if (obj->clazz == NULL || ((u4) obj->clazz) <= 65536) { 330 /* probable heap corruption */ 331 LOGE("Invalid object class %p (in %p)", obj->clazz, obj); 332 dvmAbort(); 333 } 334#endif 335 return true; 336} 337 338/* 339 * Check to see if "obj" is NULL. If so, export the PC into the stack 340 * frame and throw an exception. 341 * 342 * Perform additional checks on debug builds. 343 * 344 * Use this to check for NULL when the instruction handler doesn't do 345 * anything else that can throw an exception. 346 */ 347static inline bool checkForNullExportPC(Object* obj, u4* fp, const u2* pc) 348{ 349 if (obj == NULL) { 350 EXPORT_PC(); 351 dvmThrowNullPointerException(NULL); 352 return false; 353 } 354#ifdef WITH_EXTRA_OBJECT_VALIDATION 355 if (!dvmIsValidObject(obj)) { 356 LOGE("Invalid object %p", obj); 357 dvmAbort(); 358 } 359#endif 360#ifndef NDEBUG 361 if (obj->clazz == NULL || ((u4) obj->clazz) <= 65536) { 362 /* probable heap corruption */ 363 LOGE("Invalid object class %p (in %p)", obj->clazz, obj); 364 dvmAbort(); 365 } 366#endif 367 return true; 368} 369 370/* File: cstubs/stubdefs.cpp */ 371/* 372 * In the C mterp stubs, "goto" is a function call followed immediately 373 * by a return. 374 */ 375 376#define GOTO_TARGET_DECL(_target, ...) \ 377 extern "C" void dvmMterp_##_target(Thread* self, ## __VA_ARGS__); 378 379/* (void)xxx to quiet unused variable compiler warnings. */ 380#define GOTO_TARGET(_target, ...) \ 381 void dvmMterp_##_target(Thread* self, ## __VA_ARGS__) { \ 382 u2 ref, vsrc1, vsrc2, vdst; \ 383 u2 inst = FETCH(0); \ 384 const Method* methodToCall; \ 385 StackSaveArea* debugSaveArea; \ 386 (void)ref; (void)vsrc1; (void)vsrc2; (void)vdst; (void)inst; \ 387 (void)methodToCall; (void)debugSaveArea; 388 389#define GOTO_TARGET_END } 390 391/* 392 * Redefine what used to be local variable accesses into Thread struct 393 * references. (These are undefined down in "footer.cpp".) 394 */ 395#define retval self->interpSave.retval 396#define pc self->interpSave.pc 397#define fp self->interpSave.curFrame 398#define curMethod self->interpSave.method 399#define methodClassDex self->interpSave.methodClassDex 400#define debugTrackedRefStart self->interpSave.debugTrackedRefStart 401 402/* ugh */ 403#define STUB_HACK(x) x 404#if defined(WITH_JIT) 405#define JIT_STUB_HACK(x) x 406#else 407#define JIT_STUB_HACK(x) 408#endif 409 410/* 411 * InterpSave's pc and fp must be valid when breaking out to a 412 * "Reportxxx" routine. Because the portable interpreter uses local 413 * variables for these, we must flush prior. Stubs, however, use 414 * the interpSave vars directly, so this is a nop for stubs. 415 */ 416#define PC_FP_TO_SELF() 417#define PC_TO_SELF() 418 419/* 420 * Opcode handler framing macros. Here, each opcode is a separate function 421 * that takes a "self" argument and returns void. We can't declare 422 * these "static" because they may be called from an assembly stub. 423 * (void)xxx to quiet unused variable compiler warnings. 424 */ 425#define HANDLE_OPCODE(_op) \ 426 extern "C" void dvmMterp_##_op(Thread* self); \ 427 void dvmMterp_##_op(Thread* self) { \ 428 u4 ref; \ 429 u2 vsrc1, vsrc2, vdst; \ 430 u2 inst = FETCH(0); \ 431 (void)ref; (void)vsrc1; (void)vsrc2; (void)vdst; (void)inst; 432 433#define OP_END } 434 435/* 436 * Like the "portable" FINISH, but don't reload "inst", and return to caller 437 * when done. Further, debugger/profiler checks are handled 438 * before handler execution in mterp, so we don't do them here either. 439 */ 440#if defined(WITH_JIT) 441#define FINISH(_offset) { \ 442 ADJUST_PC(_offset); \ 443 if (self->interpBreak.ctl.subMode & kSubModeJitTraceBuild) { \ 444 dvmCheckJit(pc, self); \ 445 } \ 446 return; \ 447 } 448#else 449#define FINISH(_offset) { \ 450 ADJUST_PC(_offset); \ 451 return; \ 452 } 453#endif 454 455 456/* 457 * The "goto label" statements turn into function calls followed by 458 * return statements. Some of the functions take arguments, which in the 459 * portable interpreter are handled by assigning values to globals. 460 */ 461 462#define GOTO_exceptionThrown() \ 463 do { \ 464 dvmMterp_exceptionThrown(self); \ 465 return; \ 466 } while(false) 467 468#define GOTO_returnFromMethod() \ 469 do { \ 470 dvmMterp_returnFromMethod(self); \ 471 return; \ 472 } while(false) 473 474#define GOTO_invoke(_target, _methodCallRange, _jumboFormat) \ 475 do { \ 476 dvmMterp_##_target(self, _methodCallRange, _jumboFormat); \ 477 return; \ 478 } while(false) 479 480#define GOTO_invokeMethod(_methodCallRange, _methodToCall, _vsrc1, _vdst) \ 481 do { \ 482 dvmMterp_invokeMethod(self, _methodCallRange, _methodToCall, \ 483 _vsrc1, _vdst); \ 484 return; \ 485 } while(false) 486 487/* 488 * As a special case, "goto bail" turns into a longjmp. 489 */ 490#define GOTO_bail() \ 491 dvmMterpStdBail(self, false); 492 493/* 494 * Periodically check for thread suspension. 495 * 496 * While we're at it, see if a debugger has attached or the profiler has 497 * started. 498 */ 499#define PERIODIC_CHECKS(_pcadj) { \ 500 if (dvmCheckSuspendQuick(self)) { \ 501 EXPORT_PC(); /* need for precise GC */ \ 502 dvmCheckSuspendPending(self); \ 503 } \ 504 } 505 506/* File: c/opcommon.cpp */ 507/* forward declarations of goto targets */ 508GOTO_TARGET_DECL(filledNewArray, bool methodCallRange, bool jumboFormat); 509GOTO_TARGET_DECL(invokeVirtual, bool methodCallRange, bool jumboFormat); 510GOTO_TARGET_DECL(invokeSuper, bool methodCallRange, bool jumboFormat); 511GOTO_TARGET_DECL(invokeInterface, bool methodCallRange, bool jumboFormat); 512GOTO_TARGET_DECL(invokeDirect, bool methodCallRange, bool jumboFormat); 513GOTO_TARGET_DECL(invokeStatic, bool methodCallRange, bool jumboFormat); 514GOTO_TARGET_DECL(invokeVirtualQuick, bool methodCallRange, bool jumboFormat); 515GOTO_TARGET_DECL(invokeSuperQuick, bool methodCallRange, bool jumboFormat); 516GOTO_TARGET_DECL(invokeMethod, bool methodCallRange, const Method* methodToCall, 517 u2 count, u2 regs); 518GOTO_TARGET_DECL(returnFromMethod); 519GOTO_TARGET_DECL(exceptionThrown); 520 521/* 522 * =========================================================================== 523 * 524 * What follows are opcode definitions shared between multiple opcodes with 525 * minor substitutions handled by the C pre-processor. These should probably 526 * use the mterp substitution mechanism instead, with the code here moved 527 * into common fragment files (like the asm "binop.S"), although it's hard 528 * to give up the C preprocessor in favor of the much simpler text subst. 529 * 530 * =========================================================================== 531 */ 532 533#define HANDLE_NUMCONV(_opcode, _opname, _fromtype, _totype) \ 534 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 535 vdst = INST_A(inst); \ 536 vsrc1 = INST_B(inst); \ 537 ILOGV("|%s v%d,v%d", (_opname), vdst, vsrc1); \ 538 SET_REGISTER##_totype(vdst, \ 539 GET_REGISTER##_fromtype(vsrc1)); \ 540 FINISH(1); 541 542#define HANDLE_FLOAT_TO_INT(_opcode, _opname, _fromvtype, _fromrtype, \ 543 _tovtype, _tortype) \ 544 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 545 { \ 546 /* spec defines specific handling for +/- inf and NaN values */ \ 547 _fromvtype val; \ 548 _tovtype intMin, intMax, result; \ 549 vdst = INST_A(inst); \ 550 vsrc1 = INST_B(inst); \ 551 ILOGV("|%s v%d,v%d", (_opname), vdst, vsrc1); \ 552 val = GET_REGISTER##_fromrtype(vsrc1); \ 553 intMin = (_tovtype) 1 << (sizeof(_tovtype) * 8 -1); \ 554 intMax = ~intMin; \ 555 result = (_tovtype) val; \ 556 if (val >= intMax) /* +inf */ \ 557 result = intMax; \ 558 else if (val <= intMin) /* -inf */ \ 559 result = intMin; \ 560 else if (val != val) /* NaN */ \ 561 result = 0; \ 562 else \ 563 result = (_tovtype) val; \ 564 SET_REGISTER##_tortype(vdst, result); \ 565 } \ 566 FINISH(1); 567 568#define HANDLE_INT_TO_SMALL(_opcode, _opname, _type) \ 569 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 570 vdst = INST_A(inst); \ 571 vsrc1 = INST_B(inst); \ 572 ILOGV("|int-to-%s v%d,v%d", (_opname), vdst, vsrc1); \ 573 SET_REGISTER(vdst, (_type) GET_REGISTER(vsrc1)); \ 574 FINISH(1); 575 576/* NOTE: the comparison result is always a signed 4-byte integer */ 577#define HANDLE_OP_CMPX(_opcode, _opname, _varType, _type, _nanVal) \ 578 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 579 { \ 580 int result; \ 581 u2 regs; \ 582 _varType val1, val2; \ 583 vdst = INST_AA(inst); \ 584 regs = FETCH(1); \ 585 vsrc1 = regs & 0xff; \ 586 vsrc2 = regs >> 8; \ 587 ILOGV("|cmp%s v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 588 val1 = GET_REGISTER##_type(vsrc1); \ 589 val2 = GET_REGISTER##_type(vsrc2); \ 590 if (val1 == val2) \ 591 result = 0; \ 592 else if (val1 < val2) \ 593 result = -1; \ 594 else if (val1 > val2) \ 595 result = 1; \ 596 else \ 597 result = (_nanVal); \ 598 ILOGV("+ result=%d", result); \ 599 SET_REGISTER(vdst, result); \ 600 } \ 601 FINISH(2); 602 603#define HANDLE_OP_IF_XX(_opcode, _opname, _cmp) \ 604 HANDLE_OPCODE(_opcode /*vA, vB, +CCCC*/) \ 605 vsrc1 = INST_A(inst); \ 606 vsrc2 = INST_B(inst); \ 607 if ((s4) GET_REGISTER(vsrc1) _cmp (s4) GET_REGISTER(vsrc2)) { \ 608 int branchOffset = (s2)FETCH(1); /* sign-extended */ \ 609 ILOGV("|if-%s v%d,v%d,+0x%04x", (_opname), vsrc1, vsrc2, \ 610 branchOffset); \ 611 ILOGV("> branch taken"); \ 612 if (branchOffset < 0) \ 613 PERIODIC_CHECKS(branchOffset); \ 614 FINISH(branchOffset); \ 615 } else { \ 616 ILOGV("|if-%s v%d,v%d,-", (_opname), vsrc1, vsrc2); \ 617 FINISH(2); \ 618 } 619 620#define HANDLE_OP_IF_XXZ(_opcode, _opname, _cmp) \ 621 HANDLE_OPCODE(_opcode /*vAA, +BBBB*/) \ 622 vsrc1 = INST_AA(inst); \ 623 if ((s4) GET_REGISTER(vsrc1) _cmp 0) { \ 624 int branchOffset = (s2)FETCH(1); /* sign-extended */ \ 625 ILOGV("|if-%s v%d,+0x%04x", (_opname), vsrc1, branchOffset); \ 626 ILOGV("> branch taken"); \ 627 if (branchOffset < 0) \ 628 PERIODIC_CHECKS(branchOffset); \ 629 FINISH(branchOffset); \ 630 } else { \ 631 ILOGV("|if-%s v%d,-", (_opname), vsrc1); \ 632 FINISH(2); \ 633 } 634 635#define HANDLE_UNOP(_opcode, _opname, _pfx, _sfx, _type) \ 636 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 637 vdst = INST_A(inst); \ 638 vsrc1 = INST_B(inst); \ 639 ILOGV("|%s v%d,v%d", (_opname), vdst, vsrc1); \ 640 SET_REGISTER##_type(vdst, _pfx GET_REGISTER##_type(vsrc1) _sfx); \ 641 FINISH(1); 642 643#define HANDLE_OP_X_INT(_opcode, _opname, _op, _chkdiv) \ 644 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 645 { \ 646 u2 srcRegs; \ 647 vdst = INST_AA(inst); \ 648 srcRegs = FETCH(1); \ 649 vsrc1 = srcRegs & 0xff; \ 650 vsrc2 = srcRegs >> 8; \ 651 ILOGV("|%s-int v%d,v%d", (_opname), vdst, vsrc1); \ 652 if (_chkdiv != 0) { \ 653 s4 firstVal, secondVal, result; \ 654 firstVal = GET_REGISTER(vsrc1); \ 655 secondVal = GET_REGISTER(vsrc2); \ 656 if (secondVal == 0) { \ 657 EXPORT_PC(); \ 658 dvmThrowArithmeticException("divide by zero"); \ 659 GOTO_exceptionThrown(); \ 660 } \ 661 if ((u4)firstVal == 0x80000000 && secondVal == -1) { \ 662 if (_chkdiv == 1) \ 663 result = firstVal; /* division */ \ 664 else \ 665 result = 0; /* remainder */ \ 666 } else { \ 667 result = firstVal _op secondVal; \ 668 } \ 669 SET_REGISTER(vdst, result); \ 670 } else { \ 671 /* non-div/rem case */ \ 672 SET_REGISTER(vdst, \ 673 (s4) GET_REGISTER(vsrc1) _op (s4) GET_REGISTER(vsrc2)); \ 674 } \ 675 } \ 676 FINISH(2); 677 678#define HANDLE_OP_SHX_INT(_opcode, _opname, _cast, _op) \ 679 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 680 { \ 681 u2 srcRegs; \ 682 vdst = INST_AA(inst); \ 683 srcRegs = FETCH(1); \ 684 vsrc1 = srcRegs & 0xff; \ 685 vsrc2 = srcRegs >> 8; \ 686 ILOGV("|%s-int v%d,v%d", (_opname), vdst, vsrc1); \ 687 SET_REGISTER(vdst, \ 688 _cast GET_REGISTER(vsrc1) _op (GET_REGISTER(vsrc2) & 0x1f)); \ 689 } \ 690 FINISH(2); 691 692#define HANDLE_OP_X_INT_LIT16(_opcode, _opname, _op, _chkdiv) \ 693 HANDLE_OPCODE(_opcode /*vA, vB, #+CCCC*/) \ 694 vdst = INST_A(inst); \ 695 vsrc1 = INST_B(inst); \ 696 vsrc2 = FETCH(1); \ 697 ILOGV("|%s-int/lit16 v%d,v%d,#+0x%04x", \ 698 (_opname), vdst, vsrc1, vsrc2); \ 699 if (_chkdiv != 0) { \ 700 s4 firstVal, result; \ 701 firstVal = GET_REGISTER(vsrc1); \ 702 if ((s2) vsrc2 == 0) { \ 703 EXPORT_PC(); \ 704 dvmThrowArithmeticException("divide by zero"); \ 705 GOTO_exceptionThrown(); \ 706 } \ 707 if ((u4)firstVal == 0x80000000 && ((s2) vsrc2) == -1) { \ 708 /* won't generate /lit16 instr for this; check anyway */ \ 709 if (_chkdiv == 1) \ 710 result = firstVal; /* division */ \ 711 else \ 712 result = 0; /* remainder */ \ 713 } else { \ 714 result = firstVal _op (s2) vsrc2; \ 715 } \ 716 SET_REGISTER(vdst, result); \ 717 } else { \ 718 /* non-div/rem case */ \ 719 SET_REGISTER(vdst, GET_REGISTER(vsrc1) _op (s2) vsrc2); \ 720 } \ 721 FINISH(2); 722 723#define HANDLE_OP_X_INT_LIT8(_opcode, _opname, _op, _chkdiv) \ 724 HANDLE_OPCODE(_opcode /*vAA, vBB, #+CC*/) \ 725 { \ 726 u2 litInfo; \ 727 vdst = INST_AA(inst); \ 728 litInfo = FETCH(1); \ 729 vsrc1 = litInfo & 0xff; \ 730 vsrc2 = litInfo >> 8; /* constant */ \ 731 ILOGV("|%s-int/lit8 v%d,v%d,#+0x%02x", \ 732 (_opname), vdst, vsrc1, vsrc2); \ 733 if (_chkdiv != 0) { \ 734 s4 firstVal, result; \ 735 firstVal = GET_REGISTER(vsrc1); \ 736 if ((s1) vsrc2 == 0) { \ 737 EXPORT_PC(); \ 738 dvmThrowArithmeticException("divide by zero"); \ 739 GOTO_exceptionThrown(); \ 740 } \ 741 if ((u4)firstVal == 0x80000000 && ((s1) vsrc2) == -1) { \ 742 if (_chkdiv == 1) \ 743 result = firstVal; /* division */ \ 744 else \ 745 result = 0; /* remainder */ \ 746 } else { \ 747 result = firstVal _op ((s1) vsrc2); \ 748 } \ 749 SET_REGISTER(vdst, result); \ 750 } else { \ 751 SET_REGISTER(vdst, \ 752 (s4) GET_REGISTER(vsrc1) _op (s1) vsrc2); \ 753 } \ 754 } \ 755 FINISH(2); 756 757#define HANDLE_OP_SHX_INT_LIT8(_opcode, _opname, _cast, _op) \ 758 HANDLE_OPCODE(_opcode /*vAA, vBB, #+CC*/) \ 759 { \ 760 u2 litInfo; \ 761 vdst = INST_AA(inst); \ 762 litInfo = FETCH(1); \ 763 vsrc1 = litInfo & 0xff; \ 764 vsrc2 = litInfo >> 8; /* constant */ \ 765 ILOGV("|%s-int/lit8 v%d,v%d,#+0x%02x", \ 766 (_opname), vdst, vsrc1, vsrc2); \ 767 SET_REGISTER(vdst, \ 768 _cast GET_REGISTER(vsrc1) _op (vsrc2 & 0x1f)); \ 769 } \ 770 FINISH(2); 771 772#define HANDLE_OP_X_INT_2ADDR(_opcode, _opname, _op, _chkdiv) \ 773 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 774 vdst = INST_A(inst); \ 775 vsrc1 = INST_B(inst); \ 776 ILOGV("|%s-int-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 777 if (_chkdiv != 0) { \ 778 s4 firstVal, secondVal, result; \ 779 firstVal = GET_REGISTER(vdst); \ 780 secondVal = GET_REGISTER(vsrc1); \ 781 if (secondVal == 0) { \ 782 EXPORT_PC(); \ 783 dvmThrowArithmeticException("divide by zero"); \ 784 GOTO_exceptionThrown(); \ 785 } \ 786 if ((u4)firstVal == 0x80000000 && secondVal == -1) { \ 787 if (_chkdiv == 1) \ 788 result = firstVal; /* division */ \ 789 else \ 790 result = 0; /* remainder */ \ 791 } else { \ 792 result = firstVal _op secondVal; \ 793 } \ 794 SET_REGISTER(vdst, result); \ 795 } else { \ 796 SET_REGISTER(vdst, \ 797 (s4) GET_REGISTER(vdst) _op (s4) GET_REGISTER(vsrc1)); \ 798 } \ 799 FINISH(1); 800 801#define HANDLE_OP_SHX_INT_2ADDR(_opcode, _opname, _cast, _op) \ 802 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 803 vdst = INST_A(inst); \ 804 vsrc1 = INST_B(inst); \ 805 ILOGV("|%s-int-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 806 SET_REGISTER(vdst, \ 807 _cast GET_REGISTER(vdst) _op (GET_REGISTER(vsrc1) & 0x1f)); \ 808 FINISH(1); 809 810#define HANDLE_OP_X_LONG(_opcode, _opname, _op, _chkdiv) \ 811 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 812 { \ 813 u2 srcRegs; \ 814 vdst = INST_AA(inst); \ 815 srcRegs = FETCH(1); \ 816 vsrc1 = srcRegs & 0xff; \ 817 vsrc2 = srcRegs >> 8; \ 818 ILOGV("|%s-long v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 819 if (_chkdiv != 0) { \ 820 s8 firstVal, secondVal, result; \ 821 firstVal = GET_REGISTER_WIDE(vsrc1); \ 822 secondVal = GET_REGISTER_WIDE(vsrc2); \ 823 if (secondVal == 0LL) { \ 824 EXPORT_PC(); \ 825 dvmThrowArithmeticException("divide by zero"); \ 826 GOTO_exceptionThrown(); \ 827 } \ 828 if ((u8)firstVal == 0x8000000000000000ULL && \ 829 secondVal == -1LL) \ 830 { \ 831 if (_chkdiv == 1) \ 832 result = firstVal; /* division */ \ 833 else \ 834 result = 0; /* remainder */ \ 835 } else { \ 836 result = firstVal _op secondVal; \ 837 } \ 838 SET_REGISTER_WIDE(vdst, result); \ 839 } else { \ 840 SET_REGISTER_WIDE(vdst, \ 841 (s8) GET_REGISTER_WIDE(vsrc1) _op (s8) GET_REGISTER_WIDE(vsrc2)); \ 842 } \ 843 } \ 844 FINISH(2); 845 846#define HANDLE_OP_SHX_LONG(_opcode, _opname, _cast, _op) \ 847 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 848 { \ 849 u2 srcRegs; \ 850 vdst = INST_AA(inst); \ 851 srcRegs = FETCH(1); \ 852 vsrc1 = srcRegs & 0xff; \ 853 vsrc2 = srcRegs >> 8; \ 854 ILOGV("|%s-long v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 855 SET_REGISTER_WIDE(vdst, \ 856 _cast GET_REGISTER_WIDE(vsrc1) _op (GET_REGISTER(vsrc2) & 0x3f)); \ 857 } \ 858 FINISH(2); 859 860#define HANDLE_OP_X_LONG_2ADDR(_opcode, _opname, _op, _chkdiv) \ 861 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 862 vdst = INST_A(inst); \ 863 vsrc1 = INST_B(inst); \ 864 ILOGV("|%s-long-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 865 if (_chkdiv != 0) { \ 866 s8 firstVal, secondVal, result; \ 867 firstVal = GET_REGISTER_WIDE(vdst); \ 868 secondVal = GET_REGISTER_WIDE(vsrc1); \ 869 if (secondVal == 0LL) { \ 870 EXPORT_PC(); \ 871 dvmThrowArithmeticException("divide by zero"); \ 872 GOTO_exceptionThrown(); \ 873 } \ 874 if ((u8)firstVal == 0x8000000000000000ULL && \ 875 secondVal == -1LL) \ 876 { \ 877 if (_chkdiv == 1) \ 878 result = firstVal; /* division */ \ 879 else \ 880 result = 0; /* remainder */ \ 881 } else { \ 882 result = firstVal _op secondVal; \ 883 } \ 884 SET_REGISTER_WIDE(vdst, result); \ 885 } else { \ 886 SET_REGISTER_WIDE(vdst, \ 887 (s8) GET_REGISTER_WIDE(vdst) _op (s8)GET_REGISTER_WIDE(vsrc1));\ 888 } \ 889 FINISH(1); 890 891#define HANDLE_OP_SHX_LONG_2ADDR(_opcode, _opname, _cast, _op) \ 892 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 893 vdst = INST_A(inst); \ 894 vsrc1 = INST_B(inst); \ 895 ILOGV("|%s-long-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 896 SET_REGISTER_WIDE(vdst, \ 897 _cast GET_REGISTER_WIDE(vdst) _op (GET_REGISTER(vsrc1) & 0x3f)); \ 898 FINISH(1); 899 900#define HANDLE_OP_X_FLOAT(_opcode, _opname, _op) \ 901 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 902 { \ 903 u2 srcRegs; \ 904 vdst = INST_AA(inst); \ 905 srcRegs = FETCH(1); \ 906 vsrc1 = srcRegs & 0xff; \ 907 vsrc2 = srcRegs >> 8; \ 908 ILOGV("|%s-float v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 909 SET_REGISTER_FLOAT(vdst, \ 910 GET_REGISTER_FLOAT(vsrc1) _op GET_REGISTER_FLOAT(vsrc2)); \ 911 } \ 912 FINISH(2); 913 914#define HANDLE_OP_X_DOUBLE(_opcode, _opname, _op) \ 915 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 916 { \ 917 u2 srcRegs; \ 918 vdst = INST_AA(inst); \ 919 srcRegs = FETCH(1); \ 920 vsrc1 = srcRegs & 0xff; \ 921 vsrc2 = srcRegs >> 8; \ 922 ILOGV("|%s-double v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 923 SET_REGISTER_DOUBLE(vdst, \ 924 GET_REGISTER_DOUBLE(vsrc1) _op GET_REGISTER_DOUBLE(vsrc2)); \ 925 } \ 926 FINISH(2); 927 928#define HANDLE_OP_X_FLOAT_2ADDR(_opcode, _opname, _op) \ 929 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 930 vdst = INST_A(inst); \ 931 vsrc1 = INST_B(inst); \ 932 ILOGV("|%s-float-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 933 SET_REGISTER_FLOAT(vdst, \ 934 GET_REGISTER_FLOAT(vdst) _op GET_REGISTER_FLOAT(vsrc1)); \ 935 FINISH(1); 936 937#define HANDLE_OP_X_DOUBLE_2ADDR(_opcode, _opname, _op) \ 938 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 939 vdst = INST_A(inst); \ 940 vsrc1 = INST_B(inst); \ 941 ILOGV("|%s-double-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 942 SET_REGISTER_DOUBLE(vdst, \ 943 GET_REGISTER_DOUBLE(vdst) _op GET_REGISTER_DOUBLE(vsrc1)); \ 944 FINISH(1); 945 946#define HANDLE_OP_AGET(_opcode, _opname, _type, _regsize) \ 947 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 948 { \ 949 ArrayObject* arrayObj; \ 950 u2 arrayInfo; \ 951 EXPORT_PC(); \ 952 vdst = INST_AA(inst); \ 953 arrayInfo = FETCH(1); \ 954 vsrc1 = arrayInfo & 0xff; /* array ptr */ \ 955 vsrc2 = arrayInfo >> 8; /* index */ \ 956 ILOGV("|aget%s v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 957 arrayObj = (ArrayObject*) GET_REGISTER(vsrc1); \ 958 if (!checkForNull((Object*) arrayObj)) \ 959 GOTO_exceptionThrown(); \ 960 if (GET_REGISTER(vsrc2) >= arrayObj->length) { \ 961 dvmThrowArrayIndexOutOfBoundsException( \ 962 arrayObj->length, GET_REGISTER(vsrc2)); \ 963 GOTO_exceptionThrown(); \ 964 } \ 965 SET_REGISTER##_regsize(vdst, \ 966 ((_type*)(void*)arrayObj->contents)[GET_REGISTER(vsrc2)]); \ 967 ILOGV("+ AGET[%d]=0x%x", GET_REGISTER(vsrc2), GET_REGISTER(vdst)); \ 968 } \ 969 FINISH(2); 970 971#define HANDLE_OP_APUT(_opcode, _opname, _type, _regsize) \ 972 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 973 { \ 974 ArrayObject* arrayObj; \ 975 u2 arrayInfo; \ 976 EXPORT_PC(); \ 977 vdst = INST_AA(inst); /* AA: source value */ \ 978 arrayInfo = FETCH(1); \ 979 vsrc1 = arrayInfo & 0xff; /* BB: array ptr */ \ 980 vsrc2 = arrayInfo >> 8; /* CC: index */ \ 981 ILOGV("|aput%s v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 982 arrayObj = (ArrayObject*) GET_REGISTER(vsrc1); \ 983 if (!checkForNull((Object*) arrayObj)) \ 984 GOTO_exceptionThrown(); \ 985 if (GET_REGISTER(vsrc2) >= arrayObj->length) { \ 986 dvmThrowArrayIndexOutOfBoundsException( \ 987 arrayObj->length, GET_REGISTER(vsrc2)); \ 988 GOTO_exceptionThrown(); \ 989 } \ 990 ILOGV("+ APUT[%d]=0x%08x", GET_REGISTER(vsrc2), GET_REGISTER(vdst));\ 991 ((_type*)(void*)arrayObj->contents)[GET_REGISTER(vsrc2)] = \ 992 GET_REGISTER##_regsize(vdst); \ 993 } \ 994 FINISH(2); 995 996/* 997 * It's possible to get a bad value out of a field with sub-32-bit stores 998 * because the -quick versions always operate on 32 bits. Consider: 999 * short foo = -1 (sets a 32-bit register to 0xffffffff) 1000 * iput-quick foo (writes all 32 bits to the field) 1001 * short bar = 1 (sets a 32-bit register to 0x00000001) 1002 * iput-short (writes the low 16 bits to the field) 1003 * iget-quick foo (reads all 32 bits from the field, yielding 0xffff0001) 1004 * This can only happen when optimized and non-optimized code has interleaved 1005 * access to the same field. This is unlikely but possible. 1006 * 1007 * The easiest way to fix this is to always read/write 32 bits at a time. On 1008 * a device with a 16-bit data bus this is sub-optimal. (The alternative 1009 * approach is to have sub-int versions of iget-quick, but now we're wasting 1010 * Dalvik instruction space and making it less likely that handler code will 1011 * already be in the CPU i-cache.) 1012 */ 1013#define HANDLE_IGET_X(_opcode, _opname, _ftype, _regsize) \ 1014 HANDLE_OPCODE(_opcode /*vA, vB, field@CCCC*/) \ 1015 { \ 1016 InstField* ifield; \ 1017 Object* obj; \ 1018 EXPORT_PC(); \ 1019 vdst = INST_A(inst); \ 1020 vsrc1 = INST_B(inst); /* object ptr */ \ 1021 ref = FETCH(1); /* field ref */ \ 1022 ILOGV("|iget%s v%d,v%d,field@0x%04x", (_opname), vdst, vsrc1, ref); \ 1023 obj = (Object*) GET_REGISTER(vsrc1); \ 1024 if (!checkForNull(obj)) \ 1025 GOTO_exceptionThrown(); \ 1026 ifield = (InstField*) dvmDexGetResolvedField(methodClassDex, ref); \ 1027 if (ifield == NULL) { \ 1028 ifield = dvmResolveInstField(curMethod->clazz, ref); \ 1029 if (ifield == NULL) \ 1030 GOTO_exceptionThrown(); \ 1031 } \ 1032 SET_REGISTER##_regsize(vdst, \ 1033 dvmGetField##_ftype(obj, ifield->byteOffset)); \ 1034 ILOGV("+ IGET '%s'=0x%08llx", ifield->field.name, \ 1035 (u8) GET_REGISTER##_regsize(vdst)); \ 1036 } \ 1037 FINISH(2); 1038 1039#define HANDLE_IGET_X_JUMBO(_opcode, _opname, _ftype, _regsize) \ 1040 HANDLE_OPCODE(_opcode /*vBBBB, vCCCC, class@AAAAAAAA*/) \ 1041 { \ 1042 InstField* ifield; \ 1043 Object* obj; \ 1044 EXPORT_PC(); \ 1045 ref = FETCH(1) | (u4)FETCH(2) << 16; /* field ref */ \ 1046 vdst = FETCH(3); \ 1047 vsrc1 = FETCH(4); /* object ptr */ \ 1048 ILOGV("|iget%s/jumbo v%d,v%d,field@0x%08x", \ 1049 (_opname), vdst, vsrc1, ref); \ 1050 obj = (Object*) GET_REGISTER(vsrc1); \ 1051 if (!checkForNull(obj)) \ 1052 GOTO_exceptionThrown(); \ 1053 ifield = (InstField*) dvmDexGetResolvedField(methodClassDex, ref); \ 1054 if (ifield == NULL) { \ 1055 ifield = dvmResolveInstField(curMethod->clazz, ref); \ 1056 if (ifield == NULL) \ 1057 GOTO_exceptionThrown(); \ 1058 } \ 1059 SET_REGISTER##_regsize(vdst, \ 1060 dvmGetField##_ftype(obj, ifield->byteOffset)); \ 1061 ILOGV("+ IGET '%s'=0x%08llx", ifield->field.name, \ 1062 (u8) GET_REGISTER##_regsize(vdst)); \ 1063 } \ 1064 FINISH(5); 1065 1066#define HANDLE_IGET_X_QUICK(_opcode, _opname, _ftype, _regsize) \ 1067 HANDLE_OPCODE(_opcode /*vA, vB, field@CCCC*/) \ 1068 { \ 1069 Object* obj; \ 1070 vdst = INST_A(inst); \ 1071 vsrc1 = INST_B(inst); /* object ptr */ \ 1072 ref = FETCH(1); /* field offset */ \ 1073 ILOGV("|iget%s-quick v%d,v%d,field@+%u", \ 1074 (_opname), vdst, vsrc1, ref); \ 1075 obj = (Object*) GET_REGISTER(vsrc1); \ 1076 if (!checkForNullExportPC(obj, fp, pc)) \ 1077 GOTO_exceptionThrown(); \ 1078 SET_REGISTER##_regsize(vdst, dvmGetField##_ftype(obj, ref)); \ 1079 ILOGV("+ IGETQ %d=0x%08llx", ref, \ 1080 (u8) GET_REGISTER##_regsize(vdst)); \ 1081 } \ 1082 FINISH(2); 1083 1084#define HANDLE_IPUT_X(_opcode, _opname, _ftype, _regsize) \ 1085 HANDLE_OPCODE(_opcode /*vA, vB, field@CCCC*/) \ 1086 { \ 1087 InstField* ifield; \ 1088 Object* obj; \ 1089 EXPORT_PC(); \ 1090 vdst = INST_A(inst); \ 1091 vsrc1 = INST_B(inst); /* object ptr */ \ 1092 ref = FETCH(1); /* field ref */ \ 1093 ILOGV("|iput%s v%d,v%d,field@0x%04x", (_opname), vdst, vsrc1, ref); \ 1094 obj = (Object*) GET_REGISTER(vsrc1); \ 1095 if (!checkForNull(obj)) \ 1096 GOTO_exceptionThrown(); \ 1097 ifield = (InstField*) dvmDexGetResolvedField(methodClassDex, ref); \ 1098 if (ifield == NULL) { \ 1099 ifield = dvmResolveInstField(curMethod->clazz, ref); \ 1100 if (ifield == NULL) \ 1101 GOTO_exceptionThrown(); \ 1102 } \ 1103 dvmSetField##_ftype(obj, ifield->byteOffset, \ 1104 GET_REGISTER##_regsize(vdst)); \ 1105 ILOGV("+ IPUT '%s'=0x%08llx", ifield->field.name, \ 1106 (u8) GET_REGISTER##_regsize(vdst)); \ 1107 } \ 1108 FINISH(2); 1109 1110#define HANDLE_IPUT_X_JUMBO(_opcode, _opname, _ftype, _regsize) \ 1111 HANDLE_OPCODE(_opcode /*vBBBB, vCCCC, class@AAAAAAAA*/) \ 1112 { \ 1113 InstField* ifield; \ 1114 Object* obj; \ 1115 EXPORT_PC(); \ 1116 ref = FETCH(1) | (u4)FETCH(2) << 16; /* field ref */ \ 1117 vdst = FETCH(3); \ 1118 vsrc1 = FETCH(4); /* object ptr */ \ 1119 ILOGV("|iput%s/jumbo v%d,v%d,field@0x%08x", \ 1120 (_opname), vdst, vsrc1, ref); \ 1121 obj = (Object*) GET_REGISTER(vsrc1); \ 1122 if (!checkForNull(obj)) \ 1123 GOTO_exceptionThrown(); \ 1124 ifield = (InstField*) dvmDexGetResolvedField(methodClassDex, ref); \ 1125 if (ifield == NULL) { \ 1126 ifield = dvmResolveInstField(curMethod->clazz, ref); \ 1127 if (ifield == NULL) \ 1128 GOTO_exceptionThrown(); \ 1129 } \ 1130 dvmSetField##_ftype(obj, ifield->byteOffset, \ 1131 GET_REGISTER##_regsize(vdst)); \ 1132 ILOGV("+ IPUT '%s'=0x%08llx", ifield->field.name, \ 1133 (u8) GET_REGISTER##_regsize(vdst)); \ 1134 } \ 1135 FINISH(5); 1136 1137#define HANDLE_IPUT_X_QUICK(_opcode, _opname, _ftype, _regsize) \ 1138 HANDLE_OPCODE(_opcode /*vA, vB, field@CCCC*/) \ 1139 { \ 1140 Object* obj; \ 1141 vdst = INST_A(inst); \ 1142 vsrc1 = INST_B(inst); /* object ptr */ \ 1143 ref = FETCH(1); /* field offset */ \ 1144 ILOGV("|iput%s-quick v%d,v%d,field@0x%04x", \ 1145 (_opname), vdst, vsrc1, ref); \ 1146 obj = (Object*) GET_REGISTER(vsrc1); \ 1147 if (!checkForNullExportPC(obj, fp, pc)) \ 1148 GOTO_exceptionThrown(); \ 1149 dvmSetField##_ftype(obj, ref, GET_REGISTER##_regsize(vdst)); \ 1150 ILOGV("+ IPUTQ %d=0x%08llx", ref, \ 1151 (u8) GET_REGISTER##_regsize(vdst)); \ 1152 } \ 1153 FINISH(2); 1154 1155/* 1156 * The JIT needs dvmDexGetResolvedField() to return non-null. 1157 * Because the portable interpreter is not involved with the JIT 1158 * and trace building, we only need the extra check here when this 1159 * code is massaged into a stub called from an assembly interpreter. 1160 * This is controlled by the JIT_STUB_HACK maco. 1161 */ 1162 1163#define HANDLE_SGET_X(_opcode, _opname, _ftype, _regsize) \ 1164 HANDLE_OPCODE(_opcode /*vAA, field@BBBB*/) \ 1165 { \ 1166 StaticField* sfield; \ 1167 vdst = INST_AA(inst); \ 1168 ref = FETCH(1); /* field ref */ \ 1169 ILOGV("|sget%s v%d,sfield@0x%04x", (_opname), vdst, ref); \ 1170 sfield = (StaticField*)dvmDexGetResolvedField(methodClassDex, ref); \ 1171 if (sfield == NULL) { \ 1172 EXPORT_PC(); \ 1173 sfield = dvmResolveStaticField(curMethod->clazz, ref); \ 1174 if (sfield == NULL) \ 1175 GOTO_exceptionThrown(); \ 1176 if (dvmDexGetResolvedField(methodClassDex, ref) == NULL) { \ 1177 JIT_STUB_HACK(dvmJitEndTraceSelect(self,pc)); \ 1178 } \ 1179 } \ 1180 SET_REGISTER##_regsize(vdst, dvmGetStaticField##_ftype(sfield)); \ 1181 ILOGV("+ SGET '%s'=0x%08llx", \ 1182 sfield->field.name, (u8)GET_REGISTER##_regsize(vdst)); \ 1183 } \ 1184 FINISH(2); 1185 1186#define HANDLE_SGET_X_JUMBO(_opcode, _opname, _ftype, _regsize) \ 1187 HANDLE_OPCODE(_opcode /*vBBBB, class@AAAAAAAA*/) \ 1188 { \ 1189 StaticField* sfield; \ 1190 ref = FETCH(1) | (u4)FETCH(2) << 16; /* field ref */ \ 1191 vdst = FETCH(3); \ 1192 ILOGV("|sget%s/jumbo v%d,sfield@0x%08x", (_opname), vdst, ref); \ 1193 sfield = (StaticField*)dvmDexGetResolvedField(methodClassDex, ref); \ 1194 if (sfield == NULL) { \ 1195 EXPORT_PC(); \ 1196 sfield = dvmResolveStaticField(curMethod->clazz, ref); \ 1197 if (sfield == NULL) \ 1198 GOTO_exceptionThrown(); \ 1199 if (dvmDexGetResolvedField(methodClassDex, ref) == NULL) { \ 1200 JIT_STUB_HACK(dvmJitEndTraceSelect(self,pc)); \ 1201 } \ 1202 } \ 1203 SET_REGISTER##_regsize(vdst, dvmGetStaticField##_ftype(sfield)); \ 1204 ILOGV("+ SGET '%s'=0x%08llx", \ 1205 sfield->field.name, (u8)GET_REGISTER##_regsize(vdst)); \ 1206 } \ 1207 FINISH(4); 1208 1209#define HANDLE_SPUT_X(_opcode, _opname, _ftype, _regsize) \ 1210 HANDLE_OPCODE(_opcode /*vAA, field@BBBB*/) \ 1211 { \ 1212 StaticField* sfield; \ 1213 vdst = INST_AA(inst); \ 1214 ref = FETCH(1); /* field ref */ \ 1215 ILOGV("|sput%s v%d,sfield@0x%04x", (_opname), vdst, ref); \ 1216 sfield = (StaticField*)dvmDexGetResolvedField(methodClassDex, ref); \ 1217 if (sfield == NULL) { \ 1218 EXPORT_PC(); \ 1219 sfield = dvmResolveStaticField(curMethod->clazz, ref); \ 1220 if (sfield == NULL) \ 1221 GOTO_exceptionThrown(); \ 1222 if (dvmDexGetResolvedField(methodClassDex, ref) == NULL) { \ 1223 JIT_STUB_HACK(dvmJitEndTraceSelect(self,pc)); \ 1224 } \ 1225 } \ 1226 dvmSetStaticField##_ftype(sfield, GET_REGISTER##_regsize(vdst)); \ 1227 ILOGV("+ SPUT '%s'=0x%08llx", \ 1228 sfield->field.name, (u8)GET_REGISTER##_regsize(vdst)); \ 1229 } \ 1230 FINISH(2); 1231 1232#define HANDLE_SPUT_X_JUMBO(_opcode, _opname, _ftype, _regsize) \ 1233 HANDLE_OPCODE(_opcode /*vBBBB, class@AAAAAAAA*/) \ 1234 { \ 1235 StaticField* sfield; \ 1236 ref = FETCH(1) | (u4)FETCH(2) << 16; /* field ref */ \ 1237 vdst = FETCH(3); \ 1238 ILOGV("|sput%s/jumbo v%d,sfield@0x%08x", (_opname), vdst, ref); \ 1239 sfield = (StaticField*)dvmDexGetResolvedField(methodClassDex, ref); \ 1240 if (sfield == NULL) { \ 1241 EXPORT_PC(); \ 1242 sfield = dvmResolveStaticField(curMethod->clazz, ref); \ 1243 if (sfield == NULL) \ 1244 GOTO_exceptionThrown(); \ 1245 if (dvmDexGetResolvedField(methodClassDex, ref) == NULL) { \ 1246 JIT_STUB_HACK(dvmJitEndTraceSelect(self,pc)); \ 1247 } \ 1248 } \ 1249 dvmSetStaticField##_ftype(sfield, GET_REGISTER##_regsize(vdst)); \ 1250 ILOGV("+ SPUT '%s'=0x%08llx", \ 1251 sfield->field.name, (u8)GET_REGISTER##_regsize(vdst)); \ 1252 } \ 1253 FINISH(4); 1254 1255/* File: c/OP_IGET_WIDE_VOLATILE.cpp */ 1256HANDLE_IGET_X(OP_IGET_WIDE_VOLATILE, "-wide-volatile", LongVolatile, _WIDE) 1257OP_END 1258 1259/* File: c/OP_IPUT_WIDE_VOLATILE.cpp */ 1260HANDLE_IPUT_X(OP_IPUT_WIDE_VOLATILE, "-wide-volatile", LongVolatile, _WIDE) 1261OP_END 1262 1263/* File: c/OP_SGET_WIDE_VOLATILE.cpp */ 1264HANDLE_SGET_X(OP_SGET_WIDE_VOLATILE, "-wide-volatile", LongVolatile, _WIDE) 1265OP_END 1266 1267/* File: c/OP_SPUT_WIDE_VOLATILE.cpp */ 1268HANDLE_SPUT_X(OP_SPUT_WIDE_VOLATILE, "-wide-volatile", LongVolatile, _WIDE) 1269OP_END 1270 1271/* File: c/OP_EXECUTE_INLINE_RANGE.cpp */ 1272HANDLE_OPCODE(OP_EXECUTE_INLINE_RANGE /*{vCCCC..v(CCCC+AA-1)}, inline@BBBB*/) 1273 { 1274 u4 arg0, arg1, arg2, arg3; 1275 arg0 = arg1 = arg2 = arg3 = 0; /* placate gcc */ 1276 1277 EXPORT_PC(); 1278 1279 vsrc1 = INST_AA(inst); /* #of args */ 1280 ref = FETCH(1); /* inline call "ref" */ 1281 vdst = FETCH(2); /* range base */ 1282 ILOGV("|execute-inline-range args=%d @%d {regs=v%d-v%d}", 1283 vsrc1, ref, vdst, vdst+vsrc1-1); 1284 1285 assert((vdst >> 16) == 0); // 16-bit type -or- high 16 bits clear 1286 assert(vsrc1 <= 4); 1287 1288 switch (vsrc1) { 1289 case 4: 1290 arg3 = GET_REGISTER(vdst+3); 1291 /* fall through */ 1292 case 3: 1293 arg2 = GET_REGISTER(vdst+2); 1294 /* fall through */ 1295 case 2: 1296 arg1 = GET_REGISTER(vdst+1); 1297 /* fall through */ 1298 case 1: 1299 arg0 = GET_REGISTER(vdst+0); 1300 /* fall through */ 1301 default: // case 0 1302 ; 1303 } 1304 1305 if (self->interpBreak.ctl.subMode & kSubModeDebuggerActive) { 1306 if (!dvmPerformInlineOp4Dbg(arg0, arg1, arg2, arg3, &retval, ref)) 1307 GOTO_exceptionThrown(); 1308 } else { 1309 if (!dvmPerformInlineOp4Std(arg0, arg1, arg2, arg3, &retval, ref)) 1310 GOTO_exceptionThrown(); 1311 } 1312 } 1313 FINISH(3); 1314OP_END 1315 1316/* File: c/OP_INVOKE_OBJECT_INIT_RANGE.cpp */ 1317HANDLE_OPCODE(OP_INVOKE_OBJECT_INIT_RANGE /*{vCCCC..v(CCCC+AA-1)}, meth@BBBB*/) 1318 { 1319 Object* obj; 1320 1321 vsrc1 = FETCH(2); /* reg number of "this" pointer */ 1322 obj = GET_REGISTER_AS_OBJECT(vsrc1); 1323 1324 if (!checkForNullExportPC(obj, fp, pc)) 1325 GOTO_exceptionThrown(); 1326 1327 /* 1328 * The object should be marked "finalizable" when Object.<init> 1329 * completes normally. We're going to assume it does complete 1330 * (by virtue of being nothing but a return-void) and set it now. 1331 */ 1332 if (IS_CLASS_FLAG_SET(obj->clazz, CLASS_ISFINALIZABLE)) { 1333 EXPORT_PC(); 1334 dvmSetFinalizable(obj); 1335 if (dvmGetException(self)) 1336 GOTO_exceptionThrown(); 1337 } 1338 1339 if (self->interpBreak.ctl.subMode & kSubModeDebuggerActive) { 1340 /* behave like OP_INVOKE_DIRECT_RANGE */ 1341 GOTO_invoke(invokeDirect, true, false); 1342 } 1343 FINISH(3); 1344 } 1345OP_END 1346 1347/* File: c/OP_RETURN_VOID_BARRIER.cpp */ 1348HANDLE_OPCODE(OP_RETURN_VOID_BARRIER /**/) 1349 ILOGV("|return-void"); 1350#ifndef NDEBUG 1351 retval.j = 0xababababULL; /* placate valgrind */ 1352#endif 1353 ANDROID_MEMBAR_STORE(); 1354 GOTO_returnFromMethod(); 1355OP_END 1356 1357/* File: c/OP_INVOKE_OBJECT_INIT_JUMBO.cpp */ 1358HANDLE_OPCODE(OP_INVOKE_OBJECT_INIT_JUMBO /*{vCCCC..vNNNN}, meth@AAAAAAAA*/) 1359 { 1360 Object* obj; 1361 1362 vsrc1 = FETCH(4); /* reg number of "this" pointer */ 1363 obj = GET_REGISTER_AS_OBJECT(vsrc1); 1364 1365 if (!checkForNullExportPC(obj, fp, pc)) 1366 GOTO_exceptionThrown(); 1367 1368 /* 1369 * The object should be marked "finalizable" when Object.<init> 1370 * completes normally. We're going to assume it does complete 1371 * (by virtue of being nothing but a return-void) and set it now. 1372 */ 1373 if (IS_CLASS_FLAG_SET(obj->clazz, CLASS_ISFINALIZABLE)) { 1374 EXPORT_PC(); 1375 dvmSetFinalizable(obj); 1376 if (dvmGetException(self)) 1377 GOTO_exceptionThrown(); 1378 } 1379 1380 if (self->interpBreak.ctl.subMode & kSubModeDebuggerActive) { 1381 /* behave like OP_INVOKE_DIRECT_RANGE */ 1382 GOTO_invoke(invokeDirect, true, true); 1383 } 1384 FINISH(5); 1385 } 1386OP_END 1387 1388/* File: c/OP_IGET_VOLATILE_JUMBO.cpp */ 1389HANDLE_IGET_X_JUMBO(OP_IGET_VOLATILE_JUMBO, "-volatile/jumbo", IntVolatile, ) 1390OP_END 1391 1392/* File: c/OP_IGET_WIDE_VOLATILE_JUMBO.cpp */ 1393HANDLE_IGET_X_JUMBO(OP_IGET_WIDE_VOLATILE_JUMBO, "-wide-volatile/jumbo", LongVolatile, _WIDE) 1394OP_END 1395 1396/* File: c/OP_IGET_OBJECT_VOLATILE_JUMBO.cpp */ 1397HANDLE_IGET_X_JUMBO(OP_IGET_OBJECT_VOLATILE_JUMBO, "-object-volatile/jumbo", ObjectVolatile, _AS_OBJECT) 1398OP_END 1399 1400/* File: c/OP_IPUT_VOLATILE_JUMBO.cpp */ 1401HANDLE_IPUT_X_JUMBO(OP_IPUT_VOLATILE_JUMBO, "-volatile/jumbo", IntVolatile, ) 1402OP_END 1403 1404/* File: c/OP_IPUT_WIDE_VOLATILE_JUMBO.cpp */ 1405HANDLE_IPUT_X_JUMBO(OP_IPUT_WIDE_VOLATILE_JUMBO, "-wide-volatile/jumbo", LongVolatile, _WIDE) 1406OP_END 1407 1408/* File: c/OP_IPUT_OBJECT_VOLATILE_JUMBO.cpp */ 1409HANDLE_IPUT_X_JUMBO(OP_IPUT_OBJECT_VOLATILE_JUMBO, "-object-volatile/jumbo", ObjectVolatile, _AS_OBJECT) 1410OP_END 1411 1412/* File: c/OP_SGET_VOLATILE_JUMBO.cpp */ 1413HANDLE_SGET_X_JUMBO(OP_SGET_VOLATILE_JUMBO, "-volatile/jumbo", IntVolatile, ) 1414OP_END 1415 1416/* File: c/OP_SGET_WIDE_VOLATILE_JUMBO.cpp */ 1417HANDLE_SGET_X_JUMBO(OP_SGET_WIDE_VOLATILE_JUMBO, "-wide-volatile/jumbo", LongVolatile, _WIDE) 1418OP_END 1419 1420/* File: c/OP_SGET_OBJECT_VOLATILE_JUMBO.cpp */ 1421HANDLE_SGET_X_JUMBO(OP_SGET_OBJECT_VOLATILE_JUMBO, "-object-volatile/jumbo", ObjectVolatile, _AS_OBJECT) 1422OP_END 1423 1424/* File: c/OP_SPUT_VOLATILE_JUMBO.cpp */ 1425HANDLE_SPUT_X_JUMBO(OP_SPUT_VOLATILE_JUMBO, "-volatile", IntVolatile, ) 1426OP_END 1427 1428/* File: c/OP_SPUT_WIDE_VOLATILE_JUMBO.cpp */ 1429HANDLE_SPUT_X_JUMBO(OP_SPUT_WIDE_VOLATILE_JUMBO, "-wide-volatile/jumbo", LongVolatile, _WIDE) 1430OP_END 1431 1432/* File: c/OP_SPUT_OBJECT_VOLATILE_JUMBO.cpp */ 1433HANDLE_SPUT_X_JUMBO(OP_SPUT_OBJECT_VOLATILE_JUMBO, "-object-volatile/jumbo", ObjectVolatile, _AS_OBJECT) 1434OP_END 1435 1436/* File: c/gotoTargets.cpp */ 1437/* 1438 * C footer. This has some common code shared by the various targets. 1439 */ 1440 1441/* 1442 * Everything from here on is a "goto target". In the basic interpreter 1443 * we jump into these targets and then jump directly to the handler for 1444 * next instruction. Here, these are subroutines that return to the caller. 1445 */ 1446 1447GOTO_TARGET(filledNewArray, bool methodCallRange, bool jumboFormat) 1448 { 1449 ClassObject* arrayClass; 1450 ArrayObject* newArray; 1451 u4* contents; 1452 char typeCh; 1453 int i; 1454 u4 arg5; 1455 1456 EXPORT_PC(); 1457 1458 if (jumboFormat) { 1459 ref = FETCH(1) | (u4)FETCH(2) << 16; /* class ref */ 1460 vsrc1 = FETCH(3); /* #of elements */ 1461 vdst = FETCH(4); /* range base */ 1462 arg5 = -1; /* silence compiler warning */ 1463 ILOGV("|filled-new-array/jumbo args=%d @0x%08x {regs=v%d-v%d}", 1464 vsrc1, ref, vdst, vdst+vsrc1-1); 1465 } else { 1466 ref = FETCH(1); /* class ref */ 1467 vdst = FETCH(2); /* first 4 regs -or- range base */ 1468 1469 if (methodCallRange) { 1470 vsrc1 = INST_AA(inst); /* #of elements */ 1471 arg5 = -1; /* silence compiler warning */ 1472 ILOGV("|filled-new-array-range args=%d @0x%04x {regs=v%d-v%d}", 1473 vsrc1, ref, vdst, vdst+vsrc1-1); 1474 } else { 1475 arg5 = INST_A(inst); 1476 vsrc1 = INST_B(inst); /* #of elements */ 1477 ILOGV("|filled-new-array args=%d @0x%04x {regs=0x%04x %x}", 1478 vsrc1, ref, vdst, arg5); 1479 } 1480 } 1481 1482 /* 1483 * Resolve the array class. 1484 */ 1485 arrayClass = dvmDexGetResolvedClass(methodClassDex, ref); 1486 if (arrayClass == NULL) { 1487 arrayClass = dvmResolveClass(curMethod->clazz, ref, false); 1488 if (arrayClass == NULL) 1489 GOTO_exceptionThrown(); 1490 } 1491 /* 1492 if (!dvmIsArrayClass(arrayClass)) { 1493 dvmThrowRuntimeException( 1494 "filled-new-array needs array class"); 1495 GOTO_exceptionThrown(); 1496 } 1497 */ 1498 /* verifier guarantees this is an array class */ 1499 assert(dvmIsArrayClass(arrayClass)); 1500 assert(dvmIsClassInitialized(arrayClass)); 1501 1502 /* 1503 * Create an array of the specified type. 1504 */ 1505 LOGVV("+++ filled-new-array type is '%s'", arrayClass->descriptor); 1506 typeCh = arrayClass->descriptor[1]; 1507 if (typeCh == 'D' || typeCh == 'J') { 1508 /* category 2 primitives not allowed */ 1509 dvmThrowRuntimeException("bad filled array req"); 1510 GOTO_exceptionThrown(); 1511 } else if (typeCh != 'L' && typeCh != '[' && typeCh != 'I') { 1512 /* TODO: requires multiple "fill in" loops with different widths */ 1513 LOGE("non-int primitives not implemented"); 1514 dvmThrowInternalError( 1515 "filled-new-array not implemented for anything but 'int'"); 1516 GOTO_exceptionThrown(); 1517 } 1518 1519 newArray = dvmAllocArrayByClass(arrayClass, vsrc1, ALLOC_DONT_TRACK); 1520 if (newArray == NULL) 1521 GOTO_exceptionThrown(); 1522 1523 /* 1524 * Fill in the elements. It's legal for vsrc1 to be zero. 1525 */ 1526 contents = (u4*)(void*)newArray->contents; 1527 if (methodCallRange) { 1528 for (i = 0; i < vsrc1; i++) 1529 contents[i] = GET_REGISTER(vdst+i); 1530 } else { 1531 assert(vsrc1 <= 5); 1532 if (vsrc1 == 5) { 1533 contents[4] = GET_REGISTER(arg5); 1534 vsrc1--; 1535 } 1536 for (i = 0; i < vsrc1; i++) { 1537 contents[i] = GET_REGISTER(vdst & 0x0f); 1538 vdst >>= 4; 1539 } 1540 } 1541 if (typeCh == 'L' || typeCh == '[') { 1542 dvmWriteBarrierArray(newArray, 0, newArray->length); 1543 } 1544 1545 retval.l = (Object*)newArray; 1546 } 1547 if (jumboFormat) { 1548 FINISH(5); 1549 } else { 1550 FINISH(3); 1551 } 1552GOTO_TARGET_END 1553 1554 1555GOTO_TARGET(invokeVirtual, bool methodCallRange, bool jumboFormat) 1556 { 1557 Method* baseMethod; 1558 Object* thisPtr; 1559 1560 EXPORT_PC(); 1561 1562 if (jumboFormat) { 1563 ref = FETCH(1) | (u4)FETCH(2) << 16; /* method ref */ 1564 vsrc1 = FETCH(3); /* count */ 1565 vdst = FETCH(4); /* first reg */ 1566 ADJUST_PC(2); /* advance pc partially to make returns easier */ 1567 ILOGV("|invoke-virtual/jumbo args=%d @0x%08x {regs=v%d-v%d}", 1568 vsrc1, ref, vdst, vdst+vsrc1-1); 1569 thisPtr = (Object*) GET_REGISTER(vdst); 1570 } else { 1571 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1572 ref = FETCH(1); /* method ref */ 1573 vdst = FETCH(2); /* 4 regs -or- first reg */ 1574 1575 /* 1576 * The object against which we are executing a method is always 1577 * in the first argument. 1578 */ 1579 if (methodCallRange) { 1580 assert(vsrc1 > 0); 1581 ILOGV("|invoke-virtual-range args=%d @0x%04x {regs=v%d-v%d}", 1582 vsrc1, ref, vdst, vdst+vsrc1-1); 1583 thisPtr = (Object*) GET_REGISTER(vdst); 1584 } else { 1585 assert((vsrc1>>4) > 0); 1586 ILOGV("|invoke-virtual args=%d @0x%04x {regs=0x%04x %x}", 1587 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1588 thisPtr = (Object*) GET_REGISTER(vdst & 0x0f); 1589 } 1590 } 1591 1592 if (!checkForNull(thisPtr)) 1593 GOTO_exceptionThrown(); 1594 1595 /* 1596 * Resolve the method. This is the correct method for the static 1597 * type of the object. We also verify access permissions here. 1598 */ 1599 baseMethod = dvmDexGetResolvedMethod(methodClassDex, ref); 1600 if (baseMethod == NULL) { 1601 baseMethod = dvmResolveMethod(curMethod->clazz, ref,METHOD_VIRTUAL); 1602 if (baseMethod == NULL) { 1603 ILOGV("+ unknown method or access denied"); 1604 GOTO_exceptionThrown(); 1605 } 1606 } 1607 1608 /* 1609 * Combine the object we found with the vtable offset in the 1610 * method. 1611 */ 1612 assert(baseMethod->methodIndex < thisPtr->clazz->vtableCount); 1613 methodToCall = thisPtr->clazz->vtable[baseMethod->methodIndex]; 1614 1615#if defined(WITH_JIT) && defined(MTERP_STUB) 1616 self->methodToCall = methodToCall; 1617 self->callsiteClass = thisPtr->clazz; 1618#endif 1619 1620#if 0 1621 if (dvmIsAbstractMethod(methodToCall)) { 1622 /* 1623 * This can happen if you create two classes, Base and Sub, where 1624 * Sub is a sub-class of Base. Declare a protected abstract 1625 * method foo() in Base, and invoke foo() from a method in Base. 1626 * Base is an "abstract base class" and is never instantiated 1627 * directly. Now, Override foo() in Sub, and use Sub. This 1628 * Works fine unless Sub stops providing an implementation of 1629 * the method. 1630 */ 1631 dvmThrowAbstractMethodError("abstract method not implemented"); 1632 GOTO_exceptionThrown(); 1633 } 1634#else 1635 assert(!dvmIsAbstractMethod(methodToCall) || 1636 methodToCall->nativeFunc != NULL); 1637#endif 1638 1639 LOGVV("+++ base=%s.%s virtual[%d]=%s.%s", 1640 baseMethod->clazz->descriptor, baseMethod->name, 1641 (u4) baseMethod->methodIndex, 1642 methodToCall->clazz->descriptor, methodToCall->name); 1643 assert(methodToCall != NULL); 1644 1645#if 0 1646 if (vsrc1 != methodToCall->insSize) { 1647 LOGW("WRONG METHOD: base=%s.%s virtual[%d]=%s.%s", 1648 baseMethod->clazz->descriptor, baseMethod->name, 1649 (u4) baseMethod->methodIndex, 1650 methodToCall->clazz->descriptor, methodToCall->name); 1651 //dvmDumpClass(baseMethod->clazz); 1652 //dvmDumpClass(methodToCall->clazz); 1653 dvmDumpAllClasses(0); 1654 } 1655#endif 1656 1657 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1658 } 1659GOTO_TARGET_END 1660 1661GOTO_TARGET(invokeSuper, bool methodCallRange, bool jumboFormat) 1662 { 1663 Method* baseMethod; 1664 u2 thisReg; 1665 1666 EXPORT_PC(); 1667 1668 if (jumboFormat) { 1669 ref = FETCH(1) | (u4)FETCH(2) << 16; /* method ref */ 1670 vsrc1 = FETCH(3); /* count */ 1671 vdst = FETCH(4); /* first reg */ 1672 ADJUST_PC(2); /* advance pc partially to make returns easier */ 1673 ILOGV("|invoke-super/jumbo args=%d @0x%08x {regs=v%d-v%d}", 1674 vsrc1, ref, vdst, vdst+vsrc1-1); 1675 thisReg = vdst; 1676 } else { 1677 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1678 ref = FETCH(1); /* method ref */ 1679 vdst = FETCH(2); /* 4 regs -or- first reg */ 1680 1681 if (methodCallRange) { 1682 ILOGV("|invoke-super-range args=%d @0x%04x {regs=v%d-v%d}", 1683 vsrc1, ref, vdst, vdst+vsrc1-1); 1684 thisReg = vdst; 1685 } else { 1686 ILOGV("|invoke-super args=%d @0x%04x {regs=0x%04x %x}", 1687 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1688 thisReg = vdst & 0x0f; 1689 } 1690 } 1691 1692 /* impossible in well-formed code, but we must check nevertheless */ 1693 if (!checkForNull((Object*) GET_REGISTER(thisReg))) 1694 GOTO_exceptionThrown(); 1695 1696 /* 1697 * Resolve the method. This is the correct method for the static 1698 * type of the object. We also verify access permissions here. 1699 * The first arg to dvmResolveMethod() is just the referring class 1700 * (used for class loaders and such), so we don't want to pass 1701 * the superclass into the resolution call. 1702 */ 1703 baseMethod = dvmDexGetResolvedMethod(methodClassDex, ref); 1704 if (baseMethod == NULL) { 1705 baseMethod = dvmResolveMethod(curMethod->clazz, ref,METHOD_VIRTUAL); 1706 if (baseMethod == NULL) { 1707 ILOGV("+ unknown method or access denied"); 1708 GOTO_exceptionThrown(); 1709 } 1710 } 1711 1712 /* 1713 * Combine the object we found with the vtable offset in the 1714 * method's class. 1715 * 1716 * We're using the current method's class' superclass, not the 1717 * superclass of "this". This is because we might be executing 1718 * in a method inherited from a superclass, and we want to run 1719 * in that class' superclass. 1720 */ 1721 if (baseMethod->methodIndex >= curMethod->clazz->super->vtableCount) { 1722 /* 1723 * Method does not exist in the superclass. Could happen if 1724 * superclass gets updated. 1725 */ 1726 dvmThrowNoSuchMethodError(baseMethod->name); 1727 GOTO_exceptionThrown(); 1728 } 1729 methodToCall = curMethod->clazz->super->vtable[baseMethod->methodIndex]; 1730 1731#if 0 1732 if (dvmIsAbstractMethod(methodToCall)) { 1733 dvmThrowAbstractMethodError("abstract method not implemented"); 1734 GOTO_exceptionThrown(); 1735 } 1736#else 1737 assert(!dvmIsAbstractMethod(methodToCall) || 1738 methodToCall->nativeFunc != NULL); 1739#endif 1740 LOGVV("+++ base=%s.%s super-virtual=%s.%s", 1741 baseMethod->clazz->descriptor, baseMethod->name, 1742 methodToCall->clazz->descriptor, methodToCall->name); 1743 assert(methodToCall != NULL); 1744 1745 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1746 } 1747GOTO_TARGET_END 1748 1749GOTO_TARGET(invokeInterface, bool methodCallRange, bool jumboFormat) 1750 { 1751 Object* thisPtr; 1752 ClassObject* thisClass; 1753 1754 EXPORT_PC(); 1755 1756 if (jumboFormat) { 1757 ref = FETCH(1) | (u4)FETCH(2) << 16; /* method ref */ 1758 vsrc1 = FETCH(3); /* count */ 1759 vdst = FETCH(4); /* first reg */ 1760 ADJUST_PC(2); /* advance pc partially to make returns easier */ 1761 ILOGV("|invoke-interface/jumbo args=%d @0x%08x {regs=v%d-v%d}", 1762 vsrc1, ref, vdst, vdst+vsrc1-1); 1763 thisPtr = (Object*) GET_REGISTER(vdst); 1764 } else { 1765 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1766 ref = FETCH(1); /* method ref */ 1767 vdst = FETCH(2); /* 4 regs -or- first reg */ 1768 1769 /* 1770 * The object against which we are executing a method is always 1771 * in the first argument. 1772 */ 1773 if (methodCallRange) { 1774 assert(vsrc1 > 0); 1775 ILOGV("|invoke-interface-range args=%d @0x%04x {regs=v%d-v%d}", 1776 vsrc1, ref, vdst, vdst+vsrc1-1); 1777 thisPtr = (Object*) GET_REGISTER(vdst); 1778 } else { 1779 assert((vsrc1>>4) > 0); 1780 ILOGV("|invoke-interface args=%d @0x%04x {regs=0x%04x %x}", 1781 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1782 thisPtr = (Object*) GET_REGISTER(vdst & 0x0f); 1783 } 1784 } 1785 1786 if (!checkForNull(thisPtr)) 1787 GOTO_exceptionThrown(); 1788 1789 thisClass = thisPtr->clazz; 1790 1791 1792 /* 1793 * Given a class and a method index, find the Method* with the 1794 * actual code we want to execute. 1795 */ 1796 methodToCall = dvmFindInterfaceMethodInCache(thisClass, ref, curMethod, 1797 methodClassDex); 1798#if defined(WITH_JIT) && defined(MTERP_STUB) 1799 self->callsiteClass = thisClass; 1800 self->methodToCall = methodToCall; 1801#endif 1802 if (methodToCall == NULL) { 1803 assert(dvmCheckException(self)); 1804 GOTO_exceptionThrown(); 1805 } 1806 1807 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1808 } 1809GOTO_TARGET_END 1810 1811GOTO_TARGET(invokeDirect, bool methodCallRange, bool jumboFormat) 1812 { 1813 u2 thisReg; 1814 1815 EXPORT_PC(); 1816 1817 if (jumboFormat) { 1818 ref = FETCH(1) | (u4)FETCH(2) << 16; /* method ref */ 1819 vsrc1 = FETCH(3); /* count */ 1820 vdst = FETCH(4); /* first reg */ 1821 ADJUST_PC(2); /* advance pc partially to make returns easier */ 1822 ILOGV("|invoke-direct/jumbo args=%d @0x%08x {regs=v%d-v%d}", 1823 vsrc1, ref, vdst, vdst+vsrc1-1); 1824 thisReg = vdst; 1825 } else { 1826 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1827 ref = FETCH(1); /* method ref */ 1828 vdst = FETCH(2); /* 4 regs -or- first reg */ 1829 1830 if (methodCallRange) { 1831 ILOGV("|invoke-direct-range args=%d @0x%04x {regs=v%d-v%d}", 1832 vsrc1, ref, vdst, vdst+vsrc1-1); 1833 thisReg = vdst; 1834 } else { 1835 ILOGV("|invoke-direct args=%d @0x%04x {regs=0x%04x %x}", 1836 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1837 thisReg = vdst & 0x0f; 1838 } 1839 } 1840 1841 if (!checkForNull((Object*) GET_REGISTER(thisReg))) 1842 GOTO_exceptionThrown(); 1843 1844 methodToCall = dvmDexGetResolvedMethod(methodClassDex, ref); 1845 if (methodToCall == NULL) { 1846 methodToCall = dvmResolveMethod(curMethod->clazz, ref, 1847 METHOD_DIRECT); 1848 if (methodToCall == NULL) { 1849 ILOGV("+ unknown direct method"); // should be impossible 1850 GOTO_exceptionThrown(); 1851 } 1852 } 1853 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1854 } 1855GOTO_TARGET_END 1856 1857GOTO_TARGET(invokeStatic, bool methodCallRange, bool jumboFormat) 1858 EXPORT_PC(); 1859 1860 if (jumboFormat) { 1861 ref = FETCH(1) | (u4)FETCH(2) << 16; /* method ref */ 1862 vsrc1 = FETCH(3); /* count */ 1863 vdst = FETCH(4); /* first reg */ 1864 ADJUST_PC(2); /* advance pc partially to make returns easier */ 1865 ILOGV("|invoke-static/jumbo args=%d @0x%08x {regs=v%d-v%d}", 1866 vsrc1, ref, vdst, vdst+vsrc1-1); 1867 } else { 1868 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1869 ref = FETCH(1); /* method ref */ 1870 vdst = FETCH(2); /* 4 regs -or- first reg */ 1871 1872 if (methodCallRange) 1873 ILOGV("|invoke-static-range args=%d @0x%04x {regs=v%d-v%d}", 1874 vsrc1, ref, vdst, vdst+vsrc1-1); 1875 else 1876 ILOGV("|invoke-static args=%d @0x%04x {regs=0x%04x %x}", 1877 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1878 } 1879 1880 methodToCall = dvmDexGetResolvedMethod(methodClassDex, ref); 1881 if (methodToCall == NULL) { 1882 methodToCall = dvmResolveMethod(curMethod->clazz, ref, METHOD_STATIC); 1883 if (methodToCall == NULL) { 1884 ILOGV("+ unknown method"); 1885 GOTO_exceptionThrown(); 1886 } 1887 1888#if defined(WITH_JIT) && defined(MTERP_STUB) 1889 /* 1890 * The JIT needs dvmDexGetResolvedMethod() to return non-null. 1891 * Include the check if this code is being used as a stub 1892 * called from the assembly interpreter. 1893 */ 1894 if ((self->interpBreak.ctl.subMode & kSubModeJitTraceBuild) && 1895 (dvmDexGetResolvedMethod(methodClassDex, ref) == NULL)) { 1896 /* Class initialization is still ongoing */ 1897 dvmJitEndTraceSelect(self,pc); 1898 } 1899#endif 1900 } 1901 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1902GOTO_TARGET_END 1903 1904GOTO_TARGET(invokeVirtualQuick, bool methodCallRange, bool jumboFormat) 1905 { 1906 Object* thisPtr; 1907 1908 EXPORT_PC(); 1909 1910 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1911 ref = FETCH(1); /* vtable index */ 1912 vdst = FETCH(2); /* 4 regs -or- first reg */ 1913 1914 /* 1915 * The object against which we are executing a method is always 1916 * in the first argument. 1917 */ 1918 if (methodCallRange) { 1919 assert(vsrc1 > 0); 1920 ILOGV("|invoke-virtual-quick-range args=%d @0x%04x {regs=v%d-v%d}", 1921 vsrc1, ref, vdst, vdst+vsrc1-1); 1922 thisPtr = (Object*) GET_REGISTER(vdst); 1923 } else { 1924 assert((vsrc1>>4) > 0); 1925 ILOGV("|invoke-virtual-quick args=%d @0x%04x {regs=0x%04x %x}", 1926 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1927 thisPtr = (Object*) GET_REGISTER(vdst & 0x0f); 1928 } 1929 1930 if (!checkForNull(thisPtr)) 1931 GOTO_exceptionThrown(); 1932 1933 1934 /* 1935 * Combine the object we found with the vtable offset in the 1936 * method. 1937 */ 1938 assert(ref < (unsigned int) thisPtr->clazz->vtableCount); 1939 methodToCall = thisPtr->clazz->vtable[ref]; 1940#if defined(WITH_JIT) && defined(MTERP_STUB) 1941 self->callsiteClass = thisPtr->clazz; 1942 self->methodToCall = methodToCall; 1943#endif 1944 1945#if 0 1946 if (dvmIsAbstractMethod(methodToCall)) { 1947 dvmThrowAbstractMethodError("abstract method not implemented"); 1948 GOTO_exceptionThrown(); 1949 } 1950#else 1951 assert(!dvmIsAbstractMethod(methodToCall) || 1952 methodToCall->nativeFunc != NULL); 1953#endif 1954 1955 LOGVV("+++ virtual[%d]=%s.%s", 1956 ref, methodToCall->clazz->descriptor, methodToCall->name); 1957 assert(methodToCall != NULL); 1958 1959 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1960 } 1961GOTO_TARGET_END 1962 1963GOTO_TARGET(invokeSuperQuick, bool methodCallRange, bool jumboFormat) 1964 { 1965 u2 thisReg; 1966 1967 EXPORT_PC(); 1968 1969 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1970 ref = FETCH(1); /* vtable index */ 1971 vdst = FETCH(2); /* 4 regs -or- first reg */ 1972 1973 if (methodCallRange) { 1974 ILOGV("|invoke-super-quick-range args=%d @0x%04x {regs=v%d-v%d}", 1975 vsrc1, ref, vdst, vdst+vsrc1-1); 1976 thisReg = vdst; 1977 } else { 1978 ILOGV("|invoke-super-quick args=%d @0x%04x {regs=0x%04x %x}", 1979 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1980 thisReg = vdst & 0x0f; 1981 } 1982 /* impossible in well-formed code, but we must check nevertheless */ 1983 if (!checkForNull((Object*) GET_REGISTER(thisReg))) 1984 GOTO_exceptionThrown(); 1985 1986#if 0 /* impossible in optimized + verified code */ 1987 if (ref >= curMethod->clazz->super->vtableCount) { 1988 dvmThrowNoSuchMethodError(NULL); 1989 GOTO_exceptionThrown(); 1990 } 1991#else 1992 assert(ref < (unsigned int) curMethod->clazz->super->vtableCount); 1993#endif 1994 1995 /* 1996 * Combine the object we found with the vtable offset in the 1997 * method's class. 1998 * 1999 * We're using the current method's class' superclass, not the 2000 * superclass of "this". This is because we might be executing 2001 * in a method inherited from a superclass, and we want to run 2002 * in the method's class' superclass. 2003 */ 2004 methodToCall = curMethod->clazz->super->vtable[ref]; 2005 2006#if 0 2007 if (dvmIsAbstractMethod(methodToCall)) { 2008 dvmThrowAbstractMethodError("abstract method not implemented"); 2009 GOTO_exceptionThrown(); 2010 } 2011#else 2012 assert(!dvmIsAbstractMethod(methodToCall) || 2013 methodToCall->nativeFunc != NULL); 2014#endif 2015 LOGVV("+++ super-virtual[%d]=%s.%s", 2016 ref, methodToCall->clazz->descriptor, methodToCall->name); 2017 assert(methodToCall != NULL); 2018 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 2019 } 2020GOTO_TARGET_END 2021 2022 2023 /* 2024 * General handling for return-void, return, and return-wide. Put the 2025 * return value in "retval" before jumping here. 2026 */ 2027GOTO_TARGET(returnFromMethod) 2028 { 2029 StackSaveArea* saveArea; 2030 2031 /* 2032 * We must do this BEFORE we pop the previous stack frame off, so 2033 * that the GC can see the return value (if any) in the local vars. 2034 * 2035 * Since this is now an interpreter switch point, we must do it before 2036 * we do anything at all. 2037 */ 2038 PERIODIC_CHECKS(0); 2039 2040 ILOGV("> retval=0x%llx (leaving %s.%s %s)", 2041 retval.j, curMethod->clazz->descriptor, curMethod->name, 2042 curMethod->shorty); 2043 //DUMP_REGS(curMethod, fp); 2044 2045 saveArea = SAVEAREA_FROM_FP(fp); 2046 2047#ifdef EASY_GDB 2048 debugSaveArea = saveArea; 2049#endif 2050 2051 /* back up to previous frame and see if we hit a break */ 2052 fp = (u4*)saveArea->prevFrame; 2053 assert(fp != NULL); 2054 2055 /* Handle any special subMode requirements */ 2056 if (self->interpBreak.ctl.subMode != 0) { 2057 PC_FP_TO_SELF(); 2058 dvmReportReturn(self); 2059 } 2060 2061 if (dvmIsBreakFrame(fp)) { 2062 /* bail without popping the method frame from stack */ 2063 LOGVV("+++ returned into break frame"); 2064 GOTO_bail(); 2065 } 2066 2067 /* update thread FP, and reset local variables */ 2068 self->interpSave.curFrame = fp; 2069 curMethod = SAVEAREA_FROM_FP(fp)->method; 2070 self->interpSave.method = curMethod; 2071 //methodClass = curMethod->clazz; 2072 methodClassDex = curMethod->clazz->pDvmDex; 2073 pc = saveArea->savedPc; 2074 ILOGD("> (return to %s.%s %s)", curMethod->clazz->descriptor, 2075 curMethod->name, curMethod->shorty); 2076 2077 /* use FINISH on the caller's invoke instruction */ 2078 //u2 invokeInstr = INST_INST(FETCH(0)); 2079 if (true /*invokeInstr >= OP_INVOKE_VIRTUAL && 2080 invokeInstr <= OP_INVOKE_INTERFACE*/) 2081 { 2082 FINISH(3); 2083 } else { 2084 //LOGE("Unknown invoke instr %02x at %d", 2085 // invokeInstr, (int) (pc - curMethod->insns)); 2086 assert(false); 2087 } 2088 } 2089GOTO_TARGET_END 2090 2091 2092 /* 2093 * Jump here when the code throws an exception. 2094 * 2095 * By the time we get here, the Throwable has been created and the stack 2096 * trace has been saved off. 2097 */ 2098GOTO_TARGET(exceptionThrown) 2099 { 2100 Object* exception; 2101 int catchRelPc; 2102 2103 PERIODIC_CHECKS(0); 2104 2105 /* 2106 * We save off the exception and clear the exception status. While 2107 * processing the exception we might need to load some Throwable 2108 * classes, and we don't want class loader exceptions to get 2109 * confused with this one. 2110 */ 2111 assert(dvmCheckException(self)); 2112 exception = dvmGetException(self); 2113 dvmAddTrackedAlloc(exception, self); 2114 dvmClearException(self); 2115 2116 LOGV("Handling exception %s at %s:%d", 2117 exception->clazz->descriptor, curMethod->name, 2118 dvmLineNumFromPC(curMethod, pc - curMethod->insns)); 2119 2120 /* 2121 * Report the exception throw to any "subMode" watchers. 2122 * 2123 * TODO: if the exception was thrown by interpreted code, control 2124 * fell through native, and then back to us, we will report the 2125 * exception at the point of the throw and again here. We can avoid 2126 * this by not reporting exceptions when we jump here directly from 2127 * the native call code above, but then we won't report exceptions 2128 * that were thrown *from* the JNI code (as opposed to *through* it). 2129 * 2130 * The correct solution is probably to ignore from-native exceptions 2131 * here, and have the JNI exception code do the reporting to the 2132 * debugger. 2133 */ 2134 if (self->interpBreak.ctl.subMode != 0) { 2135 PC_FP_TO_SELF(); 2136 dvmReportExceptionThrow(self, exception); 2137 } 2138 2139 /* 2140 * We need to unroll to the catch block or the nearest "break" 2141 * frame. 2142 * 2143 * A break frame could indicate that we have reached an intermediate 2144 * native call, or have gone off the top of the stack and the thread 2145 * needs to exit. Either way, we return from here, leaving the 2146 * exception raised. 2147 * 2148 * If we do find a catch block, we want to transfer execution to 2149 * that point. 2150 * 2151 * Note this can cause an exception while resolving classes in 2152 * the "catch" blocks. 2153 */ 2154 catchRelPc = dvmFindCatchBlock(self, pc - curMethod->insns, 2155 exception, false, (void**)(void*)&fp); 2156 2157 /* 2158 * Restore the stack bounds after an overflow. This isn't going to 2159 * be correct in all circumstances, e.g. if JNI code devours the 2160 * exception this won't happen until some other exception gets 2161 * thrown. If the code keeps pushing the stack bounds we'll end 2162 * up aborting the VM. 2163 * 2164 * Note we want to do this *after* the call to dvmFindCatchBlock, 2165 * because that may need extra stack space to resolve exception 2166 * classes (e.g. through a class loader). 2167 * 2168 * It's possible for the stack overflow handling to cause an 2169 * exception (specifically, class resolution in a "catch" block 2170 * during the call above), so we could see the thread's overflow 2171 * flag raised but actually be running in a "nested" interpreter 2172 * frame. We don't allow doubled-up StackOverflowErrors, so 2173 * we can check for this by just looking at the exception type 2174 * in the cleanup function. Also, we won't unroll past the SOE 2175 * point because the more-recent exception will hit a break frame 2176 * as it unrolls to here. 2177 */ 2178 if (self->stackOverflowed) 2179 dvmCleanupStackOverflow(self, exception); 2180 2181 if (catchRelPc < 0) { 2182 /* falling through to JNI code or off the bottom of the stack */ 2183#if DVM_SHOW_EXCEPTION >= 2 2184 LOGD("Exception %s from %s:%d not caught locally", 2185 exception->clazz->descriptor, dvmGetMethodSourceFile(curMethod), 2186 dvmLineNumFromPC(curMethod, pc - curMethod->insns)); 2187#endif 2188 dvmSetException(self, exception); 2189 dvmReleaseTrackedAlloc(exception, self); 2190 GOTO_bail(); 2191 } 2192 2193#if DVM_SHOW_EXCEPTION >= 3 2194 { 2195 const Method* catchMethod = SAVEAREA_FROM_FP(fp)->method; 2196 LOGD("Exception %s thrown from %s:%d to %s:%d", 2197 exception->clazz->descriptor, dvmGetMethodSourceFile(curMethod), 2198 dvmLineNumFromPC(curMethod, pc - curMethod->insns), 2199 dvmGetMethodSourceFile(catchMethod), 2200 dvmLineNumFromPC(catchMethod, catchRelPc)); 2201 } 2202#endif 2203 2204 /* 2205 * Adjust local variables to match self->interpSave.curFrame and the 2206 * updated PC. 2207 */ 2208 //fp = (u4*) self->interpSave.curFrame; 2209 curMethod = SAVEAREA_FROM_FP(fp)->method; 2210 self->interpSave.method = curMethod; 2211 //methodClass = curMethod->clazz; 2212 methodClassDex = curMethod->clazz->pDvmDex; 2213 pc = curMethod->insns + catchRelPc; 2214 ILOGV("> pc <-- %s.%s %s", curMethod->clazz->descriptor, 2215 curMethod->name, curMethod->shorty); 2216 DUMP_REGS(curMethod, fp, false); // show all regs 2217 2218 /* 2219 * Restore the exception if the handler wants it. 2220 * 2221 * The Dalvik spec mandates that, if an exception handler wants to 2222 * do something with the exception, the first instruction executed 2223 * must be "move-exception". We can pass the exception along 2224 * through the thread struct, and let the move-exception instruction 2225 * clear it for us. 2226 * 2227 * If the handler doesn't call move-exception, we don't want to 2228 * finish here with an exception still pending. 2229 */ 2230 if (INST_INST(FETCH(0)) == OP_MOVE_EXCEPTION) 2231 dvmSetException(self, exception); 2232 2233 dvmReleaseTrackedAlloc(exception, self); 2234 FINISH(0); 2235 } 2236GOTO_TARGET_END 2237 2238 2239 2240 /* 2241 * General handling for invoke-{virtual,super,direct,static,interface}, 2242 * including "quick" variants. 2243 * 2244 * Set "methodToCall" to the Method we're calling, and "methodCallRange" 2245 * depending on whether this is a "/range" instruction. 2246 * 2247 * For a range call: 2248 * "vsrc1" holds the argument count (8 bits) 2249 * "vdst" holds the first argument in the range 2250 * For a non-range call: 2251 * "vsrc1" holds the argument count (4 bits) and the 5th argument index 2252 * "vdst" holds four 4-bit register indices 2253 * 2254 * The caller must EXPORT_PC before jumping here, because any method 2255 * call can throw a stack overflow exception. 2256 */ 2257GOTO_TARGET(invokeMethod, bool methodCallRange, const Method* _methodToCall, 2258 u2 count, u2 regs) 2259 { 2260 STUB_HACK(vsrc1 = count; vdst = regs; methodToCall = _methodToCall;); 2261 2262 //printf("range=%d call=%p count=%d regs=0x%04x\n", 2263 // methodCallRange, methodToCall, count, regs); 2264 //printf(" --> %s.%s %s\n", methodToCall->clazz->descriptor, 2265 // methodToCall->name, methodToCall->shorty); 2266 2267 u4* outs; 2268 int i; 2269 2270 /* 2271 * Copy args. This may corrupt vsrc1/vdst. 2272 */ 2273 if (methodCallRange) { 2274 // could use memcpy or a "Duff's device"; most functions have 2275 // so few args it won't matter much 2276 assert(vsrc1 <= curMethod->outsSize); 2277 assert(vsrc1 == methodToCall->insSize); 2278 outs = OUTS_FROM_FP(fp, vsrc1); 2279 for (i = 0; i < vsrc1; i++) 2280 outs[i] = GET_REGISTER(vdst+i); 2281 } else { 2282 u4 count = vsrc1 >> 4; 2283 2284 assert(count <= curMethod->outsSize); 2285 assert(count == methodToCall->insSize); 2286 assert(count <= 5); 2287 2288 outs = OUTS_FROM_FP(fp, count); 2289#if 0 2290 if (count == 5) { 2291 outs[4] = GET_REGISTER(vsrc1 & 0x0f); 2292 count--; 2293 } 2294 for (i = 0; i < (int) count; i++) { 2295 outs[i] = GET_REGISTER(vdst & 0x0f); 2296 vdst >>= 4; 2297 } 2298#else 2299 // This version executes fewer instructions but is larger 2300 // overall. Seems to be a teensy bit faster. 2301 assert((vdst >> 16) == 0); // 16 bits -or- high 16 bits clear 2302 switch (count) { 2303 case 5: 2304 outs[4] = GET_REGISTER(vsrc1 & 0x0f); 2305 case 4: 2306 outs[3] = GET_REGISTER(vdst >> 12); 2307 case 3: 2308 outs[2] = GET_REGISTER((vdst & 0x0f00) >> 8); 2309 case 2: 2310 outs[1] = GET_REGISTER((vdst & 0x00f0) >> 4); 2311 case 1: 2312 outs[0] = GET_REGISTER(vdst & 0x0f); 2313 default: 2314 ; 2315 } 2316#endif 2317 } 2318 } 2319 2320 /* 2321 * (This was originally a "goto" target; I've kept it separate from the 2322 * stuff above in case we want to refactor things again.) 2323 * 2324 * At this point, we have the arguments stored in the "outs" area of 2325 * the current method's stack frame, and the method to call in 2326 * "methodToCall". Push a new stack frame. 2327 */ 2328 { 2329 StackSaveArea* newSaveArea; 2330 u4* newFp; 2331 2332 ILOGV("> %s%s.%s %s", 2333 dvmIsNativeMethod(methodToCall) ? "(NATIVE) " : "", 2334 methodToCall->clazz->descriptor, methodToCall->name, 2335 methodToCall->shorty); 2336 2337 newFp = (u4*) SAVEAREA_FROM_FP(fp) - methodToCall->registersSize; 2338 newSaveArea = SAVEAREA_FROM_FP(newFp); 2339 2340 /* verify that we have enough space */ 2341 if (true) { 2342 u1* bottom; 2343 bottom = (u1*) newSaveArea - methodToCall->outsSize * sizeof(u4); 2344 if (bottom < self->interpStackEnd) { 2345 /* stack overflow */ 2346 LOGV("Stack overflow on method call (start=%p end=%p newBot=%p(%d) size=%d '%s')", 2347 self->interpStackStart, self->interpStackEnd, bottom, 2348 (u1*) fp - bottom, self->interpStackSize, 2349 methodToCall->name); 2350 dvmHandleStackOverflow(self, methodToCall); 2351 assert(dvmCheckException(self)); 2352 GOTO_exceptionThrown(); 2353 } 2354 //LOGD("+++ fp=%p newFp=%p newSave=%p bottom=%p", 2355 // fp, newFp, newSaveArea, bottom); 2356 } 2357 2358#ifdef LOG_INSTR 2359 if (methodToCall->registersSize > methodToCall->insSize) { 2360 /* 2361 * This makes valgrind quiet when we print registers that 2362 * haven't been initialized. Turn it off when the debug 2363 * messages are disabled -- we want valgrind to report any 2364 * used-before-initialized issues. 2365 */ 2366 memset(newFp, 0xcc, 2367 (methodToCall->registersSize - methodToCall->insSize) * 4); 2368 } 2369#endif 2370 2371#ifdef EASY_GDB 2372 newSaveArea->prevSave = SAVEAREA_FROM_FP(fp); 2373#endif 2374 newSaveArea->prevFrame = fp; 2375 newSaveArea->savedPc = pc; 2376#if defined(WITH_JIT) && defined(MTERP_STUB) 2377 newSaveArea->returnAddr = 0; 2378#endif 2379 newSaveArea->method = methodToCall; 2380 2381 if (self->interpBreak.ctl.subMode != 0) { 2382 /* 2383 * We mark ENTER here for both native and non-native 2384 * calls. For native calls, we'll mark EXIT on return. 2385 * For non-native calls, EXIT is marked in the RETURN op. 2386 */ 2387 PC_TO_SELF(); 2388 dvmReportInvoke(self, methodToCall); 2389 } 2390 2391 if (!dvmIsNativeMethod(methodToCall)) { 2392 /* 2393 * "Call" interpreted code. Reposition the PC, update the 2394 * frame pointer and other local state, and continue. 2395 */ 2396 curMethod = methodToCall; 2397 self->interpSave.method = curMethod; 2398 methodClassDex = curMethod->clazz->pDvmDex; 2399 pc = methodToCall->insns; 2400 self->interpSave.curFrame = fp = newFp; 2401#ifdef EASY_GDB 2402 debugSaveArea = SAVEAREA_FROM_FP(newFp); 2403#endif 2404 self->debugIsMethodEntry = true; // profiling, debugging 2405 ILOGD("> pc <-- %s.%s %s", curMethod->clazz->descriptor, 2406 curMethod->name, curMethod->shorty); 2407 DUMP_REGS(curMethod, fp, true); // show input args 2408 FINISH(0); // jump to method start 2409 } else { 2410 /* set this up for JNI locals, even if not a JNI native */ 2411 newSaveArea->xtra.localRefCookie = self->jniLocalRefTable.segmentState.all; 2412 2413 self->interpSave.curFrame = newFp; 2414 2415 DUMP_REGS(methodToCall, newFp, true); // show input args 2416 2417 if (self->interpBreak.ctl.subMode != 0) { 2418 dvmReportPreNativeInvoke(methodToCall, self, fp); 2419 } 2420 2421 ILOGD("> native <-- %s.%s %s", methodToCall->clazz->descriptor, 2422 methodToCall->name, methodToCall->shorty); 2423 2424 /* 2425 * Jump through native call bridge. Because we leave no 2426 * space for locals on native calls, "newFp" points directly 2427 * to the method arguments. 2428 */ 2429 (*methodToCall->nativeFunc)(newFp, &retval, methodToCall, self); 2430 2431 if (self->interpBreak.ctl.subMode != 0) { 2432 dvmReportPostNativeInvoke(methodToCall, self, fp); 2433 } 2434 2435 /* pop frame off */ 2436 dvmPopJniLocals(self, newSaveArea); 2437 self->interpSave.curFrame = fp; 2438 2439 /* 2440 * If the native code threw an exception, or interpreted code 2441 * invoked by the native call threw one and nobody has cleared 2442 * it, jump to our local exception handling. 2443 */ 2444 if (dvmCheckException(self)) { 2445 LOGV("Exception thrown by/below native code"); 2446 GOTO_exceptionThrown(); 2447 } 2448 2449 ILOGD("> retval=0x%llx (leaving native)", retval.j); 2450 ILOGD("> (return from native %s.%s to %s.%s %s)", 2451 methodToCall->clazz->descriptor, methodToCall->name, 2452 curMethod->clazz->descriptor, curMethod->name, 2453 curMethod->shorty); 2454 2455 //u2 invokeInstr = INST_INST(FETCH(0)); 2456 if (true /*invokeInstr >= OP_INVOKE_VIRTUAL && 2457 invokeInstr <= OP_INVOKE_INTERFACE*/) 2458 { 2459 FINISH(3); 2460 } else { 2461 //LOGE("Unknown invoke instr %02x at %d", 2462 // invokeInstr, (int) (pc - curMethod->insns)); 2463 assert(false); 2464 } 2465 } 2466 } 2467 assert(false); // should not get here 2468GOTO_TARGET_END 2469 2470/* File: cstubs/enddefs.cpp */ 2471 2472/* undefine "magic" name remapping */ 2473#undef retval 2474#undef pc 2475#undef fp 2476#undef curMethod 2477#undef methodClassDex 2478#undef self 2479#undef debugTrackedRefStart 2480 2481