/external/v8/src/mips/ |
H A D | builtins-mips.cc | 73 __ sll(s1, s0, kPointerSizeLog2); 286 STATIC_ASSERT(kSmiTag == 0 && kSmiTagSize < kPointerSizeLog2); 287 __ sll(elements_array_end, array_size, kPointerSizeLog2 - kSmiTagSize); 608 __ sll(a0, a0, kPointerSizeLog2); 648 t0, Operand(JSValue::kSize >> kPointerSizeLog2)); 827 __ sll(t0, a3, kPointerSizeLog2); 835 __ sll(t0, a0, kPointerSizeLog2); 911 __ sll(t3, a3, kPointerSizeLog2); 989 __ sll(t0, a3, kPointerSizeLog2 - kSmiTagSize); 1055 __ sll(t0, a1, kPointerSizeLog2 [all...] |
H A D | ic-mips.cc | 342 STATIC_ASSERT(kSmiTag == 0 && kSmiTagSize < kPointerSizeLog2); 343 __ sll(at, key, kPointerSizeLog2 - kSmiTagSize); 1049 __ sll(at, a3, kPointerSizeLog2 + 1); 1078 __ sll(at, a3, kPointerSizeLog2); 1094 __ sll(at, t2, kPointerSizeLog2); 1107 __ sll(t0, t1, kPointerSizeLog2); 1300 __ sll(scratch_value, key, kPointerSizeLog2 - kSmiTagSize); 1313 __ sll(scratch_value, key, kPointerSizeLog2 - kSmiTagSize);
|
H A D | stub-cache-mips.cc | 79 __ sll(at, offset_scratch, kPointerSizeLog2); 1555 __ sll(end_elements, v0, kPointerSizeLog2 - kSmiTagSize); 1597 __ sll(end_elements, v0, kPointerSizeLog2 - kSmiTagSize); 1638 __ sll(end_elements, v0, kPointerSizeLog2 - kSmiTagSize); 1739 __ sll(t1, t0, kPointerSizeLog2 - kSmiTagSize); 3219 __ sll(a1, a0, kPointerSizeLog2); 3276 __ sll(t0, a1, kPointerSizeLog2); 4124 STATIC_ASSERT(kSmiTag == 0 && kSmiTagSize < kPointerSizeLog2); 4125 __ sll(t0, a0, kPointerSizeLog2 - kSmiTagSize); 4271 STATIC_ASSERT(kSmiTag == 0 && kSmiTagSize < kPointerSizeLog2); [all...] |
H A D | macro-assembler-mips.cc | 56 lw(destination, MemOperand(s6, index << kPointerSizeLog2)); 65 lw(destination, MemOperand(s6, index << kPointerSizeLog2)); 71 sw(source, MemOperand(s6, index << kPointerSizeLog2)); 80 sw(source, MemOperand(s6, index << kPointerSizeLog2)); 213 And(t8, dst, Operand((1 << kPointerSizeLog2) - 1)); 513 sll(at, reg2, kPointerSizeLog2); 2803 sll(a2, a2, kPointerSizeLog2); 3031 sll(scratch2, object_size, kPointerSizeLog2); 4623 sll(t8, argument_count, kPointerSizeLog2); 5151 Ext(mask_reg, addr_reg, kPointerSizeLog2, Bitma [all...] |
H A D | code-stubs-mips.cc | 313 __ sll(t0, a0, kPointerSizeLog2 - kSmiTagSize); 390 __ sll(t0, a0, kPointerSizeLog2 - kSmiTagSize); 401 __ Branch(&slow_case, ne, a0, Operand(size >> kPointerSizeLog2)); 1622 __ sll(scratch1, scratch1, kPointerSizeLog2 + 1); 1648 __ sll(scratch, scratch, kPointerSizeLog2 + 1); 4445 __ sll(t3, a3, kPointerSizeLog2 - kSmiTagSize); 4459 __ sll(t3, a3, kPointerSizeLog2 - kSmiTagSize); 4738 __ sll(at, a1, kPointerSizeLog2 - kSmiTagSize); 5328 __ sll(t1, t1, kPointerSizeLog2); // Convert num elements to num bytes. 5666 __ sll(t0, code_, kPointerSizeLog2 [all...] |
H A D | deoptimizer-mips.cc | 906 __ sll(a1, a1, kPointerSizeLog2); // Count to offset.
|
/external/v8/src/arm/ |
H A D | builtins-arm.cc | 276 STATIC_ASSERT(kSmiTag == 0 && kSmiTagSize < kPointerSizeLog2); 279 Operand(array_size, LSL, kPointerSizeLog2 - kSmiTagSize)); 594 __ ldr(r0, MemOperand(sp, r0, LSL, kPointerSizeLog2, PreIndex)); 631 __ cmp(r4, Operand(JSValue::kSize >> kPointerSizeLog2)); 802 __ add(r6, r4, Operand(r3, LSL, kPointerSizeLog2)); // End of object. 809 __ add(r0, r5, Operand(r0, LSL, kPointerSizeLog2)); 881 __ add(r6, r2, Operand(r3, LSL, kPointerSizeLog2)); // End of object. 959 __ ldr(ip, MemOperand(r2, r3, LSL, kPointerSizeLog2 - 1)); 1027 __ add(sp, sp, Operand(r1, LSL, kPointerSizeLog2 - 1)); 1080 __ add(r2, r4, Operand(r3, LSL, kPointerSizeLog2)); [all...] |
H A D | ic-arm.cc | 341 STATIC_ASSERT(kSmiTag == 0 && kSmiTagSize < kPointerSizeLog2); 343 MemOperand(scratch1, key, LSL, kPointerSizeLog2 - kSmiTagSize)); 1046 __ add(r4, r4, Operand(r3, LSL, kPointerSizeLog2 + 1)); 1083 __ ldr(r5, MemOperand(r4, r3, LSL, kPointerSizeLog2)); 1097 __ ldr(r0, MemOperand(r1, r6, LSL, kPointerSizeLog2)); 1106 __ ldr(r0, MemOperand(r1, r5, LSL, kPointerSizeLog2)); 1417 __ add(address, address, Operand(key, LSL, kPointerSizeLog2 - kSmiTagSize)); 1428 __ add(address, address, Operand(key, LSL, kPointerSizeLog2 - kSmiTagSize));
|
H A D | macro-assembler-arm.cc | 399 ldr(destination, MemOperand(kRootRegister, index << kPointerSizeLog2), cond); 406 str(source, MemOperand(kRootRegister, index << kPointerSizeLog2), cond); 459 tst(dst, Operand((1 << kPointerSizeLog2) - 1)); 928 add(sp, sp, Operand(argument_count, LSL, kPointerSizeLog2)); 1244 ldr(r2, MemOperand(r3, r2, LSL, kPointerSizeLog2)); // Smi-tagged offset. 1482 add(t2, elements, Operand(t2, LSL, kPointerSizeLog2)); 1659 add(scratch2, result, Operand(object_size, LSL, kPointerSizeLog2), SetCC); 3516 Ubfx(mask_reg, addr_reg, kPointerSizeLog2, Bitmap::kBitsPerCellLog2); 3517 const int kLowBits = kPointerSizeLog2 + Bitmap::kBitsPerCellLog2; 3519 add(bitmap_reg, bitmap_reg, Operand(ip, LSL, kPointerSizeLog2)); [all...] |
H A D | stub-cache-arm.cc | 78 __ add(base_addr, base_addr, Operand(offset_scratch, LSL, kPointerSizeLog2)); 1559 Operand(r0, LSL, kPointerSizeLog2 - kSmiTagSize)); 1600 Operand(r0, LSL, kPointerSizeLog2 - kSmiTagSize)); 1639 Operand(r0, LSL, kPointerSizeLog2 - kSmiTagSize)); 1741 __ add(elements, elements, Operand(r4, LSL, kPointerSizeLog2 - kSmiTagSize)); 3231 __ add(r1, sp, Operand(r0, LSL, kPointerSizeLog2)); 3284 __ add(sp, sp, Operand(r1, LSL, kPointerSizeLog2)); 4067 STATIC_ASSERT(kSmiTag == 0 && kSmiTagSize < kPointerSizeLog2); 4069 MemOperand(r3, r0, LSL, kPointerSizeLog2 - kSmiTagSize)); 4216 STATIC_ASSERT(kSmiTag == 0 && kSmiTagSize < kPointerSizeLog2); [all...] |
H A D | code-stubs-arm.cc | 315 __ ldr(r3, MemOperand(r3, r0, LSL, kPointerSizeLog2 - kSmiTagSize)); 393 __ ldr(r3, MemOperand(r3, r0, LSL, kPointerSizeLog2 - kSmiTagSize)); 402 __ cmp(r0, Operand(size >> kPointerSizeLog2)); 1550 Operand(scratch1, LSL, kPointerSizeLog2 + 1)); 1575 Operand(scratch, LSL, kPointerSizeLog2 + 1)); 3816 __ add(r6, sp, Operand(r0, LSL, kPointerSizeLog2)); 4262 __ add(r3, fp, Operand(r3, LSL, kPointerSizeLog2 - kSmiTagSize)); 4276 __ add(r3, r2, Operand(r3, LSL, kPointerSizeLog2 - kSmiTagSize)); 4529 __ add(r3, r2, Operand(r1, LSL, kPointerSizeLog2 - kSmiTagSize)); 5104 __ str(r2, MemOperand(r3, r5, LSL, kPointerSizeLog2)); [all...] |
H A D | lithium-codegen-arm.cc | 2618 __ ldr(result, MemOperand(arguments, length, LSL, kPointerSizeLog2)); 2629 __ add(scratch, elements, Operand(key, LSL, kPointerSizeLog2)); 3589 __ add(scratch, elements, Operand(key, LSL, kPointerSizeLog2)); 3855 __ add(result, result, Operand(char_code, LSL, kPointerSizeLog2)); 5120 STATIC_ASSERT(kPointerSizeLog2 > kSmiTagSize); 5121 __ add(scratch, object, Operand(index, LSL, kPointerSizeLog2 - kSmiTagSize)); 5129 __ sub(scratch, result, Operand(index, LSL, kPointerSizeLog2 - kSmiTagSize));
|
/external/v8/src/ |
H A D | globals.h | 242 const int kPointerSizeLog2 = 3; member in namespace:v8::internal 246 const int kPointerSizeLog2 = 2; member in namespace:v8::internal
|
H A D | v8globals.h | 43 const int kObjectAlignmentBits = kPointerSizeLog2; 48 const intptr_t kPointerAlignment = (1 << kPointerSizeLog2);
|
H A D | store-buffer.cc | 147 return (a >> kPointerSizeLog2) - (b >> kPointerSizeLog2); 687 int_addr >>= kPointerSizeLog2; local 707 *old_top_++ = reinterpret_cast<Address>(int_addr << kPointerSizeLog2);
|
H A D | store-buffer.h | 84 static const int kStoreBufferOverflowBit = 1 << (14 + kPointerSizeLog2);
|
H A D | lithium.cc | 227 return kPointerSizeLog2;
|
H A D | spaces.h | 176 (1 << kPageSizeBits) >> (kPointerSizeLog2); 179 (1 << kPageSizeBits) >> (kPointerSizeLog2 + kBitsPerByteLog2); 545 return static_cast<uint32_t>(addr - this->address()) >> kPointerSizeLog2; 552 return static_cast<uint32_t>(offset) >> kPointerSizeLog2; 556 return this->address() + (index << kPointerSizeLog2); 2167 return static_cast<uint32_t>(addr - start_) >> kPointerSizeLog2; 2171 return reinterpret_cast<Address>(index << kPointerSizeLog2);
|
H A D | objects-visiting.h | 141 base + (object_size >> kPointerSizeLog2) - kMinObjectSizeInWords);
|
/external/v8/src/x64/ |
H A D | builtins-x64.cc | 155 __ shl(rdi, Immediate(kPointerSizeLog2)); 367 SmiIndex index = masm->SmiToIndex(rbx, rbx, kPointerSizeLog2); 833 __ PositiveSmiTimesPowerOfTwoToInteger64(rdx, rax, kPointerSizeLog2); 1089 masm->SmiToIndex(kScratchRegister, array_size, kPointerSizeLog2); 1494 Immediate(JSValue::kSize >> kPointerSizeLog2)); 1590 SmiIndex index = masm->SmiToIndex(rbx, rbx, kPointerSizeLog2);
|
H A D | code-stubs-x64.cc | 294 SmiIndex index = masm->SmiToIndex(rax, rax, kPointerSizeLog2); 370 SmiIndex index = masm->SmiToIndex(rax, rax, kPointerSizeLog2); 381 __ cmpq(rax, Immediate(size >> kPointerSizeLog2)); 2284 SmiIndex index = masm->SmiToIndex(rax, rax, kPointerSizeLog2); 2286 index = masm->SmiToNegativeIndex(rdx, rdx, kPointerSizeLog2); 2299 index = masm->SmiToIndex(rax, rcx, kPointerSizeLog2); 2301 index = masm->SmiToNegativeIndex(rdx, rdx, kPointerSizeLog2); 3266 __ shl(hash, Immediate(kPointerSizeLog2 + 1)); 4267 (Heap::kTrueValueRootIndex << kPointerSizeLog2) - kRootRegisterBias; 4290 (Heap::kFalseValueRootIndex << kPointerSizeLog2) [all...] |
H A D | macro-assembler-x64.cc | 156 (index << kPointerSizeLog2) - kRootRegisterBias)); 167 (fixed_offset << kPointerSizeLog2) - kRootRegisterBias)); 173 movq(Operand(kRootRegister, (index << kPointerSizeLog2) - kRootRegisterBias), 180 push(Operand(kRootRegister, (index << kPointerSizeLog2) - kRootRegisterBias)); 187 (index << kPointerSizeLog2) - kRootRegisterBias)); 309 testb(dst, Immediate((1 << kPointerSizeLog2) - 1)); 4273 Bitmap::kBitsPerCellLog2 + kPointerSizeLog2 - Bitmap::kBytesPerCellLog2; 4281 shrl(rcx, Immediate(kPointerSizeLog2));
|
H A D | ic-x64.cc | 345 SmiIndex index = masm->SmiToIndex(scratch, key, kPointerSizeLog2); 479 __ shl(rdi, Immediate(kPointerSizeLog2 + 1));
|
/external/v8/src/ia32/ |
H A D | builtins-ia32.cc | 155 __ shl(edi, kPointerSizeLog2); 755 __ shl(edx, kPointerSizeLog2 - kSmiTagSize); 1472 JSValue::kSize >> kPointerSizeLog2); local
|
H A D | ic-ia32.cc | 550 __ shl(edi, kPointerSizeLog2 + 1); 563 __ shl(edi, kPointerSizeLog2 + 1);
|