Searched refs:left_index (Results 1 - 4 of 4) sorted by relevance
/external/eigen/unsupported/Eigen/CXX11/src/Tensor/ |
H A D | TensorConcatenation.h | 215 Index left_index; local 217 left_index = subs[0]; 219 left_index += (subs[i] % left_dims[i]) * m_leftStrides[i]; 222 left_index = subs[NumDims - 1]; 224 left_index += (subs[i] % left_dims[i]) * m_leftStrides[i]; 227 return m_leftImpl.coeff(left_index); 328 Index left_index = subs[0]; local 330 left_index += (subs[i] % left_dims[i]) * this->m_leftStrides[i]; 332 return this->m_leftImpl.coeffRef(left_index);
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
H A D | fdct32x32_neon.c | 523 #define ADD_S16_S32(a, left_index, right_index, b, b_index) \ 526 vaddl_s16(vget_low_s16(a[left_index]), vget_low_s16(a[right_index])); \ 527 b##_hi[b_index] = vaddl_s16(vget_high_s16(a[left_index]), \ 531 #define SUB_S16_S32(a, left_index, right_index, b, b_index) \ 534 vsubl_s16(vget_low_s16(a[left_index]), vget_low_s16(a[right_index])); \ 535 b##_hi[b_index] = vsubl_s16(vget_high_s16(a[left_index]), \ 553 #define ADD_S32(a, left_index, right_index, b, b_index) \ 555 b##_lo[b_index] = vaddq_s32(a##_lo[left_index], a##_lo[right_index]); \ 556 b##_hi[b_index] = vaddq_s32(a##_hi[left_index], a##_hi[right_index]); \ 559 #define SUB_S32(a, left_index, right_inde [all...] |
/external/v8/src/asmjs/ |
H A D | asm-wasm-builder.cc | 1775 TypeIndex left_index = TypeIndexOf(left, ignore_sign); local 1777 if (left_index == kFixnum) { 1778 left_index = right_index; 1781 right_index = left_index; 1783 if (left_index == kFixnum && right_index == kFixnum) { 1784 left_index = kInt32; 1787 if (left_index != right_index) { 1788 DCHECK(ignore_sign && (left_index <= 1) && (right_index <= 1)); 1790 return left_index;
|
/external/v8/src/compiler/ |
H A D | register-allocator.cc | 149 size_t left_index = 0; local 152 size_t current_index = left_index + (right_index - left_index) / 2; 157 DCHECK(left_index < current_index); 158 left_index = current_index;
|
Completed in 201 milliseconds