/external/tensorflow/tensorflow/core/kernels/ |
H A D | cwise_op_gpu_left_shift.cu.cc | 22 DEFINE_BINARY8(left_shift, int8, int16, int32, int64, uint8, uint16, uint32,
|
H A D | cwise_op_left_shift.cc | 19 REGISTER8(BinaryOp, CPU, "LeftShift", functor::left_shift, int8, int16, int32, 26 BinaryOp<SYCLDevice, functor::left_shift<TYPE>>); 40 REGISTER8(BinaryOp, GPU, "LeftShift", functor::left_shift, int8, int16, int32,
|
H A D | cwise_ops.h | 907 struct left_shift : base<T, left_shift_op<T>> {}; struct in namespace:tensorflow::functor
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/ |
H A D | quantization_util.h | 37 int* left_shift); 55 int32_t* quantized_multiplier, int* left_shift);
|
H A D | quantization_util.cc | 45 int* left_shift) { 47 QuantizeMultiplier(double_multiplier, quantized_multiplier, left_shift); 48 TFLITE_CHECK_GE(*left_shift, 0); 64 int32_t* quantized_multiplier, int* left_shift) { 78 quantized_multiplier, left_shift); 43 QuantizeMultiplierGreaterThanOne(double double_multiplier, int32_t* quantized_multiplier, int* left_shift) argument 62 PreprocessSoftmaxScaling(double beta, double input_scale, int input_integer_bits, int32_t* quantized_multiplier, int* left_shift) argument
|
H A D | common.h | 101 int32 x, int32 quantized_multiplier, int left_shift) { 103 return SaturatingRoundingDoublingHighMul(x * (1 << left_shift), 111 int left_shift = shift > 0 ? shift : 0; local 114 x * (1 << left_shift), quantized_multiplier), 100 MultiplyByQuantizedMultiplierGreaterThanOne( int32 x, int32 quantized_multiplier, int left_shift) argument
|
/external/tensorflow/tensorflow/python/ops/ |
H A D | bitwise_ops_test.py | 105 [bitwise_ops.left_shift(lhs, rhs), 107 self.assertAllEqual(left_shift_result, np.left_shift(lhs, rhs)) 118 [bitwise_ops.left_shift(lhs, rhs), 120 self.assertAllEqual(left_shift_result, np.left_shift(lhs, rhs)) 134 sess.run([bitwise_ops.left_shift(lhs, rhs), 150 ls_tensor = bitwise_ops.left_shift(lhs, rhs)
|
/external/tensorflow/tensorflow/contrib/lite/kernels/ |
H A D | add.cc | 117 const int left_shift = 20; local 125 twice_max_input_scale / ((1 << left_shift) * output->params.scale); 145 type::opname(left_shift, GetTensorData<uint8_t>(input1), \
|
/external/gemmlowp/internal/ |
H A D | output.h | 158 left_shift = std::max(0, output_stage.result_exponent); 169 ShiftLeft(input.reg[i], left_shift), 178 int left_shift; member in struct:gemmlowp::OutputStageEvalBufferImpl
|
/external/vixl/src/aarch64/ |
H A D | simulator-aarch64.cc | 386 unsigned left_shift) const { 421 return ShiftOperand(reg_size, value, LSL, left_shift); 4855 int left_shift = immhimmb - (8 << highestSetBit); local 4858 shl(vf, rd, rn, left_shift); 4861 sli(vf, rd, rn, left_shift); 4864 sqshl(vf, rd, rn, left_shift); 4867 uqshl(vf, rd, rn, left_shift); 4870 sqshlu(vf, rd, rn, left_shift); 4963 int left_shift = immhimmb - (8 << highestSetBit); local 4967 shl(vf, rd, rn, left_shift); [all...] |
H A D | assembler-aarch64.cc | 2800 void Assembler::orr(const VRegister& vd, const int imm8, const int left_shift) { 2801 NEONModifiedImmShiftLsl(vd, imm8, left_shift, NEONModifiedImmediate_ORR); 2816 void Assembler::bic(const VRegister& vd, const int imm8, const int left_shift) { 2817 NEONModifiedImmShiftLsl(vd, imm8, left_shift, NEONModifiedImmediate_BIC); 4106 const int left_shift, 4110 VIXL_ASSERT((left_shift == 0) || (left_shift == 8) || (left_shift == 16) || 4111 (left_shift == 24)); 4121 cmode_1 = (left_shift >> [all...] |
H A D | assembler-aarch64.h | 1769 void orr(const VRegister& vd, const int imm8, const int left_shift = 0); 1781 void bic(const VRegister& vd, const int imm8, const int left_shift = 0); 2760 static Instr ImmExtendShift(unsigned left_shift) { argument 2761 VIXL_ASSERT(left_shift <= 4); 2762 return left_shift << ImmExtendShift_offset; 3155 unsigned left_shift); 3237 const int left_shift,
|
H A D | macro-assembler-aarch64.h | 2486 void Bic(const VRegister& vd, const int imm8, const int left_shift = 0) { 2489 bic(vd, imm8, left_shift); 2691 void Orr(const VRegister& vd, const int imm8, const int left_shift = 0) { 2694 orr(vd, imm8, left_shift);
|
H A D | simulator-aarch64.h | 1908 unsigned left_shift = 0) const;
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/reference/ |
H A D | reference_ops.h | 45 int32 x, int32 quantized_multiplier, int left_shift) { 47 return SaturatingRoundingDoublingHighMul(x * (1 << left_shift), 830 inline void Add(int left_shift, const uint8* input1_data, 864 const int32 shifted_input1_val = input1_val * (1 << left_shift); 865 const int32 shifted_input2_val = input2_val * (1 << left_shift); 942 inline void BroadcastAdd(int left_shift, const uint8* input1_data, 977 const int32 shifted_input1_val = input1_val * (1 << left_shift); 978 const int32 shifted_input2_val = input2_val * (1 << left_shift); 1002 inline void BroadcastAdd(int left_shift, const uint8* input1_data, 1022 BroadcastAdd(left_shift, input1_dat 44 MultiplyByQuantizedMultiplierGreaterThanOne( int32 x, int32 quantized_multiplier, int left_shift) argument [all...] |
/external/v8/src/arm64/ |
H A D | assembler-arm64-inl.h | 1092 Instr Assembler::ImmExtendShift(unsigned left_shift) { argument 1093 DCHECK(left_shift <= 4); 1094 return left_shift << ImmExtendShift_offset;
|
H A D | simulator-arm64.cc | 971 T Simulator::ExtendValue(T value, Extend extend_type, unsigned left_shift) { 1001 return value << left_shift; 1476 unsigned left_shift = instr->ImmExtendShift(); 1478 uint64_t op2 = ExtendValue(xreg(instr->Rm()), ext, left_shift); 1481 uint32_t op2 = ExtendValue(wreg(instr->Rm()), ext, left_shift);
|
H A D | assembler-arm64.cc | 2488 unsigned left_shift) { 2496 unsigned non_shift_bits = (reg_size - left_shift) & (reg_size - 1); 2510 lsl(rd, rn_, left_shift); 2517 lsl(rd, rn_, left_shift); 2485 EmitExtendShift(const Register& rd, const Register& rn, Extend extend, unsigned left_shift) argument
|
H A D | simulator-arm64.h | 700 unsigned left_shift = 0);
|
H A D | assembler-arm64.h | 1774 inline static Instr ImmExtendShift(unsigned left_shift); 1923 unsigned left_shift);
|
/external/pdfium/core/fxge/dib/ |
H A D | cfx_dibsource.cpp | 626 int left_shift = rect.left % 32; local 627 int right_shift = 32 - left_shift; 634 (src_scan[i] << left_shift) | (src_scan[i + 1] >> right_shift);
|
H A D | cfx_dibitmap.cpp | 1033 int left_shift = rect.left % 8; local 1050 uint8_t left_flag = *dest_scan_top & (255 << (8 - left_shift));
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/optimized/ |
H A D | optimized_ops.h | 1403 inline void Add(int left_shift, const uint8* input1_data, 1460 const auto left_shift_dup = vdupq_n_s32(left_shift); 1493 const int32 shifted_input1_val = input1_val * (1 << left_shift); 1494 const int32 shifted_input2_val = input2_val * (1 << left_shift); 1591 inline void BroadcastAdd(int left_shift, const uint8* input1_data, 1626 const int32 shifted_input1_val = input1_val * (1 << left_shift); 1627 const int32 shifted_input2_val = input2_val * (1 << left_shift); 1651 inline void BroadcastAdd(int left_shift, const uint8* input1_data, 1671 BroadcastAdd(left_shift, input1_data, input1_dims, input1_offset, 2351 int left_shift [all...] |
/external/tensorflow/tensorflow/compiler/tests/ |
H A D | binary_ops_test.py | 228 bitwise_ops.left_shift, lhs, rhs, 229 expected=np.left_shift(lhs, rhs))
|
/external/valgrind/VEX/priv/ |
H A D | guest_x86_toIR.c | 6186 Bool left_shift ) 6228 ( left_shift ? 'l' : 'r' ), nameISize(sz), 6236 ( left_shift ? 'l' : 'r' ), nameISize(sz), 6248 getres = left_shift ? Iop_64HIto32 : Iop_64to32; 6249 shift = left_shift ? Iop_Shl64 : Iop_Shr64; 6257 getres = left_shift ? Iop_32HIto16 : Iop_32to16; 6258 shift = left_shift ? Iop_Shl32 : Iop_Shr32; 6267 if (left_shift) 6281 setFlags_DEP1_DEP2_shift ( left_shift ? Iop_Shl32 : Iop_Sar32,
|