/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | inliner_test.cc | 56 auto rhs = builder.AddInstruction( local 59 HloInstruction::CreateMap(lhs->shape(), {lhs, rhs}, max_f32.get())); 69 op::Maximum(lhs, rhs)); 128 auto rhs = builder.AddInstruction( local 131 HloInstruction::CreateMap(lhs->shape(), {lhs, rhs}, max_f32.get())); 141 op::Subtract(rhs, lhs));
|
/external/tensorflow/tensorflow/core/platform/ |
H A D | fingerprint.h | 28 inline bool operator==(const Fprint128& lhs, const Fprint128& rhs) { argument 29 return lhs.low64 == rhs.low64 && lhs.high64 == rhs.high64;
|
/external/v8/src/base/ |
H A D | bits.cc | 28 int32_t SignedMulHigh32(int32_t lhs, int32_t rhs) { argument 29 int64_t const value = static_cast<int64_t>(lhs) * static_cast<int64_t>(rhs); 34 int32_t SignedMulHighAndAdd32(int32_t lhs, int32_t rhs, int32_t acc) { argument 36 bit_cast<uint32_t>(SignedMulHigh32(lhs, rhs))); 40 int32_t SignedDiv32(int32_t lhs, int32_t rhs) { argument 41 if (rhs == 0) return 0; 42 if (rhs == -1) return -lhs; 43 return lhs / rhs; 47 int32_t SignedMod32(int32_t lhs, int32_t rhs) { argument 48 if (rhs 68 SignedSaturatedAdd64(int64_t lhs, int64_t rhs) argument 75 SignedSaturatedSub64(int64_t lhs, int64_t rhs) argument 81 SignedMulOverflow32(int32_t lhs, int32_t rhs, int32_t* val) argument 89 SignedMulOverflow64(int64_t lhs, int64_t rhs, int64_t* val) argument [all...] |
/external/webrtc/webrtc/modules/video_coding/ |
H A D | rtt_filter.cc | 29 VCMRttFilter& VCMRttFilter::operator=(const VCMRttFilter& rhs) { argument 30 if (this != &rhs) { 31 _gotNonZeroUpdate = rhs._gotNonZeroUpdate; 32 _avgRtt = rhs._avgRtt; 33 _varRtt = rhs._varRtt; 34 _maxRtt = rhs._maxRtt; 35 _filtFactCount = rhs._filtFactCount; 36 _jumpCount = rhs._jumpCount; 37 _driftCount = rhs._driftCount; 38 memcpy(_jumpBuf, rhs [all...] |
/external/clang/include/clang/Basic/ |
H A D | SourceManagerInternals.h | 63 inline bool operator<(const LineEntry &lhs, const LineEntry &rhs) { argument 65 return lhs.FileOffset < rhs.FileOffset;
|
/external/clang/test/CodeGenCXX/ |
H A D | conditional-gnu-ext.cpp | 36 _Complex int rhs; local 38 return getComplex(1+2i) ? : rhs; 58 int rhs = 10; local 59 foo (cond()? : rhs);
|
/external/clang/test/SemaCXX/ |
H A D | vector-casts.cpp | 61 void madd(const testvec& rhs) { argument 62 v = v + rhs; // expected-error {{cannot convert between vector and non-scalar values}} 64 void madd2(testvec rhs) { argument 65 v = v + rhs; // expected-error {{cannot convert between vector and non-scalar values}}
|
/external/eigen/Eigen/src/Core/products/ |
H A D | TriangularSolverVector.h | 20 static void run(Index size, const LhsScalar* _lhs, Index lhsStride, RhsScalar* rhs) argument 25 >::run(size, _lhs, lhsStride, rhs); 29 // forward and backward substitution, row-major, rhs is a vector 36 static void run(Index size, const LhsScalar* _lhs, Index lhsStride, RhsScalar* rhs) argument 68 RhsMapper(rhs + startCol, 1), 69 rhs + startRow, 1, 78 rhs[i] -= (cjLhs.row(i).segment(s,k).transpose().cwiseProduct(Map<const Matrix<RhsScalar,Dynamic,1> >(rhs+s,k))).sum(); 81 rhs[i] /= cjLhs(i,i); 87 // forward and backward substitution, column-major, rhs i 94 run(Index size, const LhsScalar* _lhs, Index lhsStride, RhsScalar* rhs) argument [all...] |
/external/eigen/Eigen/src/Geometry/ |
H A D | OrthoMethods.h | 42 typename internal::nested_eval<OtherDerived,2>::type rhs(other.derived()); 44 numext::conj(lhs.coeff(1) * rhs.coeff(2) - lhs.coeff(2) * rhs.coeff(1)), 45 numext::conj(lhs.coeff(2) * rhs.coeff(0) - lhs.coeff(0) * rhs.coeff(2)), 46 numext::conj(lhs.coeff(0) * rhs.coeff(1) - lhs.coeff(1) * rhs.coeff(0)) 57 run(const VectorLhs& lhs, const VectorRhs& rhs) argument 60 numext::conj(lhs.coeff(1) * rhs.coeff(2) - lhs.coeff(2) * rhs [all...] |
/external/eigen/Eigen/src/SparseCore/ |
H A D | SparseSolverBase.h | 19 * The rhs is decomposed into small vertical panels which are solved through dense temporaries. 23 solve_sparse_through_dense_panels(const Decomposition &dec, const Rhs& rhs, Dest &dest) argument 27 // we process the sparse rhs per block of NbColsAtOnce columns temporarily stored into a dense matrix. 29 Index rhsCols = rhs.cols(); 30 Index size = rhs.rows(); 38 tmp.leftCols(actualCols) = rhs.middleCols(k,actualCols); 44 // Overload for vector as rhs 47 solve_sparse_through_dense_panels(const Decomposition &dec, const Rhs& rhs, Dest &dest) argument 50 Index size = rhs.rows(); 51 Eigen::Matrix<DestScalar,Dynamic,1> rhs_dense(rhs); [all...] |
H A D | SparseSparseProductWithPruning.h | 20 static void sparse_sparse_product_with_pruning_impl(const Lhs& lhs, const Rhs& rhs, ResultType& res, const typename ResultType::RealScalar& tolerance) argument 22 // return sparse_sparse_product_with_pruning_impl2(lhs,rhs,res); 29 Index cols = rhs.outerSize(); 31 eigen_assert(lhs.outerSize() == rhs.innerSize()); 43 evaluator<Rhs> rhsEval(rhs); 46 // given a rhs column containing Y non zeros, we assume that the respective Y columns 48 // the product of a rhs column with the lhs is X+Y where X is the average number of non zero 50 // Therefore, we have nnz(lhs*rhs) = nnz(lhs) + nnz(rhs) 54 double ratioColRes = double(estimated_nnz_prod)/(double(lhs.rows())*double(rhs 91 run(const Lhs& lhs, const Rhs& rhs, ResultType& res, const RealScalar& tolerance) argument 103 run(const Lhs& lhs, const Rhs& rhs, ResultType& res, const RealScalar& tolerance) argument 117 run(const Lhs& lhs, const Rhs& rhs, ResultType& res, const RealScalar& tolerance) argument 130 run(const Lhs& lhs, const Rhs& rhs, ResultType& res, const RealScalar& tolerance) argument 150 run(const Lhs& lhs, const Rhs& rhs, ResultType& res, const RealScalar& tolerance) argument 162 run(const Lhs& lhs, const Rhs& rhs, ResultType& res, const RealScalar& tolerance) argument 174 run(const Lhs& lhs, const Rhs& rhs, ResultType& res, const RealScalar& tolerance) argument 186 run(const Lhs& lhs, const Rhs& rhs, ResultType& res, const RealScalar& tolerance) argument [all...] |
/external/eigen/blas/ |
H A D | PackedTriangularMatrixVector.h | 27 static void run(Index size, const LhsScalar* lhs, const RhsScalar* rhs, ResScalar* res, ResScalar alpha) argument 39 ResMap(res+(IsLower ? s+i : 0),r) += alpha * cj(rhs[i]) * ConjLhsType(LhsMap(lhs+s,r)); 41 res[i] += alpha * cj(rhs[i]); 56 static void run(Index size, const LhsScalar* lhs, const RhsScalar* rhs, ResScalar* res, ResScalar alpha) argument 69 res[i] += alpha * (ConjLhsType(LhsMap(lhs+s,r)).cwiseProduct(ConjRhsType(RhsMap(rhs+(IsLower ? 0 : s+i),r)))).sum(); 71 res[i] += alpha * cj(rhs[i]);
|
/external/emma/core/java12/com/vladium/emma/data/ |
H A D | CoverageData.java | 86 * note: rhs entries override current entries if they have different stamps; 89 public IMergeable merge (final IMergeable rhs) argument 91 if ((rhs == null) || rhs.isEmpty () || (rhs == this)) 95 final CoverageData rhscdata = (CoverageData) rhs; // TODO: redesign so that the cast is not necessary
|
/external/gemmlowp/internal/ |
H A D | simd_wrappers_common_neon_sse.h | 198 const RegBlockInt32<1, 1>& rhs) { 200 result.buf.reg[0] = Add(lhs.buf.reg[0], Dup<Int32x4>(rhs.buf.reg[0])); 209 const RegBlockInt32<1, 1>& rhs) { 211 result.buf.reg[0] = Add(lhs.buf.reg[0], Dup<Int32x4>(rhs.buf.reg[0])); 220 const RegBlockInt32<4, 1>& rhs) { 222 result.buf.reg[0] = Add(lhs.buf.reg[0], rhs.buf.reg[0]); 231 const RegBlockInt32<1, 4>& rhs) { 233 result.buf.reg[0] = Add(lhs.buf.reg[0], rhs.buf.reg[0]); 242 const RegBlockInt32<1, 4>& rhs) { 244 result.buf.reg[0] = Add(lhs.buf.reg[0], DupLane<0>(rhs 197 Run(const RegBlockInt32<4, 1>& lhs, const RegBlockInt32<1, 1>& rhs) argument 208 Run(const RegBlockInt32<1, 4>& lhs, const RegBlockInt32<1, 1>& rhs) argument 219 Run(const RegBlockInt32<4, 1>& lhs, const RegBlockInt32<4, 1>& rhs) argument 230 Run(const RegBlockInt32<1, 4>& lhs, const RegBlockInt32<1, 4>& rhs) argument 241 Run(const RegBlockInt32<4, 4>& lhs, const RegBlockInt32<1, 4>& rhs) argument 255 Run(const RegBlockInt32<4, 4>& lhs, const RegBlockInt32<4, 1>& rhs) argument 269 Run(const RegBlockInt32<8, 1>& lhs, const RegBlockInt32<1, 1>& rhs) argument 283 Run(const RegBlockInt32<8, 1>& lhs, const RegBlockInt32<8, 1>& rhs) argument 296 Run(const RegBlockInt32<8, 4>& lhs, const RegBlockInt32<1, 4>& rhs) argument 314 Run(const RegBlockInt32<8, 4>& lhs, const RegBlockInt32<8, 1>& rhs) argument 332 Run(const RegBlockInt32<1, 8>& lhs, const RegBlockInt32<1, 8>& rhs) argument 344 Run(const RegBlockInt32<1, 8>& lhs, const RegBlockInt32<1, 1>& rhs) argument 356 Run(const RegBlockInt32<4, 1>& lhs, const RegBlockInt32<1, 1>& rhs) argument 367 Run(const RegBlockInt32<4, 1>& lhs, const RegBlockInt32<4, 1>& rhs) argument 378 Run(const RegBlockInt32<1, 4>& lhs, const RegBlockInt32<1, 4>& rhs) argument 389 Run(const RegBlockInt32<1, 4>& lhs, const RegBlockInt32<1, 1>& rhs) argument 400 Run(const RegBlockInt32<4, 4>& lhs, const RegBlockInt32<1, 4>& rhs) argument 415 Run(const RegBlockInt32<4, 4>& lhs, const RegBlockInt32<4, 1>& rhs) argument 430 Run(const RegBlockInt32<8, 1>& lhs, const RegBlockInt32<1, 1>& rhs) argument 444 Run(const RegBlockInt32<8, 1>& lhs, const RegBlockInt32<8, 1>& rhs) argument 457 Run(const RegBlockInt32<8, 4>& lhs, const RegBlockInt32<1, 4>& rhs) argument 474 Run(const RegBlockInt32<8, 4>& lhs, const RegBlockInt32<8, 1>& rhs) argument 492 Run(const RegBlockInt32<Rows, 1>& lhs, const RegBlockInt32<1, 1>& rhs, RegBlockInt32<Rows, 1>* acc) argument 505 Run(const RegBlockInt32<Rows, 1>& lhs, const RegBlockInt32<1, 1>& rhs, RegBlockInt32<Rows, Cols>* acc) argument 524 Run(const RegBlockInt32<1, Cols>& lhs, const RegBlockInt32<1, 1>& rhs, RegBlockInt32<1, Cols>* acc) argument 537 Run(const RegBlockInt32<1, 1>& lhs, const RegBlockInt32<1, 1>& rhs, RegBlockInt32<Rows, Cols>* acc) argument 551 Run(const RegBlockInt32<1, 1>& lhs, const RegBlockInt32<1, 1>& rhs, RegBlockInt32<1, 1>* acc) argument 561 Run(const RegBlockInt32<Rows, 1>& lhs, const RegBlockInt32<1, 4>& rhs, RegBlockInt32<Rows, 4>* acc) argument 578 Run(const RegBlockInt32<1, 4>& lhs, const RegBlockInt32<1, 1>& rhs, RegBlockInt32<Rows, 4>* acc) argument 600 Run(const RegBlockInt32<1, 1>& lhs, const RegBlockInt32<1, 1>& rhs, RegBlockInt32<1, Cols>* acc) argument 613 Run(const RegBlockInt32<1, 4>& lhs, const RegBlockInt32<1, 1>& rhs, RegBlockInt32<1, 4>* acc) argument 624 Run(const RegBlockInt32<4, 1>& lhs, const RegBlockInt32<1, 1>& rhs, RegBlockInt32<4, Cols>* acc) argument 637 Run(const RegBlockInt32<4, 1>& lhs, const RegBlockInt32<1, 1>& rhs, RegBlockInt32<4, 1>* acc) argument [all...] |
/external/gemmlowp/meta/ |
H A D | legacy_single_thread_gemm.h | 30 const std::uint8_t* rhs, std::int32_t m, std::int32_t n, 50 params.rhs = rhs; 75 const std::uint8_t* rhs, std::int32_t n, std::int32_t k, 94 params.rhs = rhs; 123 const std::uint8_t* rhs, std::int32_t m, std::int32_t n, 143 params.rhs = rhs; 164 const std::uint8_t* rhs, st 29 gemm_q8_strided(std::uint8_t* scratch, const std::uint8_t* lhs, const std::uint8_t* rhs, std::int32_t m, std::int32_t n, std::int32_t k, std::int32_t lhs_offset, std::int32_t rhs_offset, std::int32_t result_offset, std::int32_t multiplicative_offset, std::int32_t shift, std::uint8_t* result, std::int32_t result_stride) argument 74 gemv_q8(std::uint8_t* scratch, const std::uint8_t* lhs, const std::uint8_t* rhs, std::int32_t n, std::int32_t k, std::int32_t lhs_offset, std::int32_t rhs_offset, std::int32_t result_offset, std::int32_t multiplicative_offset, std::int32_t shift, std::uint8_t* result) argument 122 gemm_i32_strided(std::uint8_t* scratch, const std::uint8_t* lhs, const std::uint8_t* rhs, std::int32_t m, std::int32_t n, std::int32_t k, std::int32_t lhs_offset, std::int32_t rhs_offset, std::int32_t* result, std::int32_t result_stride) argument 163 gemv_i32(std::uint8_t* scratch, const std::uint8_t* lhs, const std::uint8_t* rhs, std::int32_t n, std::int32_t k, std::int32_t lhs_offset, std::int32_t rhs_offset, std::int32_t* result) argument 207 gemm_f_strided(std::uint8_t* scratch, const std::uint8_t* lhs, const std::uint8_t* rhs, std::int32_t m, std::int32_t n, std::int32_t k, std::int32_t lhs_offset, std::int32_t rhs_offset, float result_offset, float* result, std::int32_t result_stride) argument 248 gemv_f(std::uint8_t* scratch, const std::uint8_t* lhs, const std::uint8_t* rhs, std::int32_t n, std::int32_t k, std::int32_t lhs_offset, std::int32_t rhs_offset, float result_offset, float* result) argument [all...] |
/external/google-breakpad/src/testing/gtest/include/gtest/internal/ |
H A D | gtest-string.h | 134 static bool CStringEquals(const char* lhs, const char* rhs); 152 static bool WideCStringEquals(const wchar_t* lhs, const wchar_t* rhs); 161 const char* rhs); 176 const wchar_t* rhs); 244 // Returns < 0 if this is less than rhs, 0 if this is equal to rhs, or > 0 245 // if this is greater than rhs. 246 int Compare(const String& rhs) const; 254 bool operator<(const String& rhs) const { return Compare(rhs) < 283 operator =(const String& rhs) argument [all...] |
/external/icu/android_icu4j/src/main/java/android/icu/impl/ |
H A D | ICURWLock.java | 86 private Stats(Stats rhs) { argument 87 this(rhs._rc, rhs._mrc, rhs._wrc, rhs._wc, rhs._wwc);
|
/external/icu/icu4c/source/test/intltest/ |
H A D | pluralmaptest.cpp | 24 static UBool strEqual(const UnicodeString &lhs, const UnicodeString &rhs) { argument 25 return lhs == rhs; 171 PluralMapForPluralMapTest *rhs = new PluralMapForPluralMapTest(); local 172 if (rhs == NULL) { 176 addVariant(PluralMapBase::OTHER, "pickles", *rhs); 177 addVariant(PluralMapBase::ONE, "pickle", *rhs); 178 addVariant(PluralMapBase::FEW, "picklefew", *rhs); 179 PluralMapForPluralMapTest lhs(*rhs); 180 delete rhs; 184 PluralMapForPluralMapTest *rhs local [all...] |
/external/icu/icu4j/eclipse-build/plugins.template/com.ibm.icu.base/src/com/ibm/icu/util/ |
H A D | Currency.java | 221 * Return true if rhs is a Currency instance, 225 public boolean equals(Object rhs) { argument 227 return currency.equals(((Currency)rhs).currency);
|
/external/icu/icu4j/eclipse-build/plugins.template/com.ibm.icu.base.tests/src/com/ibm/icu/tests/ |
H A D | ICUTestCase.java | 67 * @param rhs an object to test, may be null 69 public static void assertNotEqual(Object lhs, Object rhs) { argument 71 if (rhs == null) fail("null equals null"); 73 if (lhs.equals(rhs)) { 74 fail(lhs.toString() + " equals " + rhs); 79 public static void assertNotEqual(long lhs, long rhs) { argument 80 if (lhs == rhs) { 186 * @param rhs the right array 188 public static void assertArraysEqual(Object lhs, Object rhs) { argument 190 Class<? extends Object> rcls = rhs 206 assertArraysNotEqual(Object lhs, Object rhs) argument [all...] |
/external/icu/icu4j/main/classes/core/src/com/ibm/icu/impl/ |
H A D | ICURWLock.java | 84 private Stats(Stats rhs) { argument 85 this(rhs._rc, rhs._mrc, rhs._wrc, rhs._wc, rhs._wwc);
|
/external/libbrillo/brillo/ |
H A D | any.h | 45 Any(const Any& rhs); 46 Any(Any&& rhs); // NOLINT(build/c++11) 58 Any& operator=(const Any& rhs); 59 Any& operator=(Any&& rhs); // NOLINT(build/c++11) 70 bool operator==(const Any& rhs) const; 71 inline bool operator!=(const Any& rhs) const { return !operator==(rhs); } 208 inline void swap(brillo::Any& lhs, brillo::Any& rhs) { argument 209 lhs.Swap(rhs);
|
/external/libcxx/test/std/algorithms/alg.sorting/alg.clamp/ |
H A D | clamp.comp.pass.cpp | 30 bool eq(const Tag& rhs, const Tag& lhs) { return rhs.val == lhs.val && rhs.tag == lhs.tag; } argument 31 // bool operator==(const Tag& rhs, const Tag& lhs) { return rhs.val == lhs.val; } 32 bool comp (const Tag& rhs, const Tag& lhs) { return rhs.val < lhs.val; } argument
|
/external/libcxx/test/std/algorithms/alg.sorting/alg.merge/ |
H A D | inplace_merge_comp.pass.cpp | 39 S(const S& rhs) : i_(rhs.i_) {} argument 40 S( S&& rhs) : i_(rhs.i_) { rhs.i_ = -1; } argument 42 S& operator =(const S& rhs) { i_ = rhs.i_; return *this; } argument 43 S& operator =( S&& rhs) { i_ = rhs.i_; rhs argument 121 operator ()(const Pair& lhs, const Pair& rhs) argument [all...] |
/external/libcxx/test/std/experimental/filesystem/class.path/path.member/ |
H A D | path.append.pass.cpp | 39 MultiStringType rhs; member in struct:AppendOperatorTestcase 85 Str RShort = (Ptr)TC.rhs; 171 const Ptr R = TC.rhs; 309 path RHS((const char*)TC.rhs);
|