/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | bfloat16_support.cc | 29 case HloOpcode::kTuple: 47 case HloOpcode::kTuple: 65 case HloOpcode::kTuple: 92 case HloOpcode::kTuple:
|
H A D | tuple_simplifier.cc | 38 if (instruction->opcode() == HloOpcode::kTuple || 58 if (instruction->opcode() == HloOpcode::kTuple) { 110 if (instruction->operand(0)->opcode() == HloOpcode::kTuple) { 117 if (user->opcode() == HloOpcode::kTuple ||
|
H A D | hlo_constant_folding.cc | 57 instruction->opcode() == HloOpcode::kTuple ||
|
H A D | hlo_opcode_test.cc | 61 case HloOpcode::kTuple:
|
H A D | batchnorm_expander_test.cc | 72 EXPECT_EQ(root->opcode(), HloOpcode::kTuple); 114 EXPECT_EQ(root->opcode(), HloOpcode::kTuple);
|
H A D | bfloat16_conversion_folding_test.cc | 39 hlo.opcode() == HloOpcode::kTuple || 49 hlo.opcode() == HloOpcode::kTuple || 57 if (hlo.opcode() == HloOpcode::kAdd || hlo.opcode() == HloOpcode::kTuple ||
|
H A D | bfloat16_normalization_test.cc | 40 hlo.opcode() == HloOpcode::kTuple || 50 hlo.opcode() == HloOpcode::kTuple || 58 if (hlo.opcode() == HloOpcode::kAdd || hlo.opcode() == HloOpcode::kTuple ||
|
H A D | bfloat16_conversion_folding.cc | 146 if (hlo->opcode() == HloOpcode::kTuple || //
|
H A D | hlo_opcode.h | 129 V(kTuple, "tuple", kHloOpcodeIsVariadic) \
|
H A D | while_loop_simplifier.cc | 157 if (while_body_root->opcode() != HloOpcode::kTuple) { 184 if (while_init->opcode() != HloOpcode::kTuple) { 309 if (while_body_root->opcode() != HloOpcode::kTuple) {
|
H A D | hlo_element_type_converter.cc | 62 // non-nested tuple. This assumption is currently safe as only kTuple, kInfeed, 128 opcode == HloOpcode::kTuple || opcode == HloOpcode::kConvert ||
|
H A D | hlo_value.cc | 133 case HloOpcode::kTuple:
|
H A D | while_loop_invariant_code_motion.cc | 111 case HloOpcode::kTuple:
|
H A D | instruction_fusion.cc | 80 case HloOpcode::kTuple:
|
H A D | hlo_dataflow_analysis.cc | 515 CHECK_EQ(tuple->opcode(), HloOpcode::kTuple); 519 // kTuple instruction's value sets. 565 case HloOpcode::kTuple: 763 case HloOpcode::kTuple:
|
H A D | hlo_graph_dumper.cc | 930 case HloOpcode::kTuple: 1077 if (instr->opcode() != HloOpcode::kTuple && shape_is_multidim) { 1244 if (instr == root || instr->opcode() != HloOpcode::kTuple) {
|
H A D | bfloat16_normalization.cc | 320 if (hlo->opcode() == HloOpcode::kTuple || //
|
H A D | hlo_instruction.cc | 304 CHECK_EQ(HloOpcode::kTuple, opcode); 1024 if (fused_root->opcode() == HloOpcode::kTuple) { 1030 if (clone->opcode() == HloOpcode::kTuple) { 1041 if (fused_root->opcode() == HloOpcode::kTuple) { 1054 if (instruction_to_fuse->opcode() == HloOpcode::kTuple) { 1155 return CreateVariadic(tuple_shape, HloOpcode::kTuple, elements); 1364 case HloOpcode::kTuple: 1728 case HloOpcode::kTuple: 2592 case HloOpcode::kTuple:
|
H A D | hlo_verifier.cc | 397 case HloOpcode::kTuple: 461 case HloOpcode::kTuple:
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
H A D | tuple_thunk.h | 40 : Thunk(Kind::kTuple, hlo_instruction),
|
H A D | thunk.h | 55 kTuple, member in class:xla::gpu::Thunk::Kind
|
H A D | while_transformer.cc | 358 ExprTree(HloOpcode::kTuple, tuple_index_, 439 ExprTree tuple0(HloOpcode::kTuple, tuple_index_,
|
H A D | fusion_merger_test.cc | 292 EXPECT_EQ(HloOpcode::kTuple, root->opcode());
|
/external/v8/src/compiler/ |
H A D | types.h | 289 enum Kind { kHeapConstant, kOtherNumberConstant, kTuple, kUnion, kRange }; enumerator in enum:v8::internal::compiler::BitsetType::TypeBase::Kind 474 TupleType(int length, Zone* zone) : StructuralType(kTuple, length, zone) {} 481 DCHECK(IsKind(type, kTuple)); 581 bool IsTuple() { return IsKind(TypeBase::kTuple); }
|
/external/v8/src/ast/ |
H A D | ast-types.h | 367 kTuple, enumerator in enum:v8::internal::AstBitsetType::AstTypeBase::Kind 651 : AstStructuralType(kTuple, length, zone) {} 659 DCHECK(IsKind(type, kTuple)); 826 bool IsTuple() { return IsKind(AstTypeBase::kTuple); }
|