/external/clang/test/CodeGen/ |
H A D | regparm.c | 14 extern void FASTCALL reduced(char b, double c, foo* d, double e, int f); 23 // CHECK: call void @reduced(i8 inreg signext 0, {{.*}} %struct.foo* inreg null 24 reduced(0, 0.0, 0, 0.0, 0);
|
/external/python/cpython2/Lib/lib2to3/ |
H A D | btm_utils.py | 130 reduced = reduce_tree(child, new_node) 131 if reduced is not None: 132 new_node.children.append(reduced) 138 reduced = reduce_tree(child, new_node) 139 if reduced: 140 new_node.children.append(reduced) 142 # delete the group if all of the children were reduced to None 230 reduced = reduce_tree(child, new_node) 231 if reduced is not None: 232 new_node.children.append(reduced) [all...] |
/external/python/cpython3/Lib/lib2to3/ |
H A D | btm_utils.py | 129 reduced = reduce_tree(child, new_node) 130 if reduced is not None: 131 new_node.children.append(reduced) 137 reduced = reduce_tree(child, new_node) 138 if reduced: 139 new_node.children.append(reduced) 141 # delete the group if all of the children were reduced to None 229 reduced = reduce_tree(child, new_node) 230 if reduced is not None: 231 new_node.children.append(reduced) [all...] |
/external/tensorflow/tensorflow/compiler/xla/service/ |
H A D | reduce_precision_insertion.cc | 96 HloInstruction* reduced = instruction->parent()->AddInstruction( local 99 TF_RETURN_IF_ERROR(instruction->ReplaceAllUsesWith(reduced)); 137 HloInstruction* reduced = nullptr; local 143 reduced = user; 148 if (!reduced) { 149 reduced = instruction->parent()->AddInstruction( 154 TF_RETURN_IF_ERROR(instruction->ReplaceOperandWith(i, reduced));
|
/external/clang/utils/analyzer/ |
H A D | reducer.pl | 64 `cp $srcFile $file.reduced`; 65 print "$prog: generated '$file.reduced";
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
H A D | reduce_join_op_test.py | 113 keep_dims: Whether or not to retain reduced dimensions. 263 reduced = string_ops.reduce_join(placeholder, reduction_indices=0) 264 output_array = reduced.eval(feed_dict={placeholder.name: input_array}) 266 self.assertAllEqual(truth_shape, reduced.get_shape()) 276 reduced = string_ops.reduce_join( 278 output_array_dim_zero = reduced.eval(feed_dict={placeholder.name: [0]}) 279 output_array_dim_one = reduced.eval(feed_dict={placeholder.name: [1]}) 282 self.assertAllEqual(truth_shape, reduced.get_shape()) 368 reduced = string_ops.reduce_join( 372 reduced [all...] |
H A D | conv1d_test.py | 47 reduced = array_ops.squeeze(c) 48 output = reduced.eval()
|
H A D | linalg_grad_test.py | 50 reduced = math_ops.reduce_sum(determinants) 51 sum_grad = gradients_impl.gradients(reduced, batch_identity)[0]
|
H A D | sparse_ops_test.py | 694 reduced = sparse_ops.sparse_reduce_sum(sp_t, axes) 697 reduced, 698 reduced.eval().shape) 702 reduced = sparse_ops.sparse_reduce_sum(sp_t, -1) 704 reduced, 705 reduced.eval().shape)
|
/external/skia/tests/ |
H A D | ClipStackTest.cpp | 860 // reduced stack. 1031 // Get the reduced version of the stack. 1034 const GrReducedClip* reduced = new (storage.get()) GrReducedClip(stack, queryBounds, caps); local 1037 reduced->maskElements().isEmpty() || 1038 SkClipStack::kInvalidGenID != reduced->maskGenID(), 1041 if (!reduced->maskElements().isEmpty()) { 1042 REPORTER_ASSERT(reporter, reduced->hasScissor(), testCase.c_str()); 1046 REPORTER_ASSERT(reporter, reduced->maskRequiresAA() == doAA, testCase.c_str()); 1049 // Build a new clip stack based on the reduced clip elements 1051 if (GrReducedClip::InitialState::kAllOut == reduced 1099 const GrReducedClip* reduced = new (storage.get()) GrReducedClip(stack, bounds, caps); local [all...] |
H A D | PathOpsConicLineIntersectionTest.cpp | 110 SkPoint reduced[3]; local 113 SkPath::Verb order1 = SkReduceOrder::Conic(floatConic, reduced);
|
/external/skqp/tests/ |
H A D | ClipStackTest.cpp | 860 // reduced stack. 1031 // Get the reduced version of the stack. 1034 const GrReducedClip* reduced = new (storage.get()) GrReducedClip(stack, queryBounds, caps); local 1037 reduced->maskElements().isEmpty() || 1038 SkClipStack::kInvalidGenID != reduced->maskGenID(), 1041 if (!reduced->maskElements().isEmpty()) { 1042 REPORTER_ASSERT(reporter, reduced->hasScissor(), testCase.c_str()); 1046 REPORTER_ASSERT(reporter, reduced->maskRequiresAA() == doAA, testCase.c_str()); 1049 // Build a new clip stack based on the reduced clip elements 1051 if (GrReducedClip::InitialState::kAllOut == reduced 1099 const GrReducedClip* reduced = new (storage.get()) GrReducedClip(stack, bounds, caps); local [all...] |
H A D | PathOpsConicLineIntersectionTest.cpp | 110 SkPoint reduced[3]; local 113 SkPath::Verb order1 = SkReduceOrder::Conic(floatConic, reduced);
|
/external/eigen/unsupported/test/ |
H A D | cxx11_tensor_argmax.cpp | 64 Tensor<Tuple<DenseIndex, float>, 0, DataLayout> reduced; local 66 reduced = index_tuples.reduce( 71 VERIFY_IS_EQUAL(maxi(), reduced(0).second); 96 Tensor<Tuple<DenseIndex, float>, 0, DataLayout> reduced; local 98 reduced = index_tuples.reduce( 103 VERIFY_IS_EQUAL(mini(), reduced(0).second); 190 // Expect max to be in the first index of the reduced dimension 212 // Expect max to be in the last index of the reduced dimension 248 // Expect min to be in the first index of the reduced dimension 270 // Expect min to be in the last index of the reduced dimensio [all...] |
/external/gemmlowp/meta/generators/ |
H A D | streams_common.py | 126 reduced = aggregators[:reduced_count] 128 emitter.EmitVSumReduce('u32', len(aggregators), 4, reduced, aggregators) 130 for temp in reduced: 133 for temp in reduced: 136 emitter.EmitVStoreA(1, 32, reduced,
|
H A D | quantized_mul_kernels_common.py | 521 reduced = aggregators[:reduced_count] 522 emitter.EmitVSumReduce('u32', len(aggregators), 4, reduced, aggregators) 523 return reduced 572 reduced = _GenerateAggregatorReduce(self.asm_emitter, aggregators, kernel_m, 576 registers, reduced,
|
/external/turbine/java/com/google/turbine/deps/ |
H A D | Dependencies.java | 142 Set<String> reduced = new HashSet<>(directJarsToTargets.keySet()); 154 reduced.add(dep.getPath()); 165 return Collections2.filter(transitiveClasspath, Predicates.in(reduced));
|
/external/tensorflow/tensorflow/core/kernels/ |
H A D | reduction_ops_common.h | 125 // Permutation of reduced dims needed to put reduction dimensions at the end 175 // A temporary tensor whose size matches the size of the reduced 218 // all reduced dimensions are last and reuse the 2-D -> 1-D case. 228 const int64 reduced = shuffled.NumElements() / unreduced; variable 231 const_shuffled.shaped<T, 2>({unreduced, reduced}),
|
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/streams/ |
H A D | rewrite.rb | 213 reduced = [] 257 reduced.unshift( operation ) if operation 260 @operations.replace( reduced )
|
/external/capstone/ |
H A D | config.mk | 25 # such as @regs_read/write & @group. The amount of binary size reduced is
|
/external/eigen/Eigen/src/Core/arch/CUDA/ |
H A D | PacketMathHalf.h | 865 float reduced = predux<Packet8f>(af); 866 return Eigen::half(reduced); 871 float reduced = predux_max<Packet8f>(af); 872 return Eigen::half(reduced); 877 float reduced = predux_min<Packet8f>(af); 878 return Eigen::half(reduced); 883 float reduced = predux_mul<Packet8f>(af); 884 return Eigen::half(reduced);
|
/external/libjpeg-turbo/simd/ |
H A D | jidctred-mmx.asm | 2 ; jidctred.asm - reduced-size IDCT (MMX) 16 ; This file contains inverse-DCT routines that produce reduced-size 98 ; producing a reduced-size 4x4 output block. 492 ; producing a reduced-size 2x2 output block.
|
H A D | jidctred-sse2-64.asm | 2 ; jidctred.asm - reduced-size IDCT (64-bit SSE2) 17 ; This file contains inverse-DCT routines that produce reduced-size 99 ; producing a reduced-size 4x4 output block. 402 ; producing a reduced-size 2x2 output block.
|
H A D | jidctred-sse2.asm | 2 ; jidctred.asm - reduced-size IDCT (SSE2) 16 ; This file contains inverse-DCT routines that produce reduced-size 98 ; producing a reduced-size 4x4 output block. 413 ; producing a reduced-size 2x2 output block.
|
/external/tensorflow/tensorflow/cc/gradients/ |
H A D | math_grad.cc | 607 // input_shape: 1-D Tensor, the shape of the Tensor being reduced. 664 // The partial derivative for any input along a "reduced" dimension 714 // all gradients are also divided by the size of reduced groups. 769 // The partial derivative for any input along a "reduced" dimension 901 auto reduced = Cast(cpu_scope, reduction_indices_pos, DataType::DT_INT32); local 907 auto other = SetDiff1D(cpu_scope, idx, reduced).out; 911 Concat(cpu_scope, std::initializer_list<Input>{reduced, other}, 0); 914 auto reduced_num = Prod(cpu_scope, Gather(scope, input_shape, reduced), 0);
|